config.name,config.backend.name,config.backend.version,config.backend._target_,config.backend.task,config.backend.library,config.backend.model,config.backend.processor,config.backend.device,config.backend.device_ids,config.backend.seed,config.backend.inter_op_num_threads,config.backend.intra_op_num_threads,config.backend.model_kwargs.trust_remote_code,config.backend.processor_kwargs.trust_remote_code,config.backend.hub_kwargs.trust_remote_code,config.backend.no_weights,config.backend.device_map,config.backend.torch_dtype,config.backend.eval_mode,config.backend.to_bettertransformer,config.backend.low_cpu_mem_usage,config.backend.attn_implementation,config.backend.cache_implementation,config.backend.autocast_enabled,config.backend.autocast_dtype,config.backend.torch_compile,config.backend.torch_compile_target,config.backend.quantization_scheme,config.backend.deepspeed_inference,config.backend.peft_type,config.scenario.name,config.scenario._target_,config.scenario.iterations,config.scenario.duration,config.scenario.warmup_runs,config.scenario.input_shapes.batch_size,config.scenario.input_shapes.num_choices,config.scenario.input_shapes.sequence_length,config.scenario.new_tokens,config.scenario.latency,config.scenario.memory,config.scenario.energy,config.scenario.generate_kwargs.max_new_tokens,config.scenario.generate_kwargs.min_new_tokens,config.launcher.name,config.launcher._target_,config.launcher.device_isolation,config.launcher.device_isolation_action,config.launcher.numactl,config.launcher.start_method,config.environment.cpu,config.environment.cpu_count,config.environment.cpu_ram_mb,config.environment.system,config.environment.machine,config.environment.platform,config.environment.processor,config.environment.python_version,config.environment.gpu,config.environment.gpu_count,config.environment.gpu_vram_mb,config.environment.optimum_benchmark_version,config.environment.optimum_benchmark_commit,config.environment.transformers_version,config.environment.transformers_commit,config.environment.accelerate_version,config.environment.accelerate_commit,config.environment.diffusers_version,config.environment.diffusers_commit,config.environment.optimum_version,config.environment.optimum_commit,config.environment.timm_version,config.environment.timm_commit,config.environment.peft_version,config.environment.peft_commit,report.traceback,config.backend.model_type,report.prefill.memory.unit,report.prefill.memory.max_ram,report.prefill.memory.max_global_vram,report.prefill.memory.max_process_vram,report.prefill.memory.max_reserved,report.prefill.memory.max_allocated,report.prefill.latency.unit,report.prefill.latency.count,report.prefill.latency.total,report.prefill.latency.mean,report.prefill.latency.stdev,report.prefill.latency.p50,report.prefill.latency.p90,report.prefill.latency.p95,report.prefill.latency.p99,report.prefill.latency.values,report.prefill.throughput.unit,report.prefill.throughput.value,report.prefill.energy.unit,report.prefill.energy.cpu,report.prefill.energy.ram,report.prefill.energy.gpu,report.prefill.energy.total,report.prefill.efficiency.unit,report.prefill.efficiency.value,report.decode.memory.unit,report.decode.memory.max_ram,report.decode.memory.max_global_vram,report.decode.memory.max_process_vram,report.decode.memory.max_reserved,report.decode.memory.max_allocated,report.decode.latency.unit,report.decode.latency.count,report.decode.latency.total,report.decode.latency.mean,report.decode.latency.stdev,report.decode.latency.p50,report.decode.latency.p90,report.decode.latency.p95,report.decode.latency.p99,report.decode.latency.values,report.decode.throughput.unit,report.decode.throughput.value,report.decode.energy.unit,report.decode.energy.cpu,report.decode.energy.ram,report.decode.energy.gpu,report.decode.energy.total,report.decode.efficiency.unit,report.decode.efficiency.value,report.per_token.memory,report.per_token.latency.unit,report.per_token.latency.count,report.per_token.latency.total,report.per_token.latency.mean,report.per_token.latency.stdev,report.per_token.latency.p50,report.per_token.latency.p90,report.per_token.latency.p95,report.per_token.latency.p99,report.per_token.latency.values,report.per_token.throughput.unit,report.per_token.throughput.value,report.per_token.energy,report.per_token.efficiency,config.backend.hub_kwargs.revision,config.backend.hub_kwargs.force_download,config.backend.hub_kwargs.local_files_only,report.load.memory.unit,report.load.memory.max_ram,report.load.memory.max_global_vram,report.load.memory.max_process_vram,report.load.memory.max_reserved,report.load.memory.max_allocated,report.load.latency.unit,report.load.latency.count,report.load.latency.total,report.load.latency.mean,report.load.latency.stdev,report.load.latency.p50,report.load.latency.p90,report.load.latency.p95,report.load.latency.p99,report.load.latency.values,report.load.throughput,report.load.energy.unit,report.load.energy.cpu,report.load.energy.ram,report.load.energy.gpu,report.load.energy.total,report.load.efficiency bfloat16-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,.,.,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 373, in cached_file raise EnvironmentError( OSError: . does not appear to have a file named config.json. Checkout 'https://huggingface.co/./tree/None' for available files. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,a,a,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/a/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6694747d-2be61c275c0d8282412a95e5;a32edf69-6e44-4e64-915b-6ac371637313) Repository Not Found for url: https://huggingface.co/a/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: a is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-7B,Qwen/Qwen-7B,cuda,0,42,,,True,,,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",qwen,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-66b,facebook/opt-66b,cuda,0,42,,,True,,,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",opt,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,stabilityai/stablelm-base-alpha-7b,stabilityai/stablelm-base-alpha-7b,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1274.925056,17205.559296,0.0,16559.112192,16333.262848,s,10,0.7347490692138672,0.07347490692138672,0.0008231328224625486,0.0731901741027832,0.07358341827392578,0.0747619010925293,0.0757046873474121,"[0.07594038391113281, 0.0731588134765625, 0.07317903900146484, 0.07322681427001954, 0.07322089385986329, 0.07319683074951172, 0.07315487670898438, 0.073321533203125, 0.07318351745605468, 0.07316636657714844]",tokens/s,3484.1827057216015,kWh,8.673447870860135e-07,4.7526480536833555e-07,5.32812672802923e-06,6.670736320483579e-06,tokens/kWh,38376573.09492364,MB,1274.925056,17205.559296,0.0,16559.112192,16464.787968,s,10,23.57336938476562,2.3573369384765623,0.0017160793725913698,2.3573675537109375,2.3593944091796875,2.3600200805664064,2.3605206176757814,"[2.358213623046875, 2.35471435546875, 2.357318603515625, 2.355459228515625, 2.35741650390625, 2.35803955078125, 2.35925537109375, 2.355737548828125, 2.360645751953125, 2.35656884765625]",tokens/s,26.725072250687244,kWh,2.774767362372912e-05,1.5204941629149124e-05,0.00013034514847717112,0.00017329776373004935,tokens/kWh,363536.13943995733,,s,629,24.194589706420892,0.038465166464898094,0.008734289734392974,0.03740364837646484,0.0375412483215332,0.03766393051147461,0.11087921478271484,"[0.0372592658996582, 0.03727360153198242, 0.03724697494506836, 0.03731353759765625, 0.03726847839355469, 0.03729100799560547, 0.03736883163452148, 0.03733708953857422, 0.037364734649658206, 0.0373125114440918, 0.03750707244873047, 0.0376627197265625, 0.037473281860351565, 0.037582847595214845, 0.03777228927612305, 0.037596160888671876, 0.037525505065917966, 0.037556224822998044, 0.03783475112915039, 0.03784396743774414, 0.037443584442138675, 0.037340160369873046, 0.037313568115234376, 0.037318656921386716, 0.03735036849975586, 0.03727155303955078, 0.0373125114440918, 0.03736371231079102, 0.03734220886230469, 0.03733913421630859, 0.037379070281982424, 0.037395454406738284, 0.03734527969360352, 0.037381118774414065, 0.037370880126953124, 0.03737702560424805, 0.037372928619384765, 0.037343231201171875, 0.037379070281982424, 0.03738623809814453, 0.037408767700195314, 0.037408767700195314, 0.03741491317749023, 0.037370880126953124, 0.037410816192626956, 0.037397502899169925, 0.0373831672668457, 0.037397502899169925, 0.037440513610839846, 0.03739340972900391, 0.03740367889404297, 0.03754083251953125, 0.03745587158203125, 0.03747532653808594, 0.03743948745727539, 0.037465087890625, 0.03744153594970703, 0.03758489608764649, 0.03751116943359375, 0.03749273681640625, 0.03744870376586914, 0.037528575897216795, 0.11087769317626953, 0.037294078826904296, 0.0372490234375, 0.03722751998901367, 0.03724390411376953, 0.03727158355712891, 0.03726230239868164, 0.03728384017944336, 0.03724697494506836, 0.0372674560546875, 0.037233665466308595, 0.03732275390625, 0.037275646209716795, 0.03730636978149414, 0.03727360153198242, 0.03731558227539063, 0.03728076934814453, 0.0373043212890625, 0.037282817840576174, 0.03730944061279297, 0.03736576080322265, 0.03728998565673828, 0.03734220886230469, 0.03735859298706055, 0.037335041046142575, 0.0373309440612793, 0.037364734649658206, 0.037356544494628906, 0.03731455993652344, 0.037372928619384765, 0.03733913421630859, 0.03731660842895508, 0.0373422737121582, 0.03738412857055664, 0.03740979385375977, 0.037362686157226564, 0.037321727752685545, 0.03739033508300781, 0.037356544494628906, 0.03734220886230469, 0.037338111877441404, 0.037359710693359374, 0.037365665435791014, 0.037364734649658206, 0.037479423522949216, 0.03760844802856445, 0.03760947036743164, 0.037449726104736326, 0.03743641662597656, 0.03744563293457031, 0.03743641662597656, 0.03746099090576172, 0.037456897735595705, 0.03743436813354492, 0.037553150177001955, 0.03742617416381836, 0.037394432067871096, 0.037456897735595705, 0.03751628875732422, 0.03770880126953125, 0.03751935958862305, 0.03744153594970703, 0.037515262603759765, 0.11082342529296875, 0.03732787322998047, 0.03723571014404297, 0.03729510498046875, 0.037381118774414065, 0.037253120422363284, 0.0372408332824707, 0.03725823974609375, 0.037477375030517575, 0.03723571014404297, 0.037275646209716795, 0.037370880126953124, 0.037748737335205076, 0.037612545013427735, 0.03729919815063477, 0.037351425170898435, 0.037310462951660156, 0.037288959503173826, 0.03733708953857422, 0.03728384017944336, 0.03747225570678711, 0.03740979385375977, 0.03750809478759766, 0.03736883163452148, 0.037321727752685545, 0.037359615325927735, 0.037362686157226564, 0.03736064147949219, 0.037566463470458986, 0.03739648056030274, 0.03738521575927734, 0.037400577545166014, 0.037566463470458986, 0.037392383575439454, 0.03746201705932617, 0.037356544494628906, 0.03735551834106445, 0.03734527969360352, 0.037410816192626956, 0.037408767700195314, 0.037356544494628906, 0.03731763076782227, 0.03734732818603516, 0.03735859298706055, 0.037335041046142575, 0.037378047943115236, 0.03736678314208984, 0.037473281860351565, 0.0374917106628418, 0.037407745361328126, 0.03747840118408203, 0.03737702560424805, 0.03736371231079102, 0.037362686157226564, 0.03810201644897461, 0.03798527908325195, 0.03765350341796875, 0.03742105484008789, 0.03754291152954101, 0.03739852905273437, 0.03745587158203125, 0.03747635269165039, 0.037454849243164064, 0.11098931121826172, 0.037294078826904296, 0.03726438522338867, 0.03725209426879883, 0.03723468780517578, 0.037282817840576174, 0.03726335906982422, 0.03727769470214844, 0.037253120422363284, 0.03743334579467773, 0.03729612731933594, 0.03748556900024414, 0.037381118774414065, 0.03728998565673828, 0.03733299255371094, 0.03732992172241211, 0.03726438522338867, 0.037512191772460936, 0.037343231201171875, 0.03729305648803711, 0.037335041046142575, 0.03732582473754883, 0.037321727752685545, 0.03740979385375977, 0.03736883163452148, 0.03734838485717774, 0.037347297668457034, 0.037372928619384765, 0.037395454406738284, 0.03737497711181641, 0.03732992172241211, 0.03736883163452148, 0.037369857788085936, 0.037405696868896485, 0.03738828659057617, 0.03741183853149414, 0.03765043258666992, 0.03746815872192383, 0.03740262222290039, 0.03741183853149414, 0.03741798400878906, 0.03741388702392578, 0.03735657501220703, 0.037453857421875, 0.03741078567504883, 0.037402591705322265, 0.037400577545166014, 0.03739955139160156, 0.037407745361328126, 0.03741900634765625, 0.037397502899169925, 0.03742310333251953, 0.037438465118408204, 0.03745177459716797, 0.03748966217041016, 0.03749587249755859, 0.03741689682006836, 0.03745075225830078, 0.03743641662597656, 0.037407745361328126, 0.03743334579467773, 0.03741183853149414, 0.0374917106628418, 0.1109381103515625, 0.0372490234375, 0.03724800109863281, 0.03725414276123047, 0.03723980712890625, 0.037275646209716795, 0.03722854232788086, 0.03721420669555664, 0.03727155303955078, 0.03724595260620117, 0.03726540756225586, 0.03727974319458008, 0.03727462387084961, 0.037269504547119144, 0.03727974319458008, 0.03727462387084961, 0.03731660842895508, 0.03732582473754883, 0.037326881408691406, 0.03736470413208008, 0.037285888671875, 0.037340160369873046, 0.037310462951660156, 0.03728793716430664, 0.03741190338134766, 0.03750598526000976, 0.03734732818603516, 0.03799552154541016, 0.037874687194824216, 0.03769343948364258, 0.03743231964111328, 0.037526527404785154, 0.037482494354248046, 0.03739955139160156, 0.03745075225830078, 0.03745382308959961, 0.037381118774414065, 0.03744467163085938, 0.037461952209472654, 0.03751116943359375, 0.037482494354248046, 0.03745280075073242, 0.03748044967651367, 0.03746412658691406, 0.037415870666503905, 0.03750604629516602, 0.037373023986816405, 0.03741276931762695, 0.03756851196289063, 0.03741183853149414, 0.037473281860351565, 0.037572608947753904, 0.03753881454467774, 0.037506145477294923, 0.037513118743896484, 0.037427200317382815, 0.03748051071166992, 0.03746912002563477, 0.03751833724975586, 0.03743436813354492, 0.03743436813354492, 0.03745177459716797, 0.03749273681640625, 0.11103846740722656, 0.03733196640014649, 0.037335041046142575, 0.037370880126953124, 0.037364734649658206, 0.037348350524902346, 0.03736070251464844, 0.03725920104980469, 0.03730636978149414, 0.03732275390625, 0.03731148910522461, 0.037353473663330077, 0.03734527969360352, 0.03742924880981445, 0.037495807647705076, 0.037323776245117186, 0.03738828659057617, 0.03735859298706055, 0.03737702560424805, 0.037338111877441404, 0.03765862274169922, 0.037579776763916016, 0.037795841217041014, 0.03745177459716797, 0.03743747329711914, 0.037372894287109375, 0.037370880126953124, 0.03741593551635742, 0.037367809295654295, 0.0373831672668457, 0.0374015998840332, 0.03741491317749023, 0.03746713638305664, 0.037438465118408204, 0.03741497421264649, 0.03747423934936524, 0.037422080993652344, 0.037425151824951174, 0.03738623809814453, 0.03739955139160156, 0.03747430419921875, 0.03743436813354492, 0.03744460678100586, 0.03742822265625, 0.037400577545166014, 0.037425151824951174, 0.03741183853149414, 0.037425151824951174, 0.03740262222290039, 0.03743027114868164, 0.03742924880981445, 0.03740364837646484, 0.037437599182128904, 0.03781308746337891, 0.037515262603759765, 0.037456897735595705, 0.03742310333251953, 0.037440513610839846, 0.03739033508300781, 0.037493759155273435, 0.037427200317382815, 0.037425151824951174, 0.03755321502685547, 0.11092371368408203, 0.03732582473754883, 0.0372490234375, 0.03725414276123047, 0.037275646209716795, 0.03724697494506836, 0.037294078826904296, 0.03729612731933594, 0.037405696868896485, 0.03741798400878906, 0.03727974319458008, 0.03730636978149414, 0.03737497711181641, 0.03751321411132812, 0.037359615325927735, 0.03735756683349609, 0.03763916778564453, 0.03731455993652344, 0.037338111877441404, 0.03788185501098633, 0.03783782577514649, 0.037395454406738284, 0.037294078826904296, 0.037354496002197264, 0.037303295135498044, 0.03736576080322265, 0.037372928619384765, 0.03734527969360352, 0.03737395095825195, 0.03736883163452148, 0.03739657592773438, 0.03747625732421875, 0.03744563293457031, 0.03740262222290039, 0.037407745361328126, 0.037425151824951174, 0.03750707244873047, 0.03734630584716797, 0.03742105484008789, 0.03746918487548828, 0.03749990463256836, 0.03740467071533203, 0.03741491317749023, 0.03743743896484375, 0.037479423522949216, 0.03748147201538086, 0.03744153594970703, 0.03745894241333008, 0.03747635269165039, 0.037612545013427735, 0.037459968566894535, 0.03749785614013672, 0.03745280075073242, 0.03753062438964844, 0.03751628875732422, 0.03775692749023438, 0.03777536010742188, 0.03748454284667969, 0.037548030853271484, 0.03748863983154297, 0.03743129730224609, 0.037533695220947266, 0.037817344665527344, 0.1111551971435547, 0.03733606338500976, 0.03726847839355469, 0.03726438522338867, 0.037233665466308595, 0.03728384017944336, 0.03726540756225586, 0.037275646209716795, 0.03727360153198242, 0.0373196792602539, 0.037272575378417966, 0.03729510498046875, 0.03735244750976562, 0.03727360153198242, 0.03730640029907226, 0.037335006713867185, 0.03738521575927734, 0.037384193420410154, 0.03743641662597656, 0.037356544494628906, 0.03731148910522461, 0.037338111877441404, 0.037375038146972656, 0.03748140716552734, 0.037375999450683595, 0.03730745697021484, 0.037350334167480466, 0.037369857788085936, 0.037359615325927735, 0.03738726425170898, 0.037364734649658206, 0.03737395095825195, 0.037338111877441404, 0.0373493766784668, 0.03739136123657227, 0.03740364837646484, 0.037359649658203126, 0.037349407196044924, 0.037357505798339845, 0.03742931365966797, 0.03738822555541992, 0.037577728271484374, 0.037482494354248046, 0.03742310333251953, 0.03749068832397461, 0.03738726425170898, 0.037416961669921874, 0.03744460678100586, 0.03744768142700195, 0.03745280075073242, 0.03751935958862305, 0.03744153594970703, 0.03749683380126953, 0.037440513610839846, 0.0374466552734375, 0.037483551025390624, 0.03744559860229492, 0.03745792007446289, 0.03748863983154297, 0.0374835205078125, 0.03746815872192383, 0.03750809478759766, 0.03751731109619141, 0.11094220733642578, 0.037303295135498044, 0.03727769470214844, 0.03729817581176758, 0.03722649765014648, 0.03727360153198242, 0.03731558227539063, 0.03727872085571289, 0.03729510498046875, 0.037351425170898435, 0.03748147201538086, 0.037465087890625, 0.03740671920776367, 0.03735756683349609, 0.03744563293457031, 0.03743539047241211, 0.037465087890625, 0.037449726104736326, 0.037407745361328126, 0.03743027114868164, 0.03739136123657227, 0.03744460678100586, 0.03751935958862305, 0.0374917106628418, 0.03756851196289063, 0.03755929565429687, 0.03742617416381836, 0.037498878479003905, 0.03753062438964844, 0.03747225570678711, 0.03749683380126953, 0.037422080993652344, 0.03742310333251953, 0.03748454284667969, 0.03759414291381836, 0.0381429443359375, 0.03773235321044922, 0.037479423522949216, 0.03750604629516602, 0.03741183853149414, 0.037397502899169925, 0.037572608947753904, 0.037972991943359374, 0.037544960021972655, 0.037456897735595705, 0.03743743896484375, 0.037438465118408204, 0.037440513610839846, 0.0374917106628418, 0.03743743896484375, 0.03743641662597656, 0.037459968566894535, 0.03745075225830078, 0.037456897735595705, 0.03746201705932617, 0.03748863983154297, 0.037440513610839846, 0.03746201705932617, 0.03743948745727539, 0.037512191772460936, 0.03748966217041016, 0.03749478530883789, 0.03760947036743164, 0.11087980651855468, 0.03724800109863281, 0.03723667144775391, 0.03723161697387695, 0.03729817581176758, 0.03725107192993164, 0.0373043212890625, 0.037272575378417966, 0.03727667236328125, 0.03723980712890625, 0.0373043212890625, 0.03730124664306641, 0.03733708953857422, 0.037348350524902346, 0.037269504547119144, 0.037336097717285154, 0.03731760025024414, 0.037292030334472655, 0.0373196792602539, 0.037275646209716795, 0.0372768325805664, 0.037329822540283206, 0.0374005126953125, 0.03751935958862305, 0.03744153594970703, 0.037372928619384765, 0.03737702560424805, 0.037379070281982424, 0.03734527969360352, 0.03729305648803711, 0.03737395095825195, 0.03738726425170898, 0.037359615325927735, 0.03736678314208984, 0.037425151824951174, 0.03746099090576172, 0.03744153594970703, 0.03742105484008789, 0.037427200317382815, 0.037449726104736326, 0.03744460678100586, 0.03741183853149414, 0.03741593551635742, 0.03743436813354492, 0.03750400161743164, 0.03740262222290039, 0.03739648056030274, 0.03743641662597656, 0.03748147201538086, 0.0374128646850586, 0.03743231964111328, 0.03746713638305664, 0.03746099090576172, 0.037554241180419924, 0.03750396728515625, 0.03766473770141601, 0.0375551986694336, 0.03748659133911133, 0.03742310333251953, 0.03751935958862305, 0.03745894241333008, 0.03803750228881836, 0.037667873382568356]",tokens/s,25.99754770104956,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/gpt-neo-2.7B,EleutherAI/gpt-neo-2.7B,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1227.554816,6568.804352,0.0,5922.357248,5577.217024,s,10,0.3875218887329101,0.03875218887329101,0.0005976895743779658,0.03850379180908203,0.03912730751037598,0.03980675868988037,0.040350319633483886,"[0.04048620986938477, 0.03849168014526367, 0.03847951889038086, 0.03851590347290039, 0.03846636962890625, 0.03852860641479492, 0.03867119979858399, 0.038453792572021486, 0.03845228958129883, 0.038976318359375]",tokens/s,6606.078455001588,kWh,4.5823433294468773e-07,2.510902127607213e-07,2.4094022289845053e-06,3.1187267746899145e-06,tokens/kWh,82084779.62147014,MB,1227.554816,6568.804352,0.0,5922.357248,5677.7216,s,10,14.478346435546875,1.4478346435546876,0.014679247415614043,1.443683837890625,1.4628969482421874,1.4730921508789063,1.4812483129882812,"[1.45308349609375, 1.436747802734375, 1.4325145263671875, 1.4430782470703125, 1.4361597900390626, 1.4442894287109376, 1.453, 1.435554443359375, 1.46063134765625, 1.483287353515625]",tokens/s,43.51325635179165,kWh,1.6928844253607194e-05,9.274113035161314e-06,5.823664628781611e-05,8.443960357658461e-05,tokens/kWh,746095.4022937894,,s,629,14.801351669311504,0.02353156068253024,0.00464661210312113,0.022737855911254882,0.023635149383544923,0.023778515243530272,0.06156992446899414,"[0.023541759490966797, 0.02326937675476074, 0.023763967514038087, 0.02347007942199707, 0.023569408416748046, 0.023447551727294923, 0.02348134422302246, 0.023415807723999024, 0.023657472610473632, 0.023631872177124022, 0.02368921661376953, 0.02353561592102051, 0.023417856216430662, 0.023410688400268553, 0.023570432662963867, 0.02377120018005371, 0.023501760482788087, 0.02326425552368164, 0.023229440689086913, 0.023142431259155275, 0.024096736907958983, 0.024209407806396483, 0.024054784774780274, 0.023744543075561522, 0.02316796875, 0.02287615966796875, 0.02266726493835449, 0.02264678382873535, 0.0227194881439209, 0.02268057632446289, 0.02264473533630371, 0.023155712127685548, 0.022921215057373046, 0.022602752685546876, 0.02266009521484375, 0.022658048629760744, 0.02270310401916504, 0.022692863464355468, 0.02267955207824707, 0.022812671661376953, 0.02267955207824707, 0.022717439651489257, 0.022617151260375976, 0.022699968338012695, 0.02269696044921875, 0.02269593620300293, 0.022761472702026365, 0.022622207641601562, 0.02267238426208496, 0.022618112564086915, 0.022635520935058592, 0.02271334457397461, 0.02271027183532715, 0.02260479927062988, 0.02275328063964844, 0.022624256134033204, 0.022590463638305663, 0.022562816619873048, 0.022625280380249024, 0.022665216445922853, 0.022592512130737305, 0.02272870445251465, 0.06170111846923828, 0.02266316795349121, 0.022772735595703125, 0.0226693115234375, 0.022541311264038084, 0.022664192199707032, 0.022609920501708985, 0.023350271224975586, 0.02389606475830078, 0.024268800735473633, 0.023440383911132814, 0.023582719802856447, 0.02329292869567871, 0.023193599700927735, 0.022578176498413087, 0.022788095474243163, 0.02267136001586914, 0.022624256134033204, 0.022623231887817383, 0.022766592025756836, 0.022615039825439453, 0.022675455093383787, 0.02265395164489746, 0.02284851264953613, 0.02269491195678711, 0.022690879821777345, 0.022737855911254882, 0.02274406433105469, 0.02265190315246582, 0.022773792266845703, 0.022687711715698243, 0.022722560882568358, 0.02269491195678711, 0.02268057632446289, 0.022788095474243163, 0.02259660720825195, 0.02267136001586914, 0.02272051239013672, 0.022606847763061523, 0.022717439651489257, 0.022617088317871094, 0.022566944122314452, 0.022665184020996095, 0.022664192199707032, 0.02262015914916992, 0.022622207641601562, 0.022811647415161132, 0.023334911346435547, 0.02293350410461426, 0.02285260772705078, 0.02262015914916992, 0.022707199096679686, 0.02267955207824707, 0.022787071228027343, 0.02265088081359863, 0.022833152770996092, 0.022649856567382814, 0.022656000137329102, 0.022572032928466795, 0.02268671989440918, 0.0226693115234375, 0.022820863723754883, 0.022583295822143554, 0.06156697463989258, 0.022623231887817383, 0.022586368560791017, 0.022709280014038085, 0.022613983154296875, 0.022527999877929687, 0.022541311264038084, 0.02273587226867676, 0.022564863204956053, 0.022603776931762694, 0.022708223342895507, 0.02267136001586914, 0.022601728439331056, 0.02281881523132324, 0.02306764793395996, 0.023434240341186522, 0.023177215576171875, 0.02305638313293457, 0.022536191940307617, 0.022701055526733398, 0.022642688751220705, 0.022675455093383787, 0.02261299133300781, 0.022829055786132812, 0.02269900894165039, 0.022610944747924806, 0.022675455093383787, 0.022591552734375, 0.022380544662475587, 0.022483903884887694, 0.022589439392089843, 0.022814720153808594, 0.022915071487426757, 0.022384639739990234, 0.02262937545776367, 0.022640640258789063, 0.022664192199707032, 0.022597631454467772, 0.02270310401916504, 0.02270207977294922, 0.022607872009277344, 0.022591487884521484, 0.022817792892456053, 0.022581247329711913, 0.022595584869384764, 0.02260585594177246, 0.023783391952514648, 0.023414783477783203, 0.02351206398010254, 0.02270310401916504, 0.022571008682250978, 0.02270515251159668, 0.02266214370727539, 0.022723583221435546, 0.022700031280517577, 0.02288025665283203, 0.02270310401916504, 0.02269388771057129, 0.02274508857727051, 0.02267852783203125, 0.02266111946105957, 0.022742015838623047, 0.02263859176635742, 0.06166425704956055, 0.022697984695434572, 0.02261299133300781, 0.02275328063964844, 0.022595584869384764, 0.022558719635009765, 0.022676479339599608, 0.022898687362670898, 0.02262015914916992, 0.02262835121154785, 0.02267955207824707, 0.02287718391418457, 0.0226693115234375, 0.022722560882568358, 0.02265088081359863, 0.022755327224731444, 0.022665216445922853, 0.02268057632446289, 0.022795263290405272, 0.02288332748413086, 0.022799360275268556, 0.022786048889160155, 0.02268057632446289, 0.02272051239013672, 0.02287718391418457, 0.02269388771057129, 0.023030784606933592, 0.024509439468383788, 0.023634944915771484, 0.023548927307128906, 0.023402496337890624, 0.023136255264282226, 0.02270412826538086, 0.022624256134033204, 0.02262835121154785, 0.022675455093383787, 0.02271232032775879, 0.023021568298339845, 0.02284339141845703, 0.02265907287597656, 0.02263859176635742, 0.02271027183532715, 0.02343731117248535, 0.023315456390380858, 0.02346188735961914, 0.02329702377319336, 0.023211008071899415, 0.023362560272216795, 0.023332927703857424, 0.023293888092041016, 0.02308403205871582, 0.022545408248901368, 0.02269491195678711, 0.022683647155761717, 0.022778879165649413, 0.023235584259033205, 0.02272051239013672, 0.023290880203247072, 0.023060480117797853, 0.02253209686279297, 0.022700031280517577, 0.0226693115234375, 0.022811647415161132, 0.06145843124389649, 0.022477823257446287, 0.02247987174987793, 0.02265190315246582, 0.022611967086791994, 0.022607872009277344, 0.02268262481689453, 0.023134208679199218, 0.023956480026245116, 0.023422975540161133, 0.02265497589111328, 0.02270412826538086, 0.023369728088378908, 0.023375871658325196, 0.023243776321411135, 0.02335436820983887, 0.02264678382873535, 0.02270310401916504, 0.022633472442626954, 0.022738943099975584, 0.022648832321166993, 0.02270412826538086, 0.02265497589111328, 0.022759424209594727, 0.02271232032775879, 0.022664192199707032, 0.022632448196411133, 0.02269081687927246, 0.022605823516845702, 0.022788095474243163, 0.022640640258789063, 0.02273587226867676, 0.02266726493835449, 0.022754304885864256, 0.023658496856689453, 0.023008256912231444, 0.022647808074951172, 0.02272768020629883, 0.022599679946899414, 0.022789119720458984, 0.02266214370727539, 0.023171072006225587, 0.023282688140869142, 0.022684736251831053, 0.022696895599365233, 0.022617088317871094, 0.02271129608154297, 0.02284339141845703, 0.022590463638305663, 0.022618112564086915, 0.022709247589111328, 0.02266009521484375, 0.022616064071655274, 0.02273587226867676, 0.022552576065063477, 0.022794240951538085, 0.022674432754516603, 0.022691839218139647, 0.022624256134033204, 0.022740991592407226, 0.022618112564086915, 0.022681600570678712, 0.022633472442626954, 0.06175539016723633, 0.02265190315246582, 0.022590463638305663, 0.02267033576965332, 0.02261299133300781, 0.022558719635009765, 0.022631423950195313, 0.02273689651489258, 0.022906879425048828, 0.02342300796508789, 0.023294944763183594, 0.023241727828979493, 0.022912000656127928, 0.023766016006469725, 0.023199743270874023, 0.02275027275085449, 0.022708192825317382, 0.022833120346069335, 0.02265292739868164, 0.022794240951538085, 0.022635520935058592, 0.02265088081359863, 0.02265497589111328, 0.02268057632446289, 0.022567935943603516, 0.02263859176635742, 0.02310348892211914, 0.02328473663330078, 0.023302143096923827, 0.023434272766113283, 0.022637535095214843, 0.022742015838623047, 0.022664192199707032, 0.022537216186523438, 0.022542335510253905, 0.02264473533630371, 0.022627328872680662, 0.022593536376953126, 0.02307072067260742, 0.023017471313476562, 0.02327347183227539, 0.023334911346435547, 0.023234560012817384, 0.023248895645141602, 0.022781951904296875, 0.02265907287597656, 0.02255462455749512, 0.022664192199707032, 0.022640640258789063, 0.022676479339599608, 0.02265395164489746, 0.02265395164489746, 0.02282499122619629, 0.023472095489501955, 0.023138303756713868, 0.022738943099975584, 0.022768640518188478, 0.02389811134338379, 0.0234967041015625, 0.023351295471191406, 0.023615488052368162, 0.023335935592651368, 0.02314035224914551, 0.06284288024902343, 0.02343731117248535, 0.02331443214416504, 0.023438335418701172, 0.02329190444946289, 0.023262208938598632, 0.023258111953735353, 0.023484479904174804, 0.023375808715820314, 0.02323967933654785, 0.022611967086791994, 0.022587392807006838, 0.022952959060668944, 0.02328883171081543, 0.023248895645141602, 0.02345471954345703, 0.023373823165893554, 0.023365631103515624, 0.023326719284057617, 0.02340355110168457, 0.023269344329833984, 0.023412736892700195, 0.023282688140869142, 0.023435264587402343, 0.023235584259033205, 0.022421503067016603, 0.022435840606689454, 0.02262326431274414, 0.0233287353515625, 0.02330931282043457, 0.02326425552368164, 0.023233535766601563, 0.023459840774536132, 0.023279680252075195, 0.02334713554382324, 0.023662591934204103, 0.022793216705322264, 0.0226079044342041, 0.022665184020996095, 0.022740991592407226, 0.022647808074951172, 0.022608896255493165, 0.022615039825439453, 0.022560768127441407, 0.023362560272216795, 0.02394316864013672, 0.023622655868530275, 0.023830591201782228, 0.022755264282226562, 0.022608928680419922, 0.0226079044342041, 0.022667200088500976, 0.022642688751220705, 0.022665216445922853, 0.022758399963378906, 0.02282700729370117, 0.022602752685546876, 0.02266214370727539, 0.022664192199707032, 0.02261299133300781, 0.022674432754516603, 0.02267033576965332, 0.022733823776245117, 0.06170624160766602, 0.022657024383544923, 0.02262835121154785, 0.02267750358581543, 0.0225218563079834, 0.022627328872680662, 0.022700031280517577, 0.02284339141845703, 0.022616064071655274, 0.022623231887817383, 0.022518783569335937, 0.022700031280517577, 0.022599679946899414, 0.022689792633056642, 0.022688831329345703, 0.022923200607299805, 0.02272153663635254, 0.022648832321166993, 0.022764543533325195, 0.02270515251159668, 0.022701055526733398, 0.02273689651489258, 0.02284339141845703, 0.022763519287109374, 0.022691839218139647, 0.02264371109008789, 0.022590463638305663, 0.022649856567382814, 0.022641664505004884, 0.022749183654785156, 0.02270515251159668, 0.022673408508300782, 0.022608896255493165, 0.02266111946105957, 0.0224849910736084, 0.02274406433105469, 0.022663232803344726, 0.02261292839050293, 0.022586368560791017, 0.02272153663635254, 0.022605823516845702, 0.02262118339538574, 0.02279840087890625, 0.02261702346801758, 0.02265193557739258, 0.02264067268371582, 0.022603744506835936, 0.022724576950073242, 0.022681600570678712, 0.02255564880371094, 0.022595584869384764, 0.02268262481689453, 0.022581247329711913, 0.02275328063964844, 0.022619136810302733, 0.022718463897705078, 0.022586368560791017, 0.024024063110351563, 0.024758272171020508, 0.023980031967163085, 0.023464960098266603, 0.023619583129882812, 0.023144447326660156, 0.061571071624755856, 0.022633472442626954, 0.022565887451171874, 0.022633472442626954, 0.022585344314575196, 0.022549503326416014, 0.022535167694091796, 0.022700031280517577, 0.02254745674133301, 0.024137727737426756, 0.023588863372802735, 0.02347110366821289, 0.02333286476135254, 0.023352319717407227, 0.023398399353027344, 0.023391231536865235, 0.02259660720825195, 0.022801408767700194, 0.022707199096679686, 0.02269388771057129, 0.02268876838684082, 0.022803455352783202, 0.023235584259033205, 0.022746112823486327, 0.022623231887817383, 0.02290995216369629, 0.023415807723999024, 0.023282688140869142, 0.023242752075195314, 0.023112703323364257, 0.023204864501953124, 0.023383039474487305, 0.02326425552368164, 0.023262208938598632, 0.023384096145629883, 0.023182304382324218, 0.022691839218139647, 0.023027711868286133, 0.02286079978942871, 0.02263859176635742, 0.022633472442626954, 0.022526975631713866, 0.02259660720825195, 0.022617088317871094, 0.023575551986694337, 0.024156160354614258, 0.023742464065551756, 0.02365644836425781, 0.0234967041015625, 0.023536640167236327, 0.02367487907409668, 0.023635967254638672, 0.023564287185668945, 0.023673856735229492, 0.023669759750366212, 0.023723007202148438, 0.024122367858886717, 0.023600128173828124, 0.02368409538269043, 0.023623680114746092, 0.023559167861938478, 0.02369331169128418, 0.023570432662963867, 0.06309478378295899, 0.023579647064208984, 0.02368409538269043, 0.023673856735229492, 0.023485471725463867, 0.023457759857177733, 0.02349772834777832, 0.02372915267944336, 0.023441408157348635, 0.023439359664916993, 0.023546880722045898, 0.02368409538269043, 0.023516159057617187, 0.023544832229614256, 0.023533567428588868, 0.024022016525268555, 0.02387353515625, 0.023682048797607422, 0.023596031188964844, 0.023601152420043944, 0.023605247497558594, 0.023573535919189453, 0.02356937599182129, 0.023679040908813478, 0.02378950309753418, 0.023529472351074218, 0.023577600479125976, 0.023599103927612306, 0.024201215744018553, 0.0237076473236084, 0.024094720840454102, 0.02288844871520996, 0.022687744140625, 0.022595584869384764, 0.022759424209594727, 0.022700031280517577, 0.02267136001586914, 0.023621631622314454, 0.023754751205444336, 0.02369843292236328, 0.0235284481048584, 0.02349260711669922, 0.02352332878112793, 0.02366262435913086, 0.023498720169067382, 0.023597055435180665, 0.02349567985534668, 0.023628799438476563, 0.02353459167480469, 0.023632896423339843, 0.023570432662963867, 0.023738367080688477, 0.023564287185668945, 0.023638015747070314, 0.0235930233001709, 0.023756736755371093, 0.023824384689331055, 0.023597055435180665, 0.023158784866333007, 0.023542783737182618, 0.02350592041015625, 0.023564287185668945, 0.023673856735229492]",tokens/s,42.496118871639354,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,1,1,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/1/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669325fc-458b920275c0dbec05aa92a0;ff6e0e69-7408-4036-819e-6ec66a26d11f) Repository Not Found for url: https://huggingface.co/1/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: 1 is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-7b,tiiuae/falcon-7b,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1272.496128,15196.48768,0.0,14550.040576,13944.957952,s,10,0.8173146514892577,0.08173146514892578,0.0015967359572456455,0.08116712188720704,0.08230683975219726,0.08438847770690917,0.0860537880706787,"[0.08647011566162109, 0.08184425354003906, 0.08111856079101562, 0.08103363037109375, 0.08127327728271484, 0.08121568298339844, 0.08125302124023437, 0.08110214233398437, 0.08101407623291015, 0.08098989105224609]",tokens/s,3132.2086241025218,kWh,9.611658960607748e-07,5.266747113459931e-07,5.150008636715389e-06,6.637849244122157e-06,tokens/kWh,38566708.97229084,MB,1272.496128,15196.48768,0.0,14550.040576,13945.037824,s,10,28.97916967773438,2.8979169677734378,0.00738291206369429,2.895778686523437,2.909371606445313,2.9102209594726562,2.9109004418945315,"[2.909182861328125, 2.896923828125, 2.8941103515625, 2.88953759765625, 2.894633544921875, 2.9110703125, 2.88955517578125, 2.901176025390625, 2.89083056640625, 2.9021494140625]",tokens/s,21.73975331267166,kWh,3.413491141034228e-05,1.870739413134244e-05,0.00012447001172588484,0.00017731231726756956,tokens/kWh,355305.2657076898,,s,629,29.66327807617189,0.047159424604406794,0.009708392241723048,0.04591308975219727,0.04647259597778321,0.04670177307128906,0.1273804360961914,"[0.04682137680053711, 0.04638412857055664, 0.04673843383789063, 0.04697292709350586, 0.046768192291259766, 0.046653377532958985, 0.04726067352294922, 0.046543872833251954, 0.04674252700805664, 0.047372287750244144, 0.04690022277832031, 0.04581171035766601, 0.04586598587036133, 0.04589363098144531, 0.04590591812133789, 0.04575027084350586, 0.045911041259765625, 0.04595199966430664, 0.04596121597290039, 0.04587212753295898, 0.045949951171875, 0.04596633529663086, 0.04597555160522461, 0.04585062408447266, 0.047710208892822265, 0.046963710784912106, 0.046769153594970705, 0.04592435073852539, 0.04597248077392578, 0.045843456268310545, 0.04584447860717773, 0.04589670562744141, 0.04594483184814453, 0.04594585418701172, 0.04620800018310547, 0.04636467361450195, 0.04595404815673828, 0.04583321762084961, 0.04589567947387695, 0.045927425384521485, 0.04598169708251953, 0.04584550476074219, 0.04599091339111328, 0.04595302581787109, 0.04592127990722656, 0.04583116912841797, 0.04643532943725586, 0.04587212753295898, 0.04591513442993164, 0.04578303909301758, 0.04583833694458008, 0.045794303894042966, 0.04605644989013672, 0.045843456268310545, 0.04585574340820313, 0.04589363098144531, 0.04593561553955078, 0.045830142974853515, 0.04591513442993164, 0.04581785583496094, 0.04637081527709961, 0.04662374496459961, 0.12934144592285157, 0.04629401779174805, 0.04579942321777344, 0.0459315185546875, 0.04580454254150391, 0.0459683837890625, 0.04575129699707031, 0.04602272033691406, 0.04591609573364258, 0.04598169708251953, 0.04579225540161133, 0.04597350311279297, 0.04591308975219727, 0.04593561553955078, 0.04588236618041992, 0.04597862243652344, 0.04598476791381836, 0.04589158248901367, 0.04582195281982422, 0.04597145462036133, 0.045848575592041016, 0.04598681640625, 0.045758464813232425, 0.045965312957763675, 0.04584447860717773, 0.04596223831176758, 0.045911041259765625, 0.045889537811279295, 0.045917182922363284, 0.04585779190063476, 0.04589875030517578, 0.045861888885498046, 0.04595507049560547, 0.045946880340576174, 0.04588339233398438, 0.04593766403198242, 0.0459233283996582, 0.04605644989013672, 0.04585574340820313, 0.045948928833007815, 0.045832191467285156, 0.046091262817382815, 0.045825023651123044, 0.04604620742797851, 0.04591923141479492, 0.045894657135009766, 0.04587519836425781, 0.04575948715209961, 0.04576870346069336, 0.04592844772338867, 0.04591308975219727, 0.045897727966308595, 0.045810688018798826, 0.04724838256835937, 0.04599603271484375, 0.045889537811279295, 0.04587116622924805, 0.04591609573364258, 0.04579020690917969, 0.046358528137207033, 0.04714700698852539, 0.04601241683959961, 0.04584755325317383, 0.12747264099121094, 0.04591513442993164, 0.045795326232910154, 0.04611379241943359, 0.04577280044555664, 0.04605132675170898, 0.04592639923095703, 0.045916160583496096, 0.04581785583496094, 0.04595814514160156, 0.04578303909301758, 0.04593664169311523, 0.045846527099609374, 0.04594790267944336, 0.04589158248901367, 0.045916160583496096, 0.045871105194091794, 0.04597350311279297, 0.045878273010253906, 0.0459950065612793, 0.045886463165283206, 0.04598988723754883, 0.04585574340820313, 0.04583119964599609, 0.045899742126464846, 0.04592230224609375, 0.045884414672851564, 0.04597043228149414, 0.04597145462036133, 0.046061569213867185, 0.0466319351196289, 0.04672512054443359, 0.04631654357910156, 0.0459233283996582, 0.04589977645874024, 0.045908992767333984, 0.045871105194091794, 0.0460052490234375, 0.045827072143554685, 0.04597452926635742, 0.04578508758544922, 0.04595814514160156, 0.045819969177246095, 0.045950912475585935, 0.04593971252441406, 0.04642611312866211, 0.046355457305908204, 0.04606771087646484, 0.045538303375244144, 0.04585779190063476, 0.04582809448242187, 0.04591513442993164, 0.04585062408447266, 0.04589056015014648, 0.04574310302734375, 0.04592236709594726, 0.045746112823486326, 0.04590591812133789, 0.04563353729248047, 0.045574176788330076, 0.04564169692993164, 0.045965312957763675, 0.045791233062744144, 0.1273671646118164, 0.04584755325317383, 0.04608102416992187, 0.04597555160522461, 0.045797374725341795, 0.045911041259765625, 0.04575743865966797, 0.045886463165283206, 0.04590387344360351, 0.04605440139770508, 0.045835262298583986, 0.045960193634033204, 0.04588236618041992, 0.04586905670166016, 0.04580044937133789, 0.045889537811279295, 0.04591001510620117, 0.04598067092895508, 0.04579840087890625, 0.04594073486328125, 0.045917182922363284, 0.04589158248901367, 0.045840385437011716, 0.04596223831176758, 0.04595199966430664, 0.04588748931884765, 0.045808639526367184, 0.045927425384521485, 0.0459048957824707, 0.045927425384521485, 0.045785152435302734, 0.045915073394775394, 0.045825023651123044, 0.04597350311279297, 0.0459048957824707, 0.04599705505371094, 0.04586393737792969, 0.04597248077392578, 0.04584243011474609, 0.045861888885498046, 0.04584550476074219, 0.045723648071289064, 0.04561417770385742, 0.045661087036132815, 0.045709312438964846, 0.04571852874755859, 0.04557926559448242, 0.04562944030761719, 0.04623769760131836, 0.04664831924438476, 0.04611388778686523, 0.04563753509521484, 0.045505535125732424, 0.04557619094848633, 0.04550656127929688, 0.04557209777832031, 0.045682689666748044, 0.04585472106933594, 0.04581990432739258, 0.04599808120727539, 0.045824001312255856, 0.04594278335571289, 0.045818878173828126, 0.12761497497558594, 0.04581171035766601, 0.04579635238647461, 0.045897727966308595, 0.04575027084350586, 0.04585881423950195, 0.04575129699707031, 0.04596736145019531, 0.04592947387695313, 0.04590182495117188, 0.04582195281982422, 0.045884414672851564, 0.045908992767333984, 0.04602265548706055, 0.045871105194091794, 0.045985790252685545, 0.045824001312255856, 0.04594585418701172, 0.04580556869506836, 0.04594585418701172, 0.04584755325317383, 0.04591206359863281, 0.04584447860717773, 0.046666751861572264, 0.046594112396240235, 0.04621615982055664, 0.046870494842529295, 0.04604006576538086, 0.045861888885498046, 0.04594278335571289, 0.045965312957763675, 0.04604620742797851, 0.04583935928344727, 0.045948928833007815, 0.045846527099609374, 0.04588544082641602, 0.04583935928344727, 0.046033920288085936, 0.04595609664916992, 0.04598169708251953, 0.04588339233398438, 0.04596121597290039, 0.04583321762084961, 0.045843456268310545, 0.04587519836425781, 0.04594585418701172, 0.04591513442993164, 0.04589977645874024, 0.04580044937133789, 0.04595814514160156, 0.04589363098144531, 0.045987838745117186, 0.045840385437011716, 0.045927425384521485, 0.04579635238647461, 0.04605644989013672, 0.04578303909301758, 0.04599193572998047, 0.04579945755004883, 0.045904865264892576, 0.04574521636962891, 0.04589459228515625, 0.04589363098144531, 0.12774092864990233, 0.04582809448242187, 0.045840385437011716, 0.04590796661376953, 0.04581990432739258, 0.04596121597290039, 0.04597248077392578, 0.045932544708251956, 0.045778976440429685, 0.04593967819213867, 0.0458158073425293, 0.04615884780883789, 0.046483455657958986, 0.04620800018310547, 0.0459048957824707, 0.04564377593994141, 0.04579328155517578, 0.04597350311279297, 0.04696780776977539, 0.0468592643737793, 0.046534656524658206, 0.04662988662719727, 0.04646092987060547, 0.04659609603881836, 0.04646604919433594, 0.046430206298828124, 0.0462110710144043, 0.04655411148071289, 0.04655411148071289, 0.04659404754638672, 0.045871105194091794, 0.04631039810180664, 0.04579635238647461, 0.04592639923095703, 0.04636979293823242, 0.04615577697753906, 0.04600012969970703, 0.04610153579711914, 0.046289886474609375, 0.04588339233398438, 0.045864959716796876, 0.04593664169311523, 0.046513153076171876, 0.04652134323120117, 0.046344192504882815, 0.046534656524658206, 0.04644454574584961, 0.04661862564086914, 0.046568447113037106, 0.046195743560791015, 0.04608918380737305, 0.04642201614379883, 0.046486526489257815, 0.0470384635925293, 0.04657561492919922, 0.04647116851806641, 0.04620083236694336, 0.04595609664916992, 0.045764606475830076, 0.04592639923095703, 0.04581785583496094, 0.04600012969970703, 0.045830142974853515, 0.12767129516601564, 0.04593971252441406, 0.045791233062744144, 0.04592230224609375, 0.04584447860717773, 0.045902847290039066, 0.04581273651123047, 0.04571033477783203, 0.04556288146972656, 0.0457237434387207, 0.04601551818847656, 0.04579110336303711, 0.0455813102722168, 0.045677566528320314, 0.04554035186767578, 0.0455731201171875, 0.04547788619995117, 0.045682815551757815, 0.04567027282714844, 0.0458158073425293, 0.04585472106933594, 0.04588339233398438, 0.045813758850097655, 0.04600012969970703, 0.045879295349121094, 0.045960193634033204, 0.04583731079101563, 0.04586700820922852, 0.04589977645874024, 0.04593971252441406, 0.04610969543457031, 0.04624076843261719, 0.045884414672851564, 0.04592639923095703, 0.045927425384521485, 0.045868030548095705, 0.045813758850097655, 0.04591308975219727, 0.04575539016723633, 0.04598886489868164, 0.04585062408447266, 0.045982719421386715, 0.045936672210693356, 0.04592227172851562, 0.045859840393066405, 0.04584755325317383, 0.045917182922363284, 0.04605542373657227, 0.04585881423950195, 0.04590694427490234, 0.04576972961425781, 0.04583833694458008, 0.0457523193359375, 0.04590182495117188, 0.04589875030517578, 0.04591820907592774, 0.04576665496826172, 0.04642816162109375, 0.045965312957763675, 0.045843456268310545, 0.045764606475830076, 0.0459048957824707, 0.045897727966308595, 0.12745216369628906, 0.04623360061645508, 0.04640671920776367, 0.04590380859375, 0.046063617706298826, 0.04595609664916992, 0.04629094314575195, 0.04687257766723633, 0.04696579360961914, 0.04659199905395508, 0.04647830581665039, 0.046445568084716796, 0.046437374114990236, 0.04647116851806641, 0.04650393676757812, 0.046496768951416016, 0.04641996765136719, 0.04637388610839844, 0.04581273651123047, 0.04592127990722656, 0.04594790267944336, 0.046042110443115236, 0.045868030548095705, 0.04598681640625, 0.045871135711669925, 0.04592022323608398, 0.04592127990722656, 0.04595507049560547, 0.04597353744506836, 0.04590895843505859, 0.04588544082641602, 0.04590182495117188, 0.04579840087890625, 0.045946880340576174, 0.04591206359863281, 0.046034942626953124, 0.04595814514160156, 0.04599808120727539, 0.045868030548095705, 0.04595404815673828, 0.04590387344360351, 0.04602675247192383, 0.045856769561767576, 0.04585369491577149, 0.045553665161132816, 0.04580556869506836, 0.04577382278442383, 0.04682963180541992, 0.04621408081054688, 0.04593971252441406, 0.04582092666625977, 0.04603289413452148, 0.045830142974853515, 0.045889537811279295, 0.045655040740966796, 0.045636768341064456, 0.045765472412109376, 0.04582092666625977, 0.0458158073425293, 0.04582915115356445, 0.04575948715209961, 0.045902816772460935, 0.046017536163330076, 0.12738559722900392, 0.045841407775878903, 0.04583321762084961, 0.045876224517822264, 0.04573081588745117, 0.04592537689208984, 0.04580147171020508, 0.045876224517822264, 0.045770751953125, 0.045879295349121094, 0.045884414672851564, 0.045930496215820314, 0.04581683349609375, 0.04592435073852539, 0.04575129699707031, 0.045884414672851564, 0.04586700820922852, 0.045949951171875, 0.04582195281982422, 0.045832191467285156, 0.045813758850097655, 0.04602880096435547, 0.04581273651123047, 0.04590591812133789, 0.04575743865966797, 0.04590182495117188, 0.04595404815673828, 0.04596633529663086, 0.0458260498046875, 0.04591513442993164, 0.045780990600585936, 0.04580556869506836, 0.0458260498046875, 0.04590591812133789, 0.04604313659667969, 0.04648038482666016, 0.04625612640380859, 0.0459683837890625, 0.04553731155395508, 0.04566422271728516, 0.04549836730957031, 0.046521438598632815, 0.0463553581237793, 0.04629401779174805, 0.045916160583496096, 0.04616294479370117, 0.045608959197998046, 0.045620223999023435, 0.04558233642578125, 0.045777919769287106, 0.04582809448242187, 0.045873153686523435, 0.04581785583496094, 0.0459048957824707, 0.04575539016723633, 0.04593664169311523, 0.04578611373901367, 0.04582809448242187, 0.04576972961425781, 0.045916160583496096, 0.04582092666625977, 0.04592947387695313, 0.04565404891967773, 0.12725142669677733, 0.04589363098144531, 0.04581683349609375, 0.04588851165771484, 0.045690879821777344, 0.04591001510620117, 0.04573081588745117, 0.04592639923095703, 0.046202880859375, 0.04597248077392578, 0.04585369491577149, 0.04687155151367187, 0.047094783782958984, 0.046371841430664064, 0.045861888885498046, 0.04590796661376953, 0.046317569732666014, 0.046048255920410154, 0.04583935928344727, 0.04591923141479492, 0.046502944946289065, 0.045970401763916015, 0.04623974227905273, 0.04628070449829102, 0.04585062408447266, 0.0459683837890625, 0.04579942321777344, 0.04652032089233398, 0.046025726318359376, 0.04589977645874024, 0.04600831985473633, 0.046266368865966793, 0.0459233283996582, 0.046018558502197264, 0.04636671829223633, 0.04596121597290039, 0.0458526725769043, 0.046053375244140625, 0.04625408172607422, 0.045827072143554685, 0.04593459320068359, 0.046311424255371096, 0.04611276626586914, 0.045871105194091794, 0.04621721649169922, 0.04650905609130859, 0.04627046585083008, 0.04590694427490234, 0.04616908645629883, 0.04654079818725586, 0.04582195281982422, 0.045835262298583986, 0.046162975311279296, 0.04636361694335937, 0.0463636474609375, 0.04615884780883789, 0.046102527618408204, 0.045548545837402345, 0.04548710250854492, 0.04593068695068359, 0.04588729476928711, 0.046152862548828125, 0.04556579208374024]",tokens/s,21.204669233953194,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,meta-llama/Llama-2-7b-hf,meta-llama/Llama-2-7b-hf,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.1,,,,MB,1253.154816,14563.147776,0.0,13916.700672,13738.740736,s,10,0.9872316818237306,0.09872316818237306,0.00029540592997361076,0.09862387084960937,0.0989060920715332,0.09923864479064941,0.09950468696594238,"[0.09957119750976562, 0.09851712036132812, 0.09855763244628907, 0.09868988800048828, 0.09853753662109375, 0.09862454223632812, 0.0986735076904297, 0.09883219146728515, 0.09862319946289062, 0.09860486602783203]",tokens/s,2593.109649065219,kWh,1.1676691196582938e-06,6.398277661366378e-07,6.082213471431282e-06,7.889710357226214e-06,tokens/kWh,32447325.49218726,MB,1253.154816,14563.147776,0.0,13916.700672,13889.779712,s,10,22.90496069335937,2.290496069335937,0.0013607297327055817,2.2909150390625,2.292089477539063,2.2921713256835936,2.2922368041992187,"[2.292253173828125, 2.29051806640625, 2.2892822265625, 2.288820068359375, 2.28832958984375, 2.28937060546875, 2.291588623046875, 2.29131201171875, 2.2914150390625, 2.2920712890625]",tokens/s,27.504958791859014,kWh,2.6973841217932878e-05,1.4782135219410359e-05,0.00011387216026976909,0.00015562813670711234,tokens/kWh,404811.11791863304,,s,629,23.755588615417466,0.037767231503048455,0.011791426483674353,0.036340736389160154,0.03649126434326172,0.03654409103393554,0.13552148864746094,"[0.03623321533203125, 0.0362342414855957, 0.036125694274902344, 0.03619123077392578, 0.036144126892089845, 0.036195327758789066, 0.036184127807617185, 0.03619219207763672, 0.03623014450073242, 0.03633356857299805, 0.03620249557495117, 0.03621478271484375, 0.03618918228149414, 0.03648409652709961, 0.036601856231689454, 0.03636019134521484, 0.03626496124267578, 0.03626803207397461, 0.036222976684570314, 0.036307968139648435, 0.03626803207397461, 0.036316158294677735, 0.036261886596679685, 0.03628646469116211, 0.036241409301757815, 0.03636019134521484, 0.03634380722045898, 0.036397056579589845, 0.036350975036621096, 0.03637452697753906, 0.03633663940429688, 0.0363397102355957, 0.03629260635375976, 0.036359169006347655, 0.03630387115478516, 0.036429824829101565, 0.036342784881591796, 0.036432895660400394, 0.03641753768920898, 0.03640627288818359, 0.03644313430786133, 0.03642060852050781, 0.03632844924926758, 0.036431873321533206, 0.03641856002807617, 0.036506622314453126, 0.036567039489746093, 0.03711180877685547, 0.03648716735839844, 0.03647488021850586, 0.03637964630126953, 0.03649126434326172, 0.03643904113769531, 0.03653836822509766, 0.03646361541748047, 0.03649433517456055, 0.03646361541748047, 0.03651583862304687, 0.036395008087158204, 0.03650764846801758, 0.03642060852050781, 0.036549633026123046, 0.1354956817626953, 0.036149246215820316, 0.03621279907226563, 0.03616966247558594, 0.036279296875, 0.03621478271484375, 0.03634175872802734, 0.03628236770629883, 0.03625471878051758, 0.03620560073852539, 0.036284385681152345, 0.03624448013305664, 0.03632128143310547, 0.03621785736083984, 0.03630080032348633, 0.036326400756835936, 0.03634688186645508, 0.0362690544128418, 0.036334590911865236, 0.03626803207397461, 0.036357120513916014, 0.03628339385986328, 0.036391937255859375, 0.03629260635375976, 0.03635507202148437, 0.036257793426513675, 0.03641241455078125, 0.036318206787109376, 0.036413440704345705, 0.03634175872802734, 0.03644313430786133, 0.0363765754699707, 0.03639295959472656, 0.036291584014892575, 0.03641446304321289, 0.036348926544189454, 0.03655680084228516, 0.03647180938720703, 0.036377601623535157, 0.036365310668945314, 0.03641753768920898, 0.03628236770629883, 0.036391937255859375, 0.036345855712890625, 0.03637247848510742, 0.03633049774169922, 0.03643084716796875, 0.03641548919677735, 0.03641548919677735, 0.0363059196472168, 0.03641958236694336, 0.03632230377197266, 0.03642572784423828, 0.036347904205322266, 0.03642879867553711, 0.036365310668945314, 0.03645337677001953, 0.036357120513916014, 0.03647590255737305, 0.036340736389160154, 0.03648716735839844, 0.0364442253112793, 0.036483009338378905, 0.1355335693359375, 0.036125694274902344, 0.03621174240112305, 0.036168670654296874, 0.036219905853271485, 0.036149246215820316, 0.03618815994262695, 0.03617279815673828, 0.03621478271484375, 0.03615948867797852, 0.036359169006347655, 0.036219905853271485, 0.03631206512451172, 0.036212734222412106, 0.03624755096435547, 0.03619635009765625, 0.03625881576538086, 0.03621068954467774, 0.03632844924926758, 0.036192256927490236, 0.036291584014892575, 0.036185089111328124, 0.03629056167602539, 0.03623219299316406, 0.0362690544128418, 0.03622195053100586, 0.03630284881591797, 0.03622505569458008, 0.036337665557861325, 0.036249568939208984, 0.03634688186645508, 0.036274177551269535, 0.03631513595581055, 0.03625676727294922, 0.03635302352905274, 0.03627724838256836, 0.03644313430786133, 0.036896766662597655, 0.036498432159423826, 0.03633152008056641, 0.0363612174987793, 0.036370433807373044, 0.03643801498413086, 0.03633049774169922, 0.03654246520996094, 0.036299774169921875, 0.036400127410888675, 0.03640524673461914, 0.03643801498413086, 0.03628543853759766, 0.03637865447998047, 0.036472801208496095, 0.036604927062988284, 0.03659366226196289, 0.03645337677001953, 0.0363765754699707, 0.036429824829101565, 0.03633663940429688, 0.03644518280029297, 0.036345855712890625, 0.03643699264526367, 0.0363612174987793, 0.036503551483154296, 0.1356441650390625, 0.03613695907592773, 0.03618409729003906, 0.03610620880126953, 0.03628339385986328, 0.03621478271484375, 0.03617587280273438, 0.03619126510620117, 0.03615024185180664, 0.03612876892089844, 0.03623833465576172, 0.03612876892089844, 0.036296703338623046, 0.03655372619628906, 0.03707596969604492, 0.03640217590332031, 0.036324352264404294, 0.03628339385986328, 0.03628851318359375, 0.03618304061889648, 0.0362690544128418, 0.036222976684570314, 0.03627212905883789, 0.03621887969970703, 0.03631206512451172, 0.03625471878051758, 0.0362977294921875, 0.03625369644165039, 0.036365310668945314, 0.03627212905883789, 0.03640422439575195, 0.036354049682617184, 0.03629875183105469, 0.03623833465576172, 0.03633152008056641, 0.03623526382446289, 0.03631411361694336, 0.036245502471923825, 0.036340736389160154, 0.03631411361694336, 0.036380672454833986, 0.03630182266235352, 0.036345855712890625, 0.03628441619873047, 0.036431873321533206, 0.03633152008056641, 0.03640115356445312, 0.036364288330078126, 0.03638784027099609, 0.036337696075439456, 0.0364062385559082, 0.036310016632080076, 0.036424705505371094, 0.036364288330078126, 0.03646054458618164, 0.036367359161376955, 0.036416511535644534, 0.03636633682250977, 0.03644211196899414, 0.03633561706542969, 0.03643904113769531, 0.03642367935180664, 0.036446208953857424, 0.13547007751464843, 0.03619635009765625, 0.0363765754699707, 0.036117504119873044, 0.036192256927490236, 0.0361338882446289, 0.036206592559814454, 0.03622604751586914, 0.03621887969970703, 0.036157440185546875, 0.03621683120727539, 0.03617488098144531, 0.036219905853271485, 0.03617174530029297, 0.03623014450073242, 0.036201473236083984, 0.03621376037597656, 0.03624038314819336, 0.03624345779418945, 0.03621068954467774, 0.03624857711791992, 0.03623116683959961, 0.03625881576538086, 0.03623526382446289, 0.0363612174987793, 0.0362342414855957, 0.036291584014892575, 0.036212734222412106, 0.03631718444824219, 0.03630387115478516, 0.03630182266235352, 0.03626496124267578, 0.036279296875, 0.03625471878051758, 0.036321311950683596, 0.03628335952758789, 0.036332542419433594, 0.03630284881591797, 0.03635200119018555, 0.03628646469116211, 0.03632332611083984, 0.03628441619873047, 0.03634380722045898, 0.03629260635375976, 0.03636633682250977, 0.036307968139648435, 0.036410369873046876, 0.03644723129272461, 0.03637452697753906, 0.036318206787109376, 0.036432895660400394, 0.03629363250732422, 0.03638784027099609, 0.03632332611083984, 0.03640627288818359, 0.0363765754699707, 0.03642777633666992, 0.03640422439575195, 0.03644416046142578, 0.036365345001220704, 0.03656803131103516, 0.03731763076782227, 0.036580352783203124, 0.13553152465820312, 0.036132865905761716, 0.036171775817871094, 0.03615948867797852, 0.036225025177001956, 0.03613081741333008, 0.03623833465576172, 0.036203521728515625, 0.03622092819213867, 0.03616972732543945, 0.03623526382446289, 0.036162559509277346, 0.03624857711791992, 0.03625574493408203, 0.03620761489868164, 0.03619430541992188, 0.036239360809326174, 0.03617279815673828, 0.03625574493408203, 0.03621478271484375, 0.0362690544128418, 0.036185089111328124, 0.03629875183105469, 0.0362496337890625, 0.036306911468505856, 0.03625881576538086, 0.036316158294677735, 0.0363059196472168, 0.036441089630126954, 0.03625267028808594, 0.03632230377197266, 0.03629363250732422, 0.03628543853759766, 0.03624959945678711, 0.03637247848510742, 0.03623628616333008, 0.036348926544189454, 0.036326400756835936, 0.036544513702392575, 0.03644927978515625, 0.03643494415283203, 0.036359169006347655, 0.03643699264526367, 0.03630182266235352, 0.03643392181396484, 0.0363765754699707, 0.03645849609375, 0.036435966491699216, 0.03671244812011719, 0.03657011032104492, 0.03645849609375, 0.036441089630126954, 0.0364820785522461, 0.03640009689331054, 0.03646771240234375, 0.03645337677001953, 0.036432895660400394, 0.036400127410888675, 0.036478977203369144, 0.036429824829101565, 0.03650457763671875, 0.03637350463867187, 0.03649228668212891, 0.13559706115722656, 0.03614310455322266, 0.03620454406738281, 0.036144126892089845, 0.03620556640625, 0.03619123077392578, 0.03628236770629883, 0.03633152008056641, 0.03630489730834961, 0.03616972732543945, 0.03622707366943359, 0.03627212905883789, 0.0362762222290039, 0.03625369644165039, 0.03626496124267578, 0.03622604751586914, 0.036318206787109376, 0.03620249557495117, 0.03632336044311523, 0.036282337188720704, 0.036337665557861325, 0.036279296875, 0.03634380722045898, 0.03642675018310547, 0.03636326217651367, 0.03623116683959961, 0.036362239837646484, 0.03625881576538086, 0.03640940856933594, 0.036348865509033206, 0.03642572784423828, 0.03633868789672851, 0.036370433807373044, 0.036354049682617184, 0.03641139221191406, 0.03632332611083984, 0.036421630859375, 0.03632332611083984, 0.03642777633666992, 0.0363694076538086, 0.03633663940429688, 0.036377601623535157, 0.03642367935180664, 0.03638681411743164, 0.03644931030273438, 0.03635094451904297, 0.03649228668212891, 0.036395008087158204, 0.03648102569580078, 0.03637452697753906, 0.03648102569580078, 0.036410369873046876, 0.036506622314453126, 0.03637145614624023, 0.03650559997558594, 0.03673088073730469, 0.037015552520751956, 0.03646156692504883, 0.03647999954223633, 0.036431873321533206, 0.0365291519165039, 0.036425758361816406, 0.03654345703125, 0.13622373962402343, 0.03623526382446289, 0.03624652862548828, 0.036201473236083984, 0.03625881576538086, 0.03614828872680664, 0.036270015716552736, 0.036257854461669924, 0.03630585479736328, 0.036245502471923825, 0.0363397102355957, 0.03622918319702149, 0.036332481384277346, 0.036291584014892575, 0.03630387115478516, 0.036212734222412106, 0.036274177551269535, 0.036187137603759766, 0.03628851318359375, 0.03622809600830078, 0.03626598358154297, 0.036178943634033206, 0.036279296875, 0.036209663391113284, 0.036278270721435545, 0.036345855712890625, 0.03661004638671875, 0.03644416046142578, 0.03638988876342773, 0.03632128143310547, 0.036377601623535157, 0.03633868789672851, 0.036393985748291016, 0.03639910507202149, 0.03642572784423828, 0.03630387115478516, 0.03640729522705078, 0.03633561706542969, 0.03647283172607422, 0.036365345001220704, 0.036372447967529295, 0.03631718444824219, 0.03642777633666992, 0.03640524673461914, 0.036822017669677735, 0.036400127410888675, 0.03647283172607422, 0.036380672454833986, 0.03639910507202149, 0.03632844924926758, 0.03641753768920898, 0.03637964630126953, 0.036511745452880856, 0.03631718444824219, 0.03643801498413086, 0.03637452697753906, 0.03646464157104492, 0.036429824829101565, 0.036441089630126954, 0.036377601623535157, 0.036503551483154296, 0.03645542526245117, 0.03651583862304687, 0.13559295654296874, 0.0361451530456543, 0.03627724838256836, 0.036245502471923825, 0.03632537460327148, 0.036208641052246096, 0.03627008056640625, 0.03625369644165039, 0.03631513595581055, 0.036201473236083984, 0.03630899047851562, 0.03622195053100586, 0.03632025527954102, 0.03626700973510742, 0.03628851318359375, 0.03631206512451172, 0.036299774169921875, 0.03625983810424805, 0.036322334289550784, 0.03626287841796875, 0.0363581428527832, 0.036334590911865236, 0.036347904205322266, 0.036280319213867186, 0.036364288330078126, 0.0363397102355957, 0.036367359161376955, 0.036278270721435545, 0.036397056579589845, 0.03634380722045898, 0.036483070373535154, 0.03635302352905274, 0.036364288330078126, 0.036350975036621096, 0.036362239837646484, 0.03634688186645508, 0.03638988876342773, 0.036279296875, 0.03640934371948242, 0.03638476943969727, 0.03646361541748047, 0.03638169479370117, 0.03641446304321289, 0.03638579177856445, 0.03642675018310547, 0.03638784027099609, 0.03642572784423828, 0.03638070297241211, 0.036494304656982425, 0.03639295959472656, 0.03650867080688477, 0.036437023162841795, 0.036506591796875, 0.03640627288818359, 0.03649638366699219, 0.036408321380615234, 0.03648102569580078, 0.03642060852050781, 0.03649126434326172, 0.03643801498413086, 0.03650969696044922, 0.036506622314453126, 0.03649433517456055, 0.13557452392578126, 0.03611238479614258, 0.03620556640625, 0.0361973762512207, 0.03626393508911133, 0.03621376037597656, 0.036275199890136715, 0.03624038314819336, 0.03709747314453125, 0.036370433807373044, 0.036362239837646484, 0.03622604751586914, 0.03623526382446289, 0.03623116683959961, 0.03626700973510742, 0.036334590911865236, 0.036370433807373044, 0.03622092819213867, 0.03630284881591797, 0.03621068954467774, 0.036326400756835936, 0.036291584014892575, 0.03632332611083984, 0.036310016632080076, 0.03632230377197266, 0.03630489730834961, 0.03640422439575195, 0.03621683120727539, 0.03638272094726563, 0.03631315231323242, 0.03639292907714844, 0.03633456039428711, 0.036354049682617184, 0.03632025527954102, 0.03638886260986328, 0.03633561706542969, 0.03640422439575195, 0.03638681411743164, 0.03639295959472656, 0.03638681411743164, 0.03645337677001953, 0.0363581428527832, 0.03639091110229492, 0.036342784881591796, 0.03642879867553711, 0.036416511535644534, 0.036506622314453126, 0.03644825744628906, 0.036410369873046876, 0.03640627288818359, 0.03647999954223633, 0.036408321380615234, 0.036483070373535154, 0.03643699264526367, 0.03649433517456055, 0.03638169479370117, 0.03651891326904297, 0.036408321380615234, 0.03652710342407227, 0.03640934371948242, 0.03659571075439453, 0.036647937774658204, 0.03654246520996094]",tokens/s,26.47797998959185,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-rw-1b,tiiuae/falcon-rw-1b,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1197.768704,3538.419712,0.0,2891.972608,2759.918592,s,10,0.22341420936584472,0.02234142093658447,0.00029111464340745864,0.022243712425231932,0.022395789527893065,0.02280235834121704,0.02312761339187622,"[0.023208927154541016, 0.022286048889160158, 0.02230544090270996, 0.022193695068359377, 0.02222368049621582, 0.022235776901245115, 0.02225164794921875, 0.02222073554992676, 0.02227689552307129, 0.022211360931396484]",tokens/s,11458.537070074872,kWh,2.6478590429809244e-07,1.4508975474799803e-07,1.2966988450313906e-06,1.706574504077481e-06,tokens/kWh,150008100.66501334,MB,1197.768704,3538.419712,0.0,2891.972608,2796.953088,s,10,10.40897344970703,1.040897344970703,0.013836335344277185,1.0432322387695312,1.0601688232421875,1.0615470581054687,1.062649645996094,"[1.0455521240234376, 1.0386646728515625, 1.0411461181640624, 1.045318359375, 1.0273031005859374, 1.0465562744140624, 1.059862548828125, 1.0212946166992187, 1.020350341796875, 1.06292529296875]",tokens/s,60.524700446587445,kWh,1.2491260500143105e-05,6.8447319731998554e-06,3.38512181621686e-05,5.3187210635511567e-05,tokens/kWh,1184495.2808624394,,s,629,10.591118341445927,0.016838025980041214,0.0027084463618811923,0.016556032180786134,0.016903577423095704,0.017006591796875,0.0388757894897461,"[0.017183744430541992, 0.017547264099121093, 0.0164454402923584, 0.01615564727783203, 0.01619046401977539, 0.016202783584594725, 0.016288736343383788, 0.01615667152404785, 0.01618636894226074, 0.015978495597839357, 0.016028671264648436, 0.01618636894226074, 0.01621196746826172, 0.016108543395996093, 0.01623961639404297, 0.016228351593017578, 0.016339967727661133, 0.01621196746826172, 0.01615667152404785, 0.016339967727661133, 0.01622425651550293, 0.01619558334350586, 0.016256032943725587, 0.016133087158203125, 0.01658880043029785, 0.01682329559326172, 0.01617715263366699, 0.0160819206237793, 0.01596723175048828, 0.016323583602905273, 0.01685196876525879, 0.016556032180786134, 0.0164771842956543, 0.016656383514404297, 0.016920576095581053, 0.016724992752075195, 0.01681305694580078, 0.01678643226623535, 0.016760831832885743, 0.016701440811157226, 0.01619865608215332, 0.01679462432861328, 0.01677107238769531, 0.01682022476196289, 0.01684787178039551, 0.016781312942504883, 0.01680076789855957, 0.016755712509155272, 0.016915456771850586, 0.01681407928466797, 0.016854015350341797, 0.016706560134887697, 0.01702195167541504, 0.017006591796875, 0.01817190361022949, 0.017079296112060546, 0.016927743911743166, 0.016397312164306642, 0.016886783599853517, 0.01675775909423828, 0.016715776443481444, 0.016750591278076172, 0.03967488098144531, 0.016723968505859374, 0.01683148765563965, 0.01675468826293945, 0.016748544692993163, 0.01680691146850586, 0.016763904571533202, 0.01634611129760742, 0.01618636894226074, 0.01624575996398926, 0.016157695770263672, 0.016228351593017578, 0.016247840881347658, 0.016202720642089843, 0.01614028739929199, 0.016332799911499024, 0.01618943977355957, 0.01623142433166504, 0.01623859214782715, 0.01620889663696289, 0.016129024505615236, 0.016438272476196288, 0.01619660758972168, 0.016285696029663087, 0.01617715263366699, 0.016285696029663087, 0.016397312164306642, 0.01671785545349121, 0.016700384140014648, 0.016703487396240235, 0.016709632873535156, 0.016746496200561522, 0.016728063583374024, 0.016755712509155272, 0.016706560134887697, 0.016745471954345705, 0.016733184814453125, 0.016503807067871093, 0.016092159271240233, 0.01623961639404297, 0.016265216827392577, 0.01618022346496582, 0.01617203140258789, 0.01619660758972168, 0.01620070457458496, 0.016243711471557617, 0.016240640640258788, 0.016275455474853515, 0.01617407989501953, 0.016175104141235352, 0.01617616081237793, 0.016330720901489258, 0.016235519409179687, 0.016309247970581055, 0.016776191711425782, 0.01681100845336914, 0.016750591278076172, 0.01665433692932129, 0.016192512512207033, 0.016539648056030275, 0.01661644744873047, 0.018372608184814454, 0.017739776611328126, 0.03963187026977539, 0.016698368072509767, 0.016691200256347655, 0.016886783599853517, 0.016745471954345705, 0.016942079544067384, 0.016694271087646484, 0.016748544692993163, 0.016698368072509767, 0.016744447708129884, 0.01618534469604492, 0.01680793571472168, 0.01674448013305664, 0.01625289535522461, 0.016152576446533205, 0.01619558334350586, 0.01617817687988281, 0.01638297653198242, 0.016021503448486327, 0.016123903274536132, 0.01622425651550293, 0.01621299171447754, 0.016350208282470705, 0.016863231658935548, 0.016655359268188476, 0.016781312942504883, 0.016699392318725585, 0.016690176010131837, 0.016614463806152342, 0.01670035171508789, 0.016703487396240235, 0.01613830375671387, 0.015962047576904295, 0.01658060836791992, 0.016659456253051756, 0.016687103271484375, 0.01673936080932617, 0.016813024520874024, 0.016729087829589845, 0.016312320709228514, 0.01619660758972168, 0.01618534469604492, 0.016153600692749022, 0.0164003849029541, 0.016694271087646484, 0.016714752197265623, 0.016680959701538087, 0.01678643226623535, 0.01618943977355957, 0.01615564727783203, 0.016157695770263672, 0.016157695770263672, 0.016149503707885742, 0.016265216827392577, 0.016666656494140626, 0.016766944885253907, 0.017048576354980468, 0.0167956485748291, 0.016656383514404297, 0.01665126419067383, 0.016630783081054687, 0.016694271087646484, 0.016151552200317384, 0.03896627044677734, 0.01616383934020996, 0.016137216567993166, 0.01620172882080078, 0.016252927780151367, 0.016313343048095702, 0.016250879287719726, 0.016242687225341796, 0.01622323226928711, 0.01658470344543457, 0.016138240814208983, 0.01617100715637207, 0.016175104141235352, 0.01619558334350586, 0.016760831832885743, 0.016792575836181642, 0.016777215957641603, 0.016708608627319335, 0.01620070457458496, 0.016202751159667968, 0.016244735717773438, 0.01617305564880371, 0.016194559097290038, 0.016558080673217773, 0.016731136322021483, 0.016686080932617187, 0.016731199264526368, 0.016703424453735353, 0.01676595115661621, 0.016717824935913086, 0.016752639770507814, 0.01679667282104492, 0.01679667282104492, 0.016694271087646484, 0.016748544692993163, 0.016714752197265623, 0.016358400344848634, 0.016690176010131837, 0.016773120880126953, 0.01678643226623535, 0.01678233528137207, 0.016738304138183592, 0.016834592819213866, 0.016740320205688476, 0.01676288032531738, 0.016677888870239257, 0.01679052734375, 0.01676697540283203, 0.016874496459960937, 0.01643724822998047, 0.016631807327270508, 0.016709632873535156, 0.016792575836181642, 0.016755712509155272, 0.0167956485748291, 0.016747520446777343, 0.016744447708129884, 0.016737279891967775, 0.016867328643798828, 0.016730112075805666, 0.0168724479675293, 0.01678335952758789, 0.016903167724609376, 0.03890892791748047, 0.01617919921875, 0.016270368576049805, 0.01621705627441406, 0.01660108757019043, 0.016745471954345705, 0.01677926445007324, 0.01689910316467285, 0.016774112701416016, 0.016761856079101564, 0.01762303924560547, 0.016952352523803713, 0.016784351348876955, 0.016491519927978517, 0.017101823806762697, 0.01685606384277344, 0.01678643226623535, 0.016763904571533202, 0.016490495681762696, 0.016471040725708007, 0.016533504486083983, 0.016028671264648436, 0.016012287139892577, 0.016111616134643555, 0.016047103881835938, 0.01603276824951172, 0.016052223205566405, 0.0160184326171875, 0.016077823638916015, 0.016087039947509766, 0.016053247451782226, 0.016058368682861326, 0.016029695510864257, 0.016048160552978516, 0.01603171157836914, 0.016046079635620117, 0.016062463760375977, 0.015996928215026854, 0.016051200866699217, 0.016035839080810545, 0.016314367294311523, 0.016440319061279296, 0.016181280136108398, 0.016033760070800783, 0.01599078369140625, 0.016002048492431642, 0.01615564727783203, 0.016094207763671875, 0.016301088333129883, 0.016063488006591797, 0.01625904083251953, 0.016184320449829103, 0.016283647537231445, 0.016259071350097656, 0.016260095596313476, 0.01620070457458496, 0.01618124771118164, 0.01619152069091797, 0.016125919342041016, 0.01602764892578125, 0.016030719757080078, 0.016119808197021485, 0.016004095077514647, 0.03882086563110351, 0.01680384063720703, 0.016724992752075195, 0.016715776443481444, 0.016729087829589845, 0.016759807586669923, 0.016257024765014647, 0.016846847534179688, 0.016252927780151367, 0.017087488174438475, 0.01804902458190918, 0.0170700798034668, 0.01677824020385742, 0.01679769515991211, 0.016745471954345705, 0.016805919647216797, 0.01670345687866211, 0.016192512512207033, 0.016159744262695314, 0.01616383934020996, 0.016281600952148437, 0.016755712509155272, 0.016730112075805666, 0.016744447708129884, 0.016729087829589845, 0.01617100715637207, 0.016146432876586913, 0.016167936325073243, 0.016194559097290038, 0.016540672302246092, 0.016738304138183592, 0.016863264083862305, 0.016797664642333985, 0.01616383934020996, 0.01623142433166504, 0.016175104141235352, 0.016187456130981444, 0.016136127471923827, 0.01621401596069336, 0.01625497627258301, 0.016740352630615234, 0.016631807327270508, 0.016356351852416993, 0.016733184814453125, 0.016703487396240235, 0.0168222713470459, 0.01686425590515137, 0.016891904830932617, 0.016915456771850586, 0.01681305694580078, 0.016349184036254884, 0.016730112075805666, 0.01680998420715332, 0.01681612777709961, 0.016928768157958983, 0.016270336151123048, 0.016333824157714845, 0.01637990379333496, 0.01679871940612793, 0.016734207153320312, 0.01683456039428711, 0.016777215957641603, 0.016881664276123046, 0.03941273498535156, 0.0168724479675293, 0.01681920051574707, 0.016965631484985352, 0.01679052734375, 0.016915456771850586, 0.016748544692993163, 0.016870399475097657, 0.016870399475097657, 0.01677414321899414, 0.01679052734375, 0.01685196876525879, 0.017006591796875, 0.016744447708129884, 0.01679155158996582, 0.01679155158996582, 0.016763904571533202, 0.01681305694580078, 0.016742399215698242, 0.01676595115661621, 0.016721920013427736, 0.016703519821166992, 0.016768991470336916, 0.016973888397216797, 0.01676691246032715, 0.016865280151367186, 0.016743423461914063, 0.016777215957641603, 0.016712703704833985, 0.01681407928466797, 0.016751615524291993, 0.017123327255249024, 0.01678950309753418, 0.016793600082397463, 0.01682329559326172, 0.016788480758666992, 0.016756736755371093, 0.016747520446777343, 0.01683148765563965, 0.016825344085693358, 0.017082368850708008, 0.01761484718322754, 0.01760665512084961, 0.017183744430541992, 0.016899072647094726, 0.016842752456665038, 0.01705779266357422, 0.016877567291259766, 0.016780288696289062, 0.01683660888671875, 0.016742399215698242, 0.016196640014648437, 0.01628976058959961, 0.01679667282104492, 0.01677824020385742, 0.016895999908447267, 0.01663692855834961, 0.016849920272827147, 0.01682636833190918, 0.016424959182739257, 0.017123327255249024, 0.01637068748474121, 0.016251903533935547, 0.0388935661315918, 0.016280576705932616, 0.016232448577880858, 0.016199680328369142, 0.01618227195739746, 0.016151552200317384, 0.016165887832641602, 0.016235519409179687, 0.01617919921875, 0.016289791107177733, 0.016252927780151367, 0.01620070457458496, 0.016154624938964843, 0.016280576705932616, 0.016151552200317384, 0.016194559097290038, 0.016161792755126952, 0.01616281509399414, 0.016157695770263672, 0.0161474552154541, 0.01618022346496582, 0.01620172882080078, 0.01617919921875, 0.016434175491333008, 0.016219135284423827, 0.01623859214782715, 0.016227327346801757, 0.01619046401977539, 0.01626316833496094, 0.016150527954101563, 0.01622425651550293, 0.01620479965209961, 0.01621299171447754, 0.01620172882080078, 0.016302080154418946, 0.016331775665283203, 0.01620582389831543, 0.016152576446533205, 0.01622118377685547, 0.016234495162963866, 0.01618739128112793, 0.01614847946166992, 0.01621401596069336, 0.016160768508911134, 0.01622118377685547, 0.01619558334350586, 0.016201759338378908, 0.016211936950683594, 0.01618739128112793, 0.016160768508911134, 0.01618227195739746, 0.016120832443237306, 0.01618227195739746, 0.016210943222045898, 0.01622528076171875, 0.01620070457458496, 0.01617715263366699, 0.016159744262695314, 0.016261119842529297, 0.016158720016479493, 0.016199680328369142, 0.01620070457458496, 0.016149503707885742, 0.038989822387695314, 0.016088064193725587, 0.016088064193725587, 0.01604198455810547, 0.01618227195739746, 0.01617203140258789, 0.01619660758972168, 0.016232448577880858, 0.01620070457458496, 0.01618124771118164, 0.016335872650146483, 0.01617203140258789, 0.01618227195739746, 0.01619660758972168, 0.01618022346496582, 0.016269311904907227, 0.016133119583129883, 0.01619660758972168, 0.01615564727783203, 0.016141311645507812, 0.01637580871582031, 0.016159744262695314, 0.016194559097290038, 0.01623347282409668, 0.01623142433166504, 0.01618124771118164, 0.016167936325073243, 0.01621504020690918, 0.016248832702636717, 0.016149503707885742, 0.016264192581176756, 0.01620172882080078, 0.016153600692749022, 0.01613926315307617, 0.01617715263366699, 0.016167936325073243, 0.016184320449829103, 0.01616281509399414, 0.01619865608215332, 0.016232511520385743, 0.016261056900024416, 0.016134143829345703, 0.016142335891723633, 0.016199680328369142, 0.01616383934020996, 0.01617100715637207, 0.01617919921875, 0.0161976318359375, 0.01619558334350586, 0.016144384384155275, 0.01619865608215332, 0.016191488265991212, 0.01620684814453125, 0.01616998481750488, 0.01619046401977539, 0.016282623291015624, 0.016149503707885742, 0.01618227195739746, 0.0162293758392334, 0.016165887832641602, 0.016191488265991212, 0.016157695770263672, 0.016175104141235352, 0.038830078125, 0.016130048751831053, 0.016566272735595702, 0.016849920272827147, 0.01683148765563965, 0.01686841583251953, 0.016824256896972655, 0.016954368591308593, 0.016969728469848632, 0.016861183166503906, 0.016913408279418944, 0.01683046340942383, 0.016915456771850586, 0.016867328643798828, 0.01682022476196289, 0.016928768157958983, 0.016930816650390625, 0.016937984466552734, 0.016951295852661134, 0.016946176528930663, 0.016891904830932617, 0.016957439422607423, 0.016866304397583007, 0.016889856338500975, 0.016884735107421875, 0.01683660888671875, 0.016837631225585938, 0.01683046340942383, 0.016837631225585938, 0.016888832092285155, 0.016913408279418944, 0.016894975662231446, 0.016833536148071288, 0.01680076789855957, 0.016859136581420898, 0.016973823547363282, 0.016882688522338866, 0.01681407928466797, 0.01683660888671875, 0.016869375228881836, 0.01683456039428711, 0.016849920272827147, 0.016921600341796874, 0.01683558464050293, 0.01681715202331543, 0.016905216217041014, 0.016888832092285155, 0.016909311294555664, 0.016899072647094726, 0.0168407039642334, 0.01680691146850586, 0.0168222713470459, 0.016913408279418944, 0.016915456771850586, 0.01686425590515137, 0.01754832077026367, 0.017020896911621095, 0.016879615783691407, 0.016910335540771485, 0.016885759353637696, 0.016912384033203123, 0.01698921585083008, 0.016853984832763673]",tokens/s,59.38938455050135,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,8,8,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/8/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6693242c-2b233f1a62002fc01034d83a;6695ebe3-fa08-4c8d-ad98-85f65090695a) Repository Not Found for url: https://huggingface.co/8/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: 8 is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-4B,,cuda,0,42,,,,,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,217063f5c507ed7cc255df7e1f64c4333a0b4dfe,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,,MB,1309.437952,10129.768448,0.0,9483.321344,8972.087296,s,10,0.4534943618774414,0.04534943618774414,0.00029954326796267103,0.04523150253295899,0.04545066413879394,0.045838788795471194,0.04614928852081299,"[0.04622691345214844, 0.04522739028930664, 0.045192062377929686, 0.04523561477661133, 0.04536441421508789, 0.04521945571899414, 0.04515244674682617, 0.04522380828857422, 0.04535683059692383, 0.045295425415039066]",tokens/s,5645.053643890395,kWh,5.35629256668254e-07,2.9350051654669867e-07,3.0042821922625344e-06,3.8334119654774866e-06,tokens/kWh,66781238.82991346,MB,1309.765632,10129.768448,0.0,9483.321344,9040.66816,s,10,23.087372558593746,2.3087372558593744,0.04120184853041573,2.3065739746093747,2.3587710693359374,2.359583288574219,2.3602330639648437,"[2.28953662109375, 2.252088623046875, 2.274980224609375, 2.30125048828125, 2.3118974609375, 2.245017578125, 2.353123046875, 2.340492431640625, 2.3603955078125, 2.358590576171875]",tokens/s,27.28764385817896,kWh,2.73109727129977e-05,1.4967316534074269e-05,7.950661714973509e-05,0.00012178490639680705,tokens/kWh,517305.4844311292,,s,629,23.45817082977295,0.037294389236522965,0.005443178154018897,0.037405696868896485,0.037789697265625,0.0379881477355957,0.08095006774902344,"[0.0353361930847168, 0.03522662353515625, 0.03511603164672852, 0.03521638488769531, 0.03536076736450195, 0.03508636856079102, 0.03518767929077148, 0.03538534545898438, 0.035244033813476565, 0.035253280639648436, 0.035331039428710936, 0.03532799911499023, 0.03532287979125977, 0.03535871887207031, 0.039669761657714846, 0.0384532470703125, 0.03801702499389648, 0.03760435104370117, 0.03749068832397461, 0.037456897735595705, 0.037510143280029294, 0.03745792007446289, 0.037579776763916016, 0.037664768218994144, 0.03749990463256836, 0.037544960021972655, 0.03743948745727539, 0.037477375030517575, 0.037482494354248046, 0.037392383575439454, 0.03743231964111328, 0.03745280075073242, 0.037405696868896485, 0.03785318374633789, 0.03738521575927734, 0.03750400161743164, 0.03730944061279297, 0.037438465118408204, 0.03737395095825195, 0.03749683380126953, 0.037362686157226564, 0.037479423522949216, 0.037324798583984374, 0.03507097625732422, 0.03511296081542969, 0.03512627029418945, 0.03502796936035156, 0.035043327331542966, 0.03505766296386719, 0.03520512008666992, 0.03510067367553711, 0.03532492828369141, 0.035154945373535154, 0.03526144027709961, 0.03535564804077149, 0.03552665710449219, 0.03511603164672852, 0.035236862182617186, 0.035102718353271486, 0.03506175994873047, 0.03511808013916016, 0.03548876953125, 0.08090009307861327, 0.03699097442626953, 0.037532737731933594, 0.03517433547973633, 0.035279872894287106, 0.03516928100585937, 0.03506175994873047, 0.03522457504272461, 0.035538944244384765, 0.035181568145751956, 0.035095550537109374, 0.035135486602783206, 0.035356670379638674, 0.03526860809326172, 0.03511603164672852, 0.03529523086547852, 0.03526041412353516, 0.03506585693359375, 0.03507814407348633, 0.03508019256591797, 0.03531980895996094, 0.035125247955322264, 0.03513651275634765, 0.035151870727539065, 0.03523788833618164, 0.0351539192199707, 0.03552870559692383, 0.03525222396850586, 0.03513248062133789, 0.03500230407714844, 0.03512934494018555, 0.03522457504272461, 0.03518463897705078, 0.03516108703613281, 0.03521331024169922, 0.03520614242553711, 0.03522150421142578, 0.03509862518310547, 0.03536383819580078, 0.03538534545898438, 0.03527475357055664, 0.03518668746948242, 0.03522355270385742, 0.035250175476074216, 0.035148799896240236, 0.035320831298828126, 0.03520819091796875, 0.035248126983642575, 0.03838259124755859, 0.03768217468261719, 0.03746303939819336, 0.037512191772460936, 0.03747532653808594, 0.037525505065917966, 0.03781836700439453, 0.03754291152954101, 0.03762995147705078, 0.03750707244873047, 0.03751321411132812, 0.03767295837402344, 0.03531980895996094, 0.03521843338012695, 0.03518668746948242, 0.08090214538574218, 0.035202079772949216, 0.035240928649902345, 0.03520409774780273, 0.03518054580688477, 0.03518873596191406, 0.03513139343261719, 0.03521331024169922, 0.03514265441894531, 0.035146751403808595, 0.03491635131835937, 0.03521331024169922, 0.03523276901245117, 0.03514265441894531, 0.035143680572509765, 0.035297279357910154, 0.03531161499023437, 0.03651583862304687, 0.03991449737548828, 0.038368255615234374, 0.037779457092285154, 0.03775283050537109, 0.03525529479980469, 0.035340286254882815, 0.0352542724609375, 0.03519385528564453, 0.03516416168212891, 0.03525222396850586, 0.03521535873413086, 0.037553150177001955, 0.03746099090576172, 0.03762790298461914, 0.03768524932861328, 0.037482494354248046, 0.037817344665527344, 0.03868979263305664, 0.03809689712524414, 0.03770675277709961, 0.03779379272460937, 0.03786137771606445, 0.03755724716186523, 0.03681484985351562, 0.03758796691894531, 0.037493759155273435, 0.037566463470458986, 0.035197952270507815, 0.03519385528564453, 0.035264511108398434, 0.035089408874511716, 0.035009536743164066, 0.035335166931152344, 0.03530854415893555, 0.03515289688110351, 0.035178497314453126, 0.03510681533813476, 0.03513859176635742, 0.03512828826904297, 0.03514780807495117, 0.035137504577636716, 0.03525529479980469, 0.03515084838867188, 0.03629875183105469, 0.0374466552734375, 0.081006591796875, 0.03506585693359375, 0.03546112060546875, 0.037599231719970705, 0.03793203353881836, 0.037473281860351565, 0.037364734649658206, 0.035178497314453126, 0.035555328369140625, 0.03515801620483398, 0.03523379135131836, 0.0353259506225586, 0.035214336395263675, 0.035093505859375, 0.03592806243896484, 0.03752243041992188, 0.035157054901123044, 0.0351077766418457, 0.0350750732421875, 0.03514265441894531, 0.035179519653320314, 0.035248126983642575, 0.03520614242553711, 0.03537510299682617, 0.035410945892333984, 0.03516108703613281, 0.03518668746948242, 0.03513139343261719, 0.03512934494018555, 0.03503104019165039, 0.03693670272827149, 0.03767705535888672, 0.03742105484008789, 0.037381118774414065, 0.03743948745727539, 0.03783987045288086, 0.037773311614990236, 0.03754905700683594, 0.03754598236083984, 0.037591041564941405, 0.03747123336791992, 0.0374835205078125, 0.037819393157958986, 0.03763302230834961, 0.03763814544677734, 0.035266559600830076, 0.037544960021972655, 0.03780915069580078, 0.037531646728515625, 0.03520614242553711, 0.035269630432128905, 0.03530035018920898, 0.03762688064575195, 0.037515262603759765, 0.03755724716186523, 0.037628929138183595, 0.03770470428466797, 0.03772723388671875, 0.037645313262939455, 0.03749478530883789, 0.037713920593261716, 0.038462528228759764, 0.037680065155029294, 0.08121241760253907, 0.03530956649780274, 0.03525529479980469, 0.0351723518371582, 0.03613695907592773, 0.03761459350585938, 0.03752755355834961, 0.03745894241333008, 0.037528575897216795, 0.03752755355834961, 0.03750400161743164, 0.037438465118408204, 0.03753676986694336, 0.03703500747680664, 0.03772825622558594, 0.03750092697143555, 0.03519385528564453, 0.035125247955322264, 0.03518873596191406, 0.03525120162963867, 0.03795558547973633, 0.0375470085144043, 0.03750707244873047, 0.03738726425170898, 0.03766579055786133, 0.03758899307250976, 0.03785830307006836, 0.0374200325012207, 0.03750297546386719, 0.03753881454467774, 0.037427200317382815, 0.037645313262939455, 0.03749990463256836, 0.03754086303710937, 0.037651454925537106, 0.03746099090576172, 0.03741183853149414, 0.037684223175048825, 0.037533695220947266, 0.03745587158203125, 0.03532799911499023, 0.036462593078613284, 0.03516723251342774, 0.03515596771240234, 0.0350904312133789, 0.0356864013671875, 0.035386367797851564, 0.035318782806396484, 0.03516211318969727, 0.03504537582397461, 0.03704934310913086, 0.03751935958862305, 0.03517750549316406, 0.03559318542480469, 0.036746238708496096, 0.037577728271484374, 0.03735756683349609, 0.037556289672851566, 0.03772102355957031, 0.037797889709472655, 0.03536588668823242, 0.03527372741699219, 0.03525632095336914, 0.08106598663330078, 0.035122177124023435, 0.0351723518371582, 0.035381248474121094, 0.03520307159423828, 0.03510681533813476, 0.0351539192199707, 0.03515084838867188, 0.035304447174072266, 0.03518259048461914, 0.0354252815246582, 0.035388416290283206, 0.03516211318969727, 0.035176448822021485, 0.03513446426391602, 0.03529420852661133, 0.03536281585693359, 0.03548160171508789, 0.03531468963623047, 0.035214336395263675, 0.035076095581054685, 0.03509657669067383, 0.035151870727539065, 0.03628854370117188, 0.037541854858398435, 0.03745177459716797, 0.03765862274169922, 0.03767295837402344, 0.037407745361328126, 0.037370880126953124, 0.03757881546020508, 0.03759916687011719, 0.037407745361328126, 0.035165184020996096, 0.03513958358764648, 0.03512319946289062, 0.035138561248779294, 0.03514777755737305, 0.035165184020996096, 0.03524095916748047, 0.03524505615234375, 0.035074047088623043, 0.03516928100585937, 0.03517440032958984, 0.03517337417602539, 0.035558399200439454, 0.03515084838867188, 0.03525222396850586, 0.03516416168212891, 0.035149822235107424, 0.035279872894287106, 0.036857856750488284, 0.037526527404785154, 0.03520512008666992, 0.034936832427978515, 0.03524198532104492, 0.035269630432128905, 0.035227649688720705, 0.035250175476074216, 0.03532799911499023, 0.03524607849121094, 0.03524095916748047, 0.03535564804077149, 0.08096870422363281, 0.03565055847167969, 0.03737395095825195, 0.037784576416015625, 0.03748966217041016, 0.037323776245117186, 0.03738623809814453, 0.03736166381835938, 0.03783270263671875, 0.03738828659057617, 0.037340160369873046, 0.03743641662597656, 0.03728179168701172, 0.03735039901733399, 0.03751321411132812, 0.03753472137451172, 0.03750400161743164, 0.03726028823852539, 0.037348350524902346, 0.03742617416381836, 0.03745587158203125, 0.03743027114868164, 0.03758796691894531, 0.03756032180786133, 0.03739648056030274, 0.037343231201171875, 0.037381118774414065, 0.03735859298706055, 0.03744460678100586, 0.03727667236328125, 0.03739340972900391, 0.03745177459716797, 0.035192832946777344, 0.0351201286315918, 0.035312641143798826, 0.03923353576660156, 0.0379504623413086, 0.03841331100463867, 0.0379791374206543, 0.037804031372070314, 0.037986305236816405, 0.03824947357177735, 0.037623809814453124, 0.03764019012451172, 0.03749683380126953, 0.03742617416381836, 0.03760639953613281, 0.03781222534179687, 0.03746303939819336, 0.03508531188964844, 0.035108863830566404, 0.035119102478027346, 0.03956326293945313, 0.03830579376220703, 0.037823486328125, 0.038196224212646485, 0.037923839569091795, 0.03751424026489258, 0.03748761749267578, 0.03773235321044922, 0.03755110549926758, 0.03745075225830078, 0.037479423522949216, 0.08333618927001953, 0.03750912094116211, 0.037395454406738284, 0.03716198348999023, 0.03787161636352539, 0.03743641662597656, 0.03738009643554688, 0.037454849243164064, 0.03520409774780273, 0.03507199859619141, 0.03514169692993164, 0.03519379043579102, 0.03507820892333984, 0.03547846221923828, 0.037443584442138675, 0.03530956649780274, 0.03505766296386719, 0.03501878356933594, 0.03537916946411133, 0.03894681549072266, 0.03853414535522461, 0.037720062255859374, 0.037379070281982424, 0.03753574371337891, 0.03740467071533203, 0.03780915069580078, 0.037351425170898435, 0.03688550567626953, 0.037321727752685545, 0.03734527969360352, 0.03752243041992188, 0.037989376068115234, 0.037684223175048825, 0.037501953125, 0.037466110229492186, 0.03762073516845703, 0.03746815872192383, 0.03742924880981445, 0.03751628875732422, 0.03755417633056641, 0.03743334579467773, 0.037443584442138675, 0.0374835205078125, 0.03751833724975586, 0.03747225570678711, 0.03808358383178711, 0.03794124984741211, 0.03768012619018555, 0.03758796691894531, 0.03723571014404297, 0.037556224822998044, 0.03790950393676758, 0.037510143280029294, 0.03750092697143555, 0.036918270111083985, 0.03747532653808594, 0.03751424026489258, 0.037318656921386716, 0.03539865493774414, 0.03698995208740234, 0.03739136123657227, 0.03741798400878906, 0.03800985717773438, 0.0820674591064453, 0.037531646728515625, 0.037582847595214845, 0.03745894241333008, 0.037410816192626956, 0.03758694458007812, 0.03754905700683594, 0.03766988754272461, 0.03741798400878906, 0.037353473663330077, 0.03511500930786133, 0.03501260757446289, 0.03656604766845703, 0.03767907333374024, 0.03806003189086914, 0.03765350341796875, 0.03746201705932617, 0.037372928619384765, 0.037645313262939455, 0.03757363128662109, 0.03775590515136719, 0.037596160888671876, 0.03761459350585938, 0.037647361755371096, 0.03745587158203125, 0.03750604629516602, 0.03783987045288086, 0.03752243041992188, 0.03748659133911133, 0.03752140808105469, 0.037310462951660156, 0.03763814544677734, 0.03776716613769531, 0.037480480194091795, 0.03736880111694336, 0.03778867340087891, 0.0375623664855957, 0.03755212783813477, 0.03778355026245117, 0.03767295837402344, 0.037493759155273435, 0.03794124984741211, 0.037945343017578126, 0.037596160888671876, 0.03759513473510742, 0.03751935958862305, 0.037720062255859374, 0.037932064056396486, 0.037492702484130856, 0.0371486701965332, 0.037525505065917966, 0.03757673645019531, 0.037504993438720706, 0.03755929565429687, 0.03767910385131836, 0.03771187210083008, 0.03734527969360352, 0.03750707244873047, 0.03745792007446289, 0.03769139099121094, 0.03738521575927734, 0.037425151824951174, 0.03742617416381836, 0.08117350769042969, 0.03523379135131836, 0.03526860809326172, 0.03732070541381836, 0.03737702560424805, 0.03731558227539063, 0.03731763076782227, 0.03765862274169922, 0.03743231964111328, 0.03763916778564453, 0.03780198287963867, 0.03775692749023438, 0.03758182525634766, 0.0374128646850586, 0.03747532653808594, 0.03698175811767578, 0.03743436813354492, 0.0374835205078125, 0.03753779220581055, 0.03762073516845703, 0.03746713638305664, 0.03736576080322265, 0.037541889190673826, 0.03761459350585938, 0.038004737854003906, 0.03746406555175781, 0.037700607299804685, 0.03759718322753906, 0.03748659133911133, 0.03741388702392578, 0.03773952102661133, 0.037894142150878905, 0.0375623664855957, 0.037405696868896485, 0.0374466552734375, 0.03768832015991211, 0.03774566268920899, 0.037531646728515625, 0.03748454284667969, 0.03785420989990235, 0.03759308624267578, 0.03734732818603516, 0.03733299255371094, 0.037392383575439454, 0.03764223861694336, 0.0374917106628418, 0.037566463470458986, 0.03750809478759766, 0.03754291152954101, 0.03730636978149414, 0.037351425170898435, 0.03747840118408203, 0.03756851196289063, 0.03767705535888672, 0.03747225570678711, 0.03764022445678711, 0.03804771041870117, 0.03740467071533203, 0.03745075225830078, 0.03749785614013672, 0.03757875061035156, 0.037477375030517575, 0.03769036865234375]",tokens/s,26.813684859080215,,,main,False,False,,,,,,,,,,,,,,,,,,,,,,, bfloat16-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,internlm/internlm2-20b,internlm/internlm2-20b,cuda,0,42,,,True,,,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",internlm2,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,i,i,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/i/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669474b3-5456d6ca6c7db421315c927d;bfe57325-826e-4cd1-8410-27a5335b8ba7) Repository Not Found for url: https://huggingface.co/i/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: i is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Deci/DeciLM-7B,Deci/DeciLM-7B,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1267.232768,14990.966784,0.0,14344.51968,14290.684928,s,10,0.811465950012207,0.0811465950012207,0.00033896753053419624,0.08105659103393555,0.08131665267944335,0.08171917457580566,0.08204119209289551,"[0.08212169647216797, 0.08105721282958984, 0.08108946990966796, 0.08122720336914062, 0.08095852661132813, 0.08105596923828125, 0.08086233520507813, 0.08096918487548828, 0.08113394927978515, 0.08099040222167969]",tokens/s,3154.784251836432,kWh,9.599776731597052e-07,5.260231838820478e-07,6.137056880967739e-06,7.623057738009492e-06,tokens/kWh,33582324.67839682,MB,1267.232768,14990.966784,0.0,14344.51968,14290.687488,s,10,22.714950195312497,2.2714950195312498,0.0030080022681676934,2.27052197265625,2.2729391357421873,2.276495080566406,2.2793398364257813,"[2.280051025390625, 2.270586181640625, 2.268709228515625, 2.269547607421875, 2.27006640625, 2.27214892578125, 2.270457763671875, 2.271254150390625, 2.270385009765625, 2.271743896484375]",tokens/s,27.73503769909247,kWh,2.6764263543817744e-05,1.4667705739504467e-05,0.00011411962265763214,0.00015555159194095434,tokens/kWh,405010.31981668243,,s,629,23.40739488601686,0.03721366436568656,0.009690543906175774,0.03601919937133789,0.036112589263916015,0.036279090881347655,0.11756068969726563,"[0.036759552001953126, 0.03680255889892578, 0.036719615936279294, 0.03669811248779297, 0.036741119384765625, 0.03667763137817383, 0.036732959747314456, 0.03679536056518555, 0.0367564811706543, 0.036531200408935545, 0.03636326217651367, 0.03634995269775391, 0.036501502990722655, 0.03633356857299805, 0.03621376037597656, 0.036040702819824216, 0.03599359893798828, 0.03601510238647461, 0.036016128540039063, 0.03603968048095703, 0.03603251266479492, 0.03602022552490235, 0.03604582214355469, 0.036040702819824216, 0.03601715087890625, 0.036021247863769534, 0.03602435302734375, 0.036026336669921874, 0.036021247863769534, 0.03604582214355469, 0.03608063888549805, 0.036018177032470705, 0.036004863739013675, 0.03604275131225586, 0.036021247863769534, 0.036063232421875, 0.036034561157226565, 0.0360263671875, 0.036127742767333985, 0.03618304061889648, 0.036036609649658206, 0.03602841567993164, 0.03601100921630859, 0.036016128540039063, 0.03602739334106445, 0.03601715087890625, 0.03599155044555664, 0.03617996978759765, 0.03601100921630859, 0.03603353500366211, 0.036004863739013675, 0.03605401611328125, 0.03607244873046875, 0.035999744415283204, 0.03599769592285156, 0.03602841567993164, 0.03604582214355469, 0.03614003372192383, 0.03615846252441406, 0.036033599853515626, 0.03599148941040039, 0.035999744415283204, 0.11762073516845703, 0.03602739334106445, 0.03601715087890625, 0.0360263671875, 0.03599564743041992, 0.036017215728759766, 0.03605401611328125, 0.03616864013671875, 0.03600998306274414, 0.03599359893798828, 0.036004863739013675, 0.03600588989257812, 0.03600076675415039, 0.03597312164306641, 0.03600076675415039, 0.03599871826171875, 0.0360263671875, 0.0360263671875, 0.03602534484863281, 0.036013057708740234, 0.03599052810668945, 0.03602534484863281, 0.03614822387695313, 0.03618304061889648, 0.036057086944580076, 0.036016128540039063, 0.036029441833496094, 0.03601408004760742, 0.03613699340820312, 0.0360212173461914, 0.03603558349609375, 0.03601715087890625, 0.03605913543701172, 0.03602431869506836, 0.03600588989257812, 0.03603968048095703, 0.03605606460571289, 0.03608473587036133, 0.03599564743041992, 0.03604172897338867, 0.03603251266479492, 0.03602329635620117, 0.03600588989257812, 0.036037631988525394, 0.036034561157226565, 0.0360263671875, 0.03602329635620117, 0.03601203155517578, 0.03602329635620117, 0.03599769592285156, 0.03600998306274414, 0.03600998306274414, 0.036017215728759766, 0.03600582504272461, 0.036073471069335936, 0.03602534484863281, 0.03605606460571289, 0.036063232421875, 0.03606016159057617, 0.036063232421875, 0.03604582214355469, 0.036029441833496094, 0.03601408004760742, 0.11760537719726563, 0.03597926330566406, 0.03597721481323242, 0.03600896072387695, 0.035963905334472655, 0.035969024658203126, 0.03595673751831055, 0.03597926330566406, 0.03595980834960937, 0.03598950576782227, 0.03601510238647461, 0.036013057708740234, 0.03600281524658203, 0.03596083068847656, 0.03596083068847656, 0.03598233413696289, 0.03599257659912109, 0.03599769592285156, 0.03600281524658203, 0.03606937789916992, 0.035980289459228515, 0.03600896072387695, 0.03597824096679687, 0.03597926330566406, 0.03597619247436523, 0.0359813117980957, 0.035996673583984375, 0.03599769592285156, 0.0359741439819336, 0.03598342514038086, 0.036020160675048825, 0.03599769592285156, 0.03598233413696289, 0.03605811309814453, 0.03607244873046875, 0.035996673583984375, 0.0360263671875, 0.036018177032470705, 0.03599257659912109, 0.036013057708740234, 0.035988479614257815, 0.035975166320800785, 0.03599564743041992, 0.03600076675415039, 0.0360079345703125, 0.03601919937133789, 0.03602841567993164, 0.03606835174560547, 0.036004863739013675, 0.03600998306274414, 0.036001792907714845, 0.03600691223144531, 0.03599462509155273, 0.035985408782958986, 0.035972095489501955, 0.035999744415283204, 0.03597107315063477, 0.03621376037597656, 0.03603968048095703, 0.036034561157226565, 0.036029441833496094, 0.036034561157226565, 0.03602534484863281, 0.11777228546142578, 0.03597107315063477, 0.0359813117980957, 0.03597619247436523, 0.035999744415283204, 0.03597312164306641, 0.03610419082641601, 0.036117504119873044, 0.036018177032470705, 0.03603558349609375, 0.035999744415283204, 0.0359813117980957, 0.03599462509155273, 0.03597926330566406, 0.03598950576782227, 0.036040702819824216, 0.03599052810668945, 0.03597926330566406, 0.03599871826171875, 0.035996673583984375, 0.03599462509155273, 0.03602329635620117, 0.035969024658203126, 0.036040702819824216, 0.03604172897338867, 0.03599257659912109, 0.03601715087890625, 0.03597619247436523, 0.03600998306274414, 0.03600281524658203, 0.03598438262939453, 0.036016128540039063, 0.03599462509155273, 0.03600588989257812, 0.03604991912841797, 0.03601510238647461, 0.03603257751464844, 0.036046783447265626, 0.03603251266479492, 0.036016128540039063, 0.03602022552490235, 0.03601510238647461, 0.0360447998046875, 0.03600896072387695, 0.03605913543701172, 0.03603459167480469, 0.03612771224975586, 0.036162559509277346, 0.03601510238647461, 0.03598441696166992, 0.036006881713867185, 0.035996673583984375, 0.03599052810668945, 0.036018177032470705, 0.03602329635620117, 0.036004863739013675, 0.03600998306274414, 0.03599564743041992, 0.03598438262939453, 0.036034561157226565, 0.03602431869506836, 0.03603251266479492, 0.03604582214355469, 0.11753266906738281, 0.03599462509155273, 0.03601715087890625, 0.035975166320800785, 0.035969024658203126, 0.03601408004760742, 0.03598745727539063, 0.036004863739013675, 0.03601100921630859, 0.03600588989257812, 0.03640627288818359, 0.03599468612670898, 0.03601504135131836, 0.035996673583984375, 0.036037631988525394, 0.03601715087890625, 0.03602534484863281, 0.036039745330810544, 0.036022209167480466, 0.03606528091430664, 0.036029441833496094, 0.03604377746582031, 0.03599980926513672, 0.03600377655029297, 0.03599359893798828, 0.03605913543701172, 0.036109310150146484, 0.03609804916381836, 0.03605913543701172, 0.03600796890258789, 0.03603043365478516, 0.03600691223144531, 0.036016128540039063, 0.036021247863769534, 0.03602739334106445, 0.03602329635620117, 0.03602739334106445, 0.036029441833496094, 0.03600076675415039, 0.0360079345703125, 0.03599769592285156, 0.03600998306274414, 0.03599257659912109, 0.035996673583984375, 0.035999744415283204, 0.03603251266479492, 0.036034561157226565, 0.03609190368652344, 0.03599871826171875, 0.03602022552490235, 0.03602534484863281, 0.03600998306274414, 0.036004863739013675, 0.03600691223144531, 0.03599462509155273, 0.03604582214355469, 0.035988479614257815, 0.03603046417236328, 0.03601715087890625, 0.035999744415283204, 0.03599462509155273, 0.03602227020263672, 0.035996673583984375, 0.11761971282958984, 0.03594652938842773, 0.035944416046142576, 0.03596492767333984, 0.03599564743041992, 0.03599871826171875, 0.0359628791809082, 0.03597824096679687, 0.03604889678955078, 0.03630284881591797, 0.036141056060791016, 0.035975166320800785, 0.03597107315063477, 0.035980289459228515, 0.0359741439819336, 0.036001792907714845, 0.03599769592285156, 0.03602227020263672, 0.036067329406738284, 0.03599769592285156, 0.03601510238647461, 0.036021247863769534, 0.035972095489501955, 0.036245536804199216, 0.037525470733642576, 0.03607551956176758, 0.0360263671875, 0.0360079345703125, 0.03601510238647461, 0.036018177032470705, 0.036004863739013675, 0.03604787063598633, 0.03602534484863281, 0.036036609649658206, 0.036029441833496094, 0.036021247863769534, 0.03602022552490235, 0.03604377746582031, 0.03616153717041016, 0.0362158088684082, 0.03603353500366211, 0.03601919937133789, 0.03611238479614258, 0.03601919937133789, 0.03605196762084961, 0.036157440185546875, 0.03601408004760742, 0.03601203155517578, 0.03602022552490235, 0.0360079345703125, 0.036013057708740234, 0.03600896072387695, 0.03600691223144531, 0.03601919937133789, 0.036013057708740234, 0.0360263671875, 0.036034561157226565, 0.0360079345703125, 0.03600384140014649, 0.03605299377441406, 0.036132865905761716, 0.03605811309814453, 0.03600588989257812, 0.11749683380126953, 0.036016128540039063, 0.03600281524658203, 0.03600588989257812, 0.03598745727539063, 0.035963905334472655, 0.035996673583984375, 0.03603046417236328, 0.03602534484863281, 0.0360079345703125, 0.03604582214355469, 0.03600076675415039, 0.03603558349609375, 0.035983360290527344, 0.03604172897338867, 0.036013057708740234, 0.036040702819824216, 0.03604377746582031, 0.036421630859375, 0.036055038452148434, 0.03602739334106445, 0.03603251266479492, 0.03599359893798828, 0.03607654571533203, 0.03599052810668945, 0.036036609649658206, 0.03602022552490235, 0.03603968048095703, 0.03603251266479492, 0.03600998306274414, 0.03602534484863281, 0.03599871826171875, 0.03603251266479492, 0.03601408004760742, 0.036004863739013675, 0.03602534484863281, 0.03605811309814453, 0.03601100921630859, 0.036030529022216796, 0.036049854278564455, 0.035988479614257815, 0.03599465560913086, 0.035998687744140626, 0.036055038452148434, 0.03599769592285156, 0.03601919937133789, 0.036004863739013675, 0.03601408004760742, 0.03607756805419922, 0.03607449722290039, 0.03618201446533203, 0.03605196762084961, 0.03601715087890625, 0.0360263671875, 0.035996673583984375, 0.036018177032470705, 0.036004863739013675, 0.03600691223144531, 0.03603251266479492, 0.03601715087890625, 0.0360079345703125, 0.03601919937133789, 0.03602534484863281, 0.11757158660888672, 0.03601408004760742, 0.035988479614257815, 0.035985408782958986, 0.03598745727539063, 0.036038688659667965, 0.035992607116699216, 0.03598534393310547, 0.035999744415283204, 0.03603251266479492, 0.03641139221191406, 0.03601919937133789, 0.03601718521118164, 0.03600073623657227, 0.03601203155517578, 0.03601919937133789, 0.03606220626831055, 0.0360079345703125, 0.036002849578857424, 0.03605910491943359, 0.036101119995117184, 0.03626291275024414, 0.036141056060791016, 0.036055038452148434, 0.03602329635620117, 0.036034561157226565, 0.03611040115356445, 0.03604064178466797, 0.03602227020263672, 0.03602227020263672, 0.035988479614257815, 0.036029441833496094, 0.03599564743041992, 0.036018177032470705, 0.03600281524658203, 0.036004863739013675, 0.0360447998046875, 0.03604377746582031, 0.036036609649658206, 0.03606118392944336, 0.036280319213867186, 0.036016128540039063, 0.03603046417236328, 0.03600076675415039, 0.03600998306274414, 0.03600588989257812, 0.03601206588745117, 0.036047840118408205, 0.03600896072387695, 0.035988479614257815, 0.03604275131225586, 0.03602227020263672, 0.03602739334106445, 0.03607551956176758, 0.036087806701660154, 0.03603148651123047, 0.03601919937133789, 0.036004863739013675, 0.03600691223144531, 0.03600998306274414, 0.0360079345703125, 0.03603558349609375, 0.03611238479614258, 0.11788492584228516, 0.03599769592285156, 0.03599769592285156, 0.035985408782958986, 0.03596799850463867, 0.035996673583984375, 0.036013057708740234, 0.03598233413696289, 0.03600384140014649, 0.03603968048095703, 0.03604275131225586, 0.03606528091430664, 0.03604889678955078, 0.035972095489501955, 0.03598438262939453, 0.03600592041015625, 0.03601712036132813, 0.03605606460571289, 0.036040702819824216, 0.0360263671875, 0.036001792907714845, 0.03600281524658203, 0.03602227020263672, 0.03604787063598633, 0.036067329406738284, 0.03602534484863281, 0.036446208953857424, 0.03601203155517578, 0.03603558349609375, 0.036018177032470705, 0.03605196762084961, 0.03606630325317383, 0.03603251266479492, 0.03604787063598633, 0.03602739334106445, 0.03601100921630859, 0.036071456909179685, 0.03603148651123047, 0.036020191192626956, 0.03602022552490235, 0.03598643112182617, 0.03600384140014649, 0.03602227020263672, 0.036109310150146484, 0.036101119995117184, 0.03603968048095703, 0.03601510238647461, 0.03597926330566406, 0.03601919937133789, 0.03602739334106445, 0.03600281524658203, 0.03601203155517578, 0.036038654327392575, 0.036029441833496094, 0.03600691223144531, 0.03601715087890625, 0.03599462509155273, 0.03602739334106445, 0.036016128540039063, 0.036018177032470705, 0.03601100921630859, 0.03601408004760742, 0.03600281524658203, 0.11763712310791016, 0.036029441833496094, 0.03598950576782227, 0.03602739334106445, 0.03609088134765625, 0.03600896072387695, 0.03599462509155273, 0.03604991912841797, 0.036055038452148434, 0.036021247863769534, 0.03599462509155273, 0.03599462509155273, 0.03597107315063477, 0.03599871826171875, 0.03600998306274414, 0.03599769592285156, 0.03597926330566406, 0.03599769592285156, 0.03605401611328125, 0.03601100921630859, 0.03604787063598633, 0.036119552612304685, 0.03600076675415039, 0.03600076675415039, 0.036016128540039063, 0.036176929473876955, 0.03643081665039063, 0.036029441833496094, 0.03602431869506836, 0.03602431869506836, 0.03603148651123047, 0.03601510238647461, 0.036039745330810544, 0.03605088043212891, 0.03610726547241211, 0.03614822387695313, 0.03603251266479492, 0.03602022552490235, 0.03602534484863281, 0.036016128540039063, 0.03616358566284179, 0.03629260635375976, 0.03627724838256836, 0.03600998306274414, 0.036016128540039063, 0.03614310455322266, 0.03600384140014649, 0.036001792907714845, 0.03600384140014649, 0.03602947235107422, 0.036006881713867185, 0.03600896072387695, 0.03597312164306641, 0.03599564743041992, 0.036071422576904294, 0.03624038314819336, 0.03611340713500977, 0.03603046417236328, 0.03600896072387695, 0.03600281524658203, 0.03600076675415039, 0.03604991912841797, 0.03604991912841797]",tokens/s,26.8718498176725,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,stabilityai/stablelm-3b-4e1t,stabilityai/stablelm-3b-4e1t,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1290.440704,6715.604992,0.0,6069.157888,5767.730176,s,10,0.3468295364379883,0.03468295364379882,0.0003755381045782764,0.03456620788574219,0.03472768783569336,0.03526627540588379,0.03569714546203614,"[0.03580486297607422, 0.03456451034545899, 0.03452537536621094, 0.03455408096313477, 0.03449792098999024, 0.03458457565307617, 0.03460416030883789, 0.03451814270019531, 0.03456790542602539, 0.03460800170898438]",tokens/s,7381.147598591904,kWh,4.104567996375723e-07,2.2491133311935603e-07,2.441110170479238e-06,3.0764783032361666e-06,tokens/kWh,83212028.41921948,MB,1290.440704,6736.576512,0.0,6090.129408,5863.115264,s,10,16.423444946289063,1.6423444946289063,0.017755620258921256,1.6376460571289062,1.655046875,1.673444580078125,1.688162744140625,"[1.69184228515625, 1.63790625, 1.62473388671875, 1.6322568359375, 1.633239501953125, 1.6373858642578125, 1.63308544921875, 1.65095849609375, 1.641189697265625, 1.6408466796875]",tokens/s,38.35979613658039,kWh,1.933375718543467e-05,1.0595080320960428e-05,6.129932971071992e-05,9.122816721711502e-05,tokens/kWh,690576.1884930291,,s,629,16.706031612396245,0.026559668700153003,0.004166775535262216,0.02595123291015625,0.02671966781616211,0.02726277122497559,0.06090154006958008,"[0.02773811149597168, 0.027708415985107423, 0.02790399932861328, 0.0273438720703125, 0.027987968444824218, 0.027274240493774415, 0.02797260856628418, 0.028047359466552735, 0.027634687423706054, 0.026824703216552736, 0.02800230407714844, 0.02777907180786133, 0.026849279403686522, 0.027019264221191407, 0.026925056457519532, 0.02758963203430176, 0.027794431686401368, 0.027493375778198242, 0.027870208740234374, 0.027648000717163085, 0.027245567321777343, 0.02693939208984375, 0.027018239974975586, 0.025985023498535157, 0.026719232559204102, 0.02650726318359375, 0.02588467216491699, 0.025935871124267578, 0.02740121650695801, 0.027210784912109376, 0.027007999420166014, 0.026876895904541016, 0.026994688034057617, 0.02692403221130371, 0.02654515266418457, 0.02578950309753418, 0.02672140884399414, 0.026543935775756835, 0.026422271728515623, 0.025997312545776367, 0.026000415802001953, 0.02584163284301758, 0.025983999252319336, 0.02612838363647461, 0.026863616943359377, 0.026253311157226563, 0.02615500831604004, 0.026065919876098635, 0.026381311416625978, 0.027075584411621095, 0.026570751190185548, 0.026589183807373046, 0.027029504776000978, 0.026259456634521484, 0.026047487258911133, 0.026862592697143556, 0.026804224014282226, 0.026015743255615235, 0.026217472076416014, 0.026000383377075196, 0.025948160171508788, 0.025977855682373048, 0.060902400970458986, 0.025903104782104492, 0.025907199859619142, 0.025935871124267578, 0.026068992614746093, 0.026209280014038085, 0.02590105628967285, 0.02607923126220703, 0.02597068786621094, 0.025840639114379883, 0.025985023498535157, 0.026055679321289063, 0.025844736099243162, 0.026017791748046876, 0.025971712112426756, 0.026036224365234374, 0.02595840072631836, 0.026019840240478515, 0.025873407363891602, 0.026058752059936522, 0.025923583984375, 0.02593280029296875, 0.026006528854370117, 0.026130495071411134, 0.025897920608520506, 0.025899007797241212, 0.02589593505859375, 0.02590105628967285, 0.02593075180053711, 0.02594508743286133, 0.02590924835205078, 0.02647039985656738, 0.02629734420776367, 0.026027008056640624, 0.025975807189941406, 0.026039295196533203, 0.02593699264526367, 0.02590096092224121, 0.026020864486694335, 0.026027008056640624, 0.026013696670532226, 0.026015743255615235, 0.026064895629882814, 0.02609049606323242, 0.026008575439453126, 0.02598297691345215, 0.026192895889282225, 0.026061824798583984, 0.02594611167907715, 0.02676531219482422, 0.026205184936523438, 0.02598297691345215, 0.025967615127563477, 0.026015743255615235, 0.025968639373779297, 0.026046464920043946, 0.025944063186645508, 0.025788415908813478, 0.02593484878540039, 0.025636863708496094, 0.025825279235839844, 0.02570444869995117, 0.02572697639465332, 0.06066995239257812, 0.02569523239135742, 0.025530399322509764, 0.025489376068115233, 0.025554943084716796, 0.02552729606628418, 0.02553548812866211, 0.025847808837890625, 0.02555084800720215, 0.02568294334411621, 0.025664512634277343, 0.0256615047454834, 0.02545043182373047, 0.02548531150817871, 0.025428096771240236, 0.02555891227722168, 0.025480192184448244, 0.025689088821411132, 0.025617408752441406, 0.02546892738342285, 0.02547302436828613, 0.02585203170776367, 0.025625663757324217, 0.025685823440551758, 0.025556991577148438, 0.02552729606628418, 0.025531391143798828, 0.025822208404541015, 0.025426944732666015, 0.02572800064086914, 0.025952255249023438, 0.025969663619995118, 0.025870336532592773, 0.0259051513671875, 0.025848831176757812, 0.02591948890686035, 0.025869312286376952, 0.025829376220703124, 0.02590105628967285, 0.025891839981079103, 0.025797632217407225, 0.02592153549194336, 0.02587238311767578, 0.025915391921997072, 0.025854976654052734, 0.025869312286376952, 0.025809919357299805, 0.026016767501831056, 0.025829376220703124, 0.025883647918701173, 0.025915391921997072, 0.025825279235839844, 0.02592460823059082, 0.02594304084777832, 0.025880672454833983, 0.02594396781921387, 0.026627071380615236, 0.026254335403442384, 0.02594099235534668, 0.026689535140991212, 0.026027040481567384, 0.025885663986206054, 0.025882623672485353, 0.0610621452331543, 0.025839616775512695, 0.025971712112426756, 0.02593894386291504, 0.025865215301513672, 0.026451967239379884, 0.026521600723266602, 0.025985023498535157, 0.025999359130859375, 0.02657792091369629, 0.025977855682373048, 0.025832447052001953, 0.02588979148864746, 0.025867263793945314, 0.026012672424316406, 0.025927711486816406, 0.025851871490478514, 0.02590822410583496, 0.025826303482055665, 0.025779199600219727, 0.025709568023681642, 0.025542688369750977, 0.025422815322875977, 0.02593075180053711, 0.025819135665893556, 0.025850879669189454, 0.025955327987670897, 0.02589695930480957, 0.02583660888671875, 0.025885631561279297, 0.025887775421142577, 0.025952224731445313, 0.026027008056640624, 0.02573619270324707, 0.0257392635345459, 0.025633792877197265, 0.02546790313720703, 0.02555904006958008, 0.025988096237182616, 0.02591846466064453, 0.02589388847351074, 0.025964544296264647, 0.025867263793945314, 0.02594099235534668, 0.0259102725982666, 0.025969663619995118, 0.025853952407836913, 0.026019840240478515, 0.025838592529296874, 0.025874431610107423, 0.025939968109130858, 0.02591744041442871, 0.025857023239135742, 0.025948160171508788, 0.02587753677368164, 0.025900096893310548, 0.025982912063598634, 0.025879520416259766, 0.02592870330810547, 0.02589286422729492, 0.025907199859619142, 0.025860095977783205, 0.025968639373779297, 0.06089932632446289, 0.025798656463623046, 0.0267775993347168, 0.026195968627929687, 0.02592972755432129, 0.02595430374145508, 0.02590822410583496, 0.0259368953704834, 0.025786367416381836, 0.025845760345458983, 0.025883647918701173, 0.02593791961669922, 0.025838592529296874, 0.025903104782104492, 0.027077632904052733, 0.02731929588317871, 0.025998336791992187, 0.025878528594970703, 0.025777183532714843, 0.025926624298095703, 0.02590617561340332, 0.025785343170166015, 0.0259420166015625, 0.026010623931884767, 0.025772031784057618, 0.025867263793945314, 0.02593382453918457, 0.025881599426269532, 0.025879552841186523, 0.025907199859619142, 0.025796607971191408, 0.02596659278869629, 0.025869312286376952, 0.025866239547729493, 0.025869312286376952, 0.025883647918701173, 0.025808895111083984, 0.025850879669189454, 0.02591948890686035, 0.02596147155761719, 0.025824256896972656, 0.02590822410583496, 0.02591948890686035, 0.0259368953704834, 0.02596147155761719, 0.025960447311401368, 0.025995264053344725, 0.026013696670532226, 0.02559903907775879, 0.025696191787719726, 0.02628096008300781, 0.02589593505859375, 0.02612326431274414, 0.025935871124267578, 0.025875455856323244, 0.0257126407623291, 0.025683967590332032, 0.025358335494995117, 0.025477312088012696, 0.025576255798339845, 0.02551091194152832, 0.025610240936279297, 0.02567078399658203, 0.060916606903076174, 0.025568256378173827, 0.025819135665893556, 0.02557542419433594, 0.025630720138549806, 0.025561088562011718, 0.02551913642883301, 0.025688032150268554, 0.025475072860717773, 0.025641984939575195, 0.025667776107788087, 0.02571468734741211, 0.0276693115234375, 0.02693222427368164, 0.02638643264770508, 0.026235904693603516, 0.025766912460327147, 0.025695264816284178, 0.025629663467407228, 0.025614336013793947, 0.025827327728271485, 0.025807872772216797, 0.025886720657348632, 0.02592153549194336, 0.025989120483398437, 0.026007551193237305, 0.025960447311401368, 0.025860095977783205, 0.025785343170166015, 0.025987071990966795, 0.02609459114074707, 0.026129407882690428, 0.025935871124267578, 0.025911296844482422, 0.02590105628967285, 0.026043392181396483, 0.025859071731567384, 0.025971712112426756, 0.02592460823059082, 0.026003456115722655, 0.026278911590576173, 0.025960447311401368, 0.025993215560913087, 0.025969663619995118, 0.025859071731567384, 0.02594713592529297, 0.025996288299560546, 0.026068992614746093, 0.025863168716430664, 0.025985023498535157, 0.026010623931884767, 0.026207231521606447, 0.02597478485107422, 0.026020864486694335, 0.025977855682373048, 0.02594713592529297, 0.02592767906188965, 0.026033151626586915, 0.026688512802124024, 0.02738483238220215, 0.02636390495300293, 0.026008607864379883, 0.02600547218322754, 0.06099251174926758, 0.02591948890686035, 0.02595123291015625, 0.026001407623291017, 0.02594099235534668, 0.026032127380371094, 0.02589593505859375, 0.026055679321289063, 0.026001407623291017, 0.026049535751342775, 0.0259051513671875, 0.025977855682373048, 0.02588467216491699, 0.02593075180053711, 0.026048511505126954, 0.026023935317993165, 0.026004480361938476, 0.02590412712097168, 0.025803775787353517, 0.025972736358642577, 0.026062847137451172, 0.02592767906188965, 0.02593075180053711, 0.02609152030944824, 0.02596147155761719, 0.025879552841186523, 0.02596659278869629, 0.025835519790649415, 0.025869312286376952, 0.02550169563293457, 0.025483392715454103, 0.025745311737060548, 0.025943008422851563, 0.025923583984375, 0.025789440155029295, 0.025965568542480468, 0.0259051513671875, 0.025967615127563477, 0.026027008056640624, 0.026003456115722655, 0.02596147155761719, 0.025767936706542968, 0.02569215965270996, 0.025591808319091795, 0.02555392074584961, 0.025487360000610353, 0.02592972755432129, 0.025993215560913087, 0.025890815734863282, 0.025831424713134765, 0.026047487258911133, 0.025811967849731447, 0.025988096237182616, 0.025986047744750978, 0.025964544296264647, 0.026033151626586915, 0.026054655075073242, 0.02592255973815918, 0.025964544296264647, 0.026012672424316406, 0.025955327987670897, 0.026028032302856444, 0.02616524887084961, 0.06101299285888672, 0.02632601547241211, 0.027045888900756834, 0.026658815383911134, 0.026006528854370117, 0.0261079044342041, 0.025971712112426756, 0.026068992614746093, 0.026025983810424806, 0.025838592529296874, 0.025931776046752928, 0.026060800552368164, 0.02589593505859375, 0.026015743255615235, 0.026000383377075196, 0.026002431869506838, 0.026002431869506838, 0.026211360931396484, 0.02637001609802246, 0.025976831436157227, 0.02590105628967285, 0.02590003204345703, 0.025903104782104492, 0.026076160430908202, 0.026639360427856446, 0.02592870330810547, 0.02594099235534668, 0.026655744552612305, 0.026052608489990234, 0.02596249580383301, 0.026129407882690428, 0.02728550338745117, 0.02632089614868164, 0.02592665672302246, 0.025939968109130858, 0.025996288299560546, 0.025997312545776367, 0.02596249580383301, 0.025973760604858398, 0.025995264053344725, 0.026508287429809572, 0.027397119522094726, 0.02700595283508301, 0.02607513618469238, 0.02592665672302246, 0.025879552841186523, 0.026640384674072266, 0.026043392181396483, 0.02594918441772461, 0.026608640670776368, 0.026925056457519532, 0.02618880081176758, 0.02597478485107422, 0.025992191314697266, 0.025873407363891602, 0.02591948890686035, 0.027047935485839843, 0.026670080184936523, 0.02650214385986328, 0.025886720657348632, 0.025915391921997072, 0.02590412712097168, 0.026659839630126952, 0.06112255859375, 0.025791488647460937, 0.025644031524658203, 0.025587711334228515, 0.02568191909790039, 0.025460832595825194, 0.02550681686401367, 0.02570537567138672, 0.025448448181152345, 0.02547715187072754, 0.025895904541015625, 0.02574847984313965, 0.025667583465576172, 0.02568704032897949, 0.026033151626586915, 0.02647859191894531, 0.026089471817016603, 0.026499168395996094, 0.026289056777954102, 0.025446399688720703, 0.025554943084716796, 0.025482240676879882, 0.02546892738342285, 0.026239999771118162, 0.026400768280029296, 0.02571059226989746, 0.025789440155029295, 0.025750528335571288, 0.025585664749145507, 0.026012672424316406, 0.027062271118164064, 0.026690624237060548, 0.025881599426269532, 0.025942975997924805, 0.02590617561340332, 0.025827327728271485, 0.026603519439697267, 0.026364927291870118, 0.026084352493286132, 0.02650111961364746, 0.02632294464111328, 0.025973760604858398, 0.026434560775756837, 0.026404863357543946, 0.025944063186645508, 0.025887744903564453, 0.02609868812561035, 0.026057727813720705, 0.026324031829833984, 0.0270314884185791, 0.026195968627929687, 0.02592563247680664, 0.02591334342956543, 0.02591948890686035, 0.025955327987670897, 0.02714931106567383, 0.02632601547241211, 0.025944063186645508, 0.026030080795288086, 0.02608742332458496, 0.02595430374145508, 0.026610687255859376, 0.027182079315185546, 0.060949504852294924, 0.025862144470214843, 0.026455039978027343, 0.02697216033935547, 0.026302463531494142, 0.025980928421020507, 0.02596659278869629, 0.02613248062133789, 0.02590924835205078, 0.026129407882690428, 0.026004480361938476, 0.02588569641113281, 0.025935871124267578, 0.026214399337768556, 0.026211360931396484, 0.026075103759765624, 0.02752204895019531, 0.026335231781005858, 0.025935871124267578, 0.02595737648010254, 0.026062847137451172, 0.02592767906188965, 0.0259368953704834, 0.026081279754638673, 0.025899007797241212, 0.02590003204345703, 0.026016767501831056, 0.025989120483398437, 0.025854976654052734, 0.025925664901733397, 0.025896928787231446, 0.02595840072631836, 0.026054655075073242, 0.026071039199829102, 0.025803775787353517, 0.025980928421020507, 0.02596147155761719, 0.025944063186645508, 0.025960447311401368, 0.026081279754638673, 0.025882623672485353, 0.026183679580688478, 0.02588979148864746, 0.02593894386291504, 0.02593894386291504, 0.0259102725982666, 0.025975807189941406, 0.026003456115722655, 0.025915391921997072, 0.02592767906188965, 0.025993215560913087, 0.02591334342956543, 0.02592563247680664, 0.025955327987670897, 0.02596147155761719, 0.026076160430908202, 0.025992191314697266, 0.025847808837890625, 0.0259368953704834, 0.026066944122314452, 0.02591334342956543, 0.02614784049987793, 0.026129407882690428]",tokens/s,37.65107205551247,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,huggyllama/llama-13b,huggyllama/llama-13b,cuda,0,42,,,True,,,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",llama,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-2.7b,,cuda,0,42,,,,,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,d35829e539df8480b726c647eeabf91e41eae047,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,,MB,1200.279552,6415.712256,0.0,5769.265152,5440.254976,s,10,0.3110248985290527,0.031102489852905275,0.0003445720948210344,0.031050463676452637,0.031341987228393556,0.031678098487854,0.031946987495422366,"[0.032014209747314455, 0.030605728149414063, 0.03100707244873047, 0.031057695388793945, 0.031043231964111326, 0.030977760314941406, 0.031107872009277344, 0.03106656074523926, 0.030877471923828125, 0.03126729583740234]",tokens/s,8230.852295450139,kWh,3.6442570910494557e-07,1.9966522265559222e-07,1.8872331337170214e-06,2.4513240654775594e-06,tokens/kWh,104433356.48896624,MB,1200.730112,6415.712256,0.0,5769.265152,5540.555776,s,10,11.756425415039065,1.1756425415039062,0.002954145858079468,1.1763389282226564,1.178477746582031,1.1789994323730468,1.1794167810058593,"[1.17836181640625, 1.175759765625, 1.1742996826171874, 1.1780562744140626, 1.1769180908203125, 1.1734361572265626, 1.1795211181640626, 1.169724609375, 1.1724141845703124, 1.1779337158203125]",tokens/s,53.587717163934116,kWh,1.3630998501920293e-05,7.467455192607025e-06,4.9380724547283436e-05,7.047917824181076e-05,tokens/kWh,893881.0237521492,,s,629,12.015698955535886,0.019102860024699346,0.003737121467162467,0.01865830421447754,0.018774629974365235,0.01886719970703125,0.0500291552734375,"[0.018505727767944336, 0.018288639068603514, 0.018354175567626953, 0.018950143814086915, 0.01867366409301758, 0.01866444778442383, 0.018685951232910156, 0.01860403251647949, 0.01860710334777832, 0.018677759170532226, 0.018619392395019533, 0.018534400939941405, 0.018778112411499022, 0.018634752273559572, 0.018804800033569335, 0.018711488723754884, 0.01862451171875, 0.01860095977783203, 0.018677759170532226, 0.018686975479125977, 0.01863987159729004, 0.01863270378112793, 0.018719743728637696, 0.018695167541503906, 0.0186562557220459, 0.018710527420043945, 0.01862144088745117, 0.01864806365966797, 0.018655231475830078, 0.018739231109619142, 0.018757600784301758, 0.018753536224365236, 0.018753536224365236, 0.01866444778442383, 0.018654207229614257, 0.018691072463989256, 0.018679807662963867, 0.018670591354370117, 0.01881907272338867, 0.01866649627685547, 0.01865727996826172, 0.018718719482421875, 0.01863680076599121, 0.018788383483886718, 0.018928607940673827, 0.018918399810791017, 0.018759679794311524, 0.018735103607177735, 0.018719743728637696, 0.018742271423339844, 0.018562047958374024, 0.018547712326049806, 0.01860505676269531, 0.018667520523071288, 0.018786304473876952, 0.018723840713500976, 0.018750463485717773, 0.018752511978149415, 0.018691072463989256, 0.018708480834960937, 0.018700288772583007, 0.018670591354370117, 0.04956774520874024, 0.01861631965637207, 0.01859891128540039, 0.01864499282836914, 0.018578432083129884, 0.01859071922302246, 0.018448383331298827, 0.018290719985961913, 0.018200544357299803, 0.018242559432983398, 0.01847500801086426, 0.018655231475830078, 0.01860710334777832, 0.018750463485717773, 0.01864396858215332, 0.018725887298583984, 0.018720767974853517, 0.01863167953491211, 0.018550783157348632, 0.01863884735107422, 0.01862758445739746, 0.018653215408325194, 0.018636768341064452, 0.018746368408203123, 0.018634752273559572, 0.01866035270690918, 0.018663423538208008, 0.01861734390258789, 0.01866444778442383, 0.018682880401611326, 0.01864089584350586, 0.018765823364257812, 0.018626560211181642, 0.018716672897338867, 0.018669567108154296, 0.0187064323425293, 0.018637823104858398, 0.018662399291992187, 0.01880166435241699, 0.0189040641784668, 0.01889587211608887, 0.018735103607177735, 0.0186746883392334, 0.018695167541503906, 0.018692096710205077, 0.01868083190917969, 0.018676736831665038, 0.018732032775878905, 0.018729984283447267, 0.01864192008972168, 0.01862553596496582, 0.018647039413452148, 0.01863270378112793, 0.018966527938842775, 0.018747392654418944, 0.01877299118041992, 0.018733055114746093, 0.018692096710205077, 0.01865727996826172, 0.018749439239501953, 0.018734079360961914, 0.018703359603881836, 0.018705408096313478, 0.05008998489379883, 0.018637855529785155, 0.018595808029174803, 0.01862144088745117, 0.018545663833618165, 0.01865830421447754, 0.01863577651977539, 0.018675712585449217, 0.018626560211181642, 0.018589696884155273, 0.018587648391723634, 0.01862553596496582, 0.018508832931518556, 0.01845347213745117, 0.01842278480529785, 0.018556928634643553, 0.018554880142211915, 0.018534400939941405, 0.018588672637939452, 0.018603008270263673, 0.018752511978149415, 0.018637823104858398, 0.01864297676086426, 0.018706399917602538, 0.0186562557220459, 0.018667520523071288, 0.01863065528869629, 0.01884569549560547, 0.018721792221069337, 0.01866547203063965, 0.018678783416748047, 0.018679807662963867, 0.01867263984680176, 0.018720767974853517, 0.018687999725341797, 0.01863270378112793, 0.01865318489074707, 0.018651136398315428, 0.018671615600585938, 0.018713600158691408, 0.01864806365966797, 0.018659328460693358, 0.01863167953491211, 0.01864499282836914, 0.018679840087890625, 0.018665439605712892, 0.01858355140686035, 0.01880268859863281, 0.01867366409301758, 0.018684928894042968, 0.018647039413452148, 0.01863577651977539, 0.01861529541015625, 0.018597888946533202, 0.018514944076538087, 0.018693119049072265, 0.018585599899291993, 0.01856716728210449, 0.018571296691894532, 0.018563039779663087, 0.018588672637939452, 0.018556928634643553, 0.018728960037231446, 0.05025075149536133, 0.018568191528320312, 0.018592767715454102, 0.01860403251647949, 0.018576383590698242, 0.018626560211181642, 0.01864499282836914, 0.01865830421447754, 0.01860710334777832, 0.018638912200927733, 0.018638784408569337, 0.019170303344726563, 0.01880678367614746, 0.0186296329498291, 0.01864396858215332, 0.018767871856689454, 0.018742271423339844, 0.018817024230957033, 0.01864499282836914, 0.018572288513183592, 0.01856716728210449, 0.01862246322631836, 0.018593791961669923, 0.018702335357666015, 0.01866755294799805, 0.01892348861694336, 0.018722816467285155, 0.018700351715087892, 0.018547647476196288, 0.01862553596496582, 0.01862348747253418, 0.0186746883392334, 0.018525184631347655, 0.01857535934448242, 0.018497535705566406, 0.018519039154052733, 0.018520063400268554, 0.018527231216430663, 0.01850060844421387, 0.019082239151000976, 0.01877299118041992, 0.01865727996826172, 0.018718719482421875, 0.01863987159729004, 0.018738176345825194, 0.018716672897338867, 0.018886655807495118, 0.018742271423339844, 0.01880166435241699, 0.01864806365966797, 0.01868083190917969, 0.018663423538208008, 0.018685951232910156, 0.018728960037231446, 0.018701311111450195, 0.018726911544799805, 0.018712575912475587, 0.018932735443115235, 0.019116031646728517, 0.01886207962036133, 0.018722816467285155, 0.01869004821777344, 0.01868185615539551, 0.050098175048828124, 0.01862041664123535, 0.01859686470031738, 0.018655231475830078, 0.018678783416748047, 0.018741247177124023, 0.01904640007019043, 0.018773088455200194, 0.018667423248291015, 0.01864806365966797, 0.018626623153686524, 0.01861625671386719, 0.01862348747253418, 0.018651136398315428, 0.01860915184020996, 0.018709503173828124, 0.018597888946533202, 0.01866444778442383, 0.01861529541015625, 0.01860095977783203, 0.01861529541015625, 0.0186296329498291, 0.01859382438659668, 0.01874224090576172, 0.01862348747253418, 0.01866854476928711, 0.01860915184020996, 0.018737152099609376, 0.01866854476928711, 0.01862860870361328, 0.0186429443359375, 0.01861529541015625, 0.018503679275512695, 0.018695167541503906, 0.01866035270690918, 0.01865216064453125, 0.018691072463989256, 0.01869004821777344, 0.018619392395019533, 0.018774015426635742, 0.01862041664123535, 0.01866035270690918, 0.01862758445739746, 0.01864908790588379, 0.018662399291992187, 0.018691072463989256, 0.01864806365966797, 0.018740224838256835, 0.018563072204589845, 0.018573312759399413, 0.018487295150756835, 0.018539520263671876, 0.018634752273559572, 0.01900748825073242, 0.018974720001220705, 0.01882931137084961, 0.018696191787719727, 0.018722816467285155, 0.018747392654418944, 0.018684928894042968, 0.01886412811279297, 0.01874844741821289, 0.01868796730041504, 0.050071617126464844, 0.018650047302246092, 0.01863577651977539, 0.01860095977783203, 0.018594816207885743, 0.018570240020751954, 0.018514944076538087, 0.01858355140686035, 0.018539520263671876, 0.018456575393676757, 0.018495487213134765, 0.018588672637939452, 0.01863065528869629, 0.018594816207885743, 0.018603008270263673, 0.018721792221069337, 0.018539520263671876, 0.01861222457885742, 0.018663423538208008, 0.018585599899291993, 0.018557952880859374, 0.018700288772583007, 0.01861734390258789, 0.018683904647827147, 0.018678783416748047, 0.018592767715454102, 0.018566144943237304, 0.018534400939941405, 0.018449472427368163, 0.01859174346923828, 0.018659263610839843, 0.01868083190917969, 0.01863065528869629, 0.0186562557220459, 0.018725887298583984, 0.018661376953125, 0.018654207229614257, 0.01861734390258789, 0.018661376953125, 0.01880678367614746, 0.01859891128540039, 0.018662399291992187, 0.018692096710205077, 0.01863987159729004, 0.01865830421447754, 0.018691072463989256, 0.01864908790588379, 0.01885081672668457, 0.018700288772583007, 0.01883340835571289, 0.01863372802734375, 0.01862144088745117, 0.018309120178222657, 0.018298976898193358, 0.018296735763549805, 0.018556928634643553, 0.018714624404907225, 0.018691072463989256, 0.018687999725341797, 0.018682880401611326, 0.0186746883392334, 0.01867366409301758, 0.018695167541503906, 0.05016064071655273, 0.01860710334777832, 0.01860505676269531, 0.018619392395019533, 0.018592767715454102, 0.018593791961669923, 0.018603008270263673, 0.018691072463989256, 0.018676736831665038, 0.01863577651977539, 0.018573312759399413, 0.01863577651977539, 0.018569215774536133, 0.018803712844848632, 0.018822175979614258, 0.019521503448486327, 0.01883443260192871, 0.01869824028015137, 0.0186429443359375, 0.01862758445739746, 0.018653247833251955, 0.018623424530029298, 0.01859071922302246, 0.018795520782470702, 0.01863680076599121, 0.018687999725341797, 0.018663488388061523, 0.018649023056030275, 0.018597888946533202, 0.018646015167236327, 0.01861324882507324, 0.018745344161987306, 0.0187064323425293, 0.018637823104858398, 0.018562047958374024, 0.018750463485717773, 0.020188159942626953, 0.01893791961669922, 0.018606016159057617, 0.019317760467529296, 0.018777088165283205, 0.01866854476928711, 0.018683904647827147, 0.01865830421447754, 0.018721792221069337, 0.0186562557220459, 0.01866035270690918, 0.018778112411499022, 0.01863987159729004, 0.018672672271728516, 0.018623455047607422, 0.01864396858215332, 0.018704383850097657, 0.018670591354370117, 0.01864806365966797, 0.01864908790588379, 0.01858252716064453, 0.018682880401611326, 0.018679807662963867, 0.01865727996826172, 0.018692096710205077, 0.018750463485717773, 0.01861734390258789, 0.050179073333740234, 0.01862758445739746, 0.018603008270263673, 0.0186060791015625, 0.01858355140686035, 0.018579456329345705, 0.01865011215209961, 0.01869824028015137, 0.01865011215209961, 0.018676736831665038, 0.01866649627685547, 0.018566144943237304, 0.01855897521972656, 0.018572288513183592, 0.018545663833618165, 0.01842995262145996, 0.018150400161743165, 0.018241535186767577, 0.018266111373901366, 0.01820159912109375, 0.018206720352172853, 0.018241535186767577, 0.018257919311523436, 0.018311168670654295, 0.0182794246673584, 0.018167808532714845, 0.018334720611572267, 0.018276351928710938, 0.018225151062011717, 0.018233343124389647, 0.0182794246673584, 0.018667520523071288, 0.018652191162109377, 0.018610143661499025, 0.018718719482421875, 0.01865216064453125, 0.01866963195800781, 0.018645952224731446, 0.01865318489074707, 0.018795520782470702, 0.018661376953125, 0.01865830421447754, 0.018721792221069337, 0.018647039413452148, 0.01861222457885742, 0.01867263984680176, 0.018685951232910156, 0.018784255981445314, 0.01869824028015137, 0.01864908790588379, 0.01865011215209961, 0.018717695236206054, 0.018687999725341797, 0.018761728286743166, 0.018710527420043945, 0.01878937530517578, 0.018687999725341797, 0.018697216033935548, 0.018681888580322267, 0.018683935165405275, 0.018692031860351562, 0.018675712585449217, 0.018690080642700196, 0.049919967651367185, 0.01864192008972168, 0.01858252716064453, 0.01863884735107422, 0.018581504821777343, 0.018629695892333983, 0.018670528411865235, 0.01862246322631836, 0.018268159866333008, 0.018332672119140626, 0.01829478454589844, 0.018284543991088868, 0.018281471252441405, 0.0182476806640625, 0.01829478454589844, 0.01836851119995117, 0.01861427116394043, 0.018682880401611326, 0.01861222457885742, 0.01862041664123535, 0.018626560211181642, 0.01862553596496582, 0.018692096710205077, 0.01882111930847168, 0.018695167541503906, 0.01869004821777344, 0.018578432083129884, 0.0186296329498291, 0.01864806365966797, 0.018691072463989256, 0.018647039413452148, 0.018744384765625, 0.018640832901000978, 0.01863884735107422, 0.018717695236206054, 0.018657312393188477, 0.0186777286529541, 0.018685951232910156, 0.018671615600585938, 0.018387968063354493, 0.018275327682495117, 0.018314239501953124, 0.018363391876220703, 0.018737152099609376, 0.01862553596496582, 0.018705408096313478, 0.01863884735107422, 0.018869247436523438, 0.018722816467285155, 0.01867263984680176, 0.01864089584350586, 0.01863680076599121, 0.01864396858215332, 0.018719743728637696, 0.01866035270690918, 0.01874127960205078, 0.01865622329711914, 0.018701311111450195, 0.018748416900634765, 0.018716672897338867, 0.018697216033935548, 0.018769920349121092, 0.018691072463989256, 0.05052726364135742, 0.018670591354370117, 0.01871971130371094, 0.01860403251647949, 0.018926591873168946, 0.018700288772583007, 0.018669567108154296, 0.018685951232910156, 0.0186060791015625, 0.01860095977783203, 0.01866547203063965, 0.01864908790588379, 0.01866649627685547, 0.01862144088745117, 0.018757631301879883, 0.018679807662963867, 0.01862451171875, 0.01866547203063965, 0.01866854476928711, 0.01863577651977539, 0.01860095977783203, 0.01860505676269531, 0.01866035270690918, 0.018703359603881836, 0.01863065528869629, 0.01862553596496582, 0.018677759170532226, 0.018613311767578126, 0.018619327545166015, 0.018646015167236327, 0.018696191787719727, 0.018758655548095703, 0.01861529541015625, 0.01866444778442383, 0.01863167953491211, 0.0186429443359375, 0.01869004821777344, 0.018779136657714843, 0.018763776779174804, 0.018874368667602538, 0.01861631965637207, 0.01861324882507324, 0.01865318489074707, 0.01867366409301758, 0.018774015426635742, 0.01864192008972168, 0.018685951232910156, 0.01871776008605957, 0.018664384841918947, 0.01866649627685547, 0.018637887954711913, 0.018663423538208008, 0.018923456192016602, 0.018753536224365236, 0.01876479911804199, 0.018787328720092773, 0.018677759170532226, 0.018787328720092773, 0.01880166435241699, 0.01876479911804199, 0.018700288772583007, 0.018700288772583007, 0.018739200592041014]",tokens/s,52.348182351073824,,,main,False,False,,,,,,,,,,,,,,,,,,,,,,, bfloat16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,microsoft/phi-1_5,microsoft/phi-1_5,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1262.206976,3876.061184,0.0,3229.61408,2982.448128,s,10,0.2535144634246826,0.025351446342468265,0.0002687257791714568,0.025277584075927735,0.025410530853271485,0.025779185676574706,0.026074109535217285,"[0.02614784049987793, 0.02527123260498047, 0.02522502326965332, 0.025283935546875, 0.025167327880859375, 0.025258176803588866, 0.025247615814208986, 0.025328607559204103, 0.02528838348388672, 0.025296319961547853]",tokens/s,10098.043186244315,kWh,2.999588905991771e-07,1.6432001539953655e-07,1.442334769208118e-06,1.906613675206832e-06,tokens/kWh,134269465.97990218,MB,1262.206976,3876.061184,0.0,3229.61408,3023.634432,s,10,13.590118530273438,1.3590118530273436,0.008322280197909628,1.3580879516601563,1.3684149536132812,1.3703542053222655,1.371905606689453,"[1.37229345703125, 1.3679840087890625, 1.3493690185546876, 1.3455333251953125, 1.3541610107421875, 1.3669649658203125, 1.36407861328125, 1.3610157470703126, 1.35516015625, 1.3535582275390625]",tokens/s,46.357211572261704,kWh,1.5409029815671675e-05,8.44397595864564e-06,3.971889260419252e-05,6.357189837850984e-05,tokens/kWh,991003.9122144075,,s,629,13.794023429870627,0.021930084944150406,0.003040407702214905,0.021638143539428712,0.021858712768554688,0.022060646438598633,0.046446875,"[0.023059455871582032, 0.023569408416748046, 0.022540288925170897, 0.022071296691894532, 0.021776384353637695, 0.021738496780395508, 0.022622207641601562, 0.021922815322875978, 0.021797887802124022, 0.021767168045043944, 0.02183782386779785, 0.02168934440612793, 0.021724159240722657, 0.021766143798828123, 0.021687360763549806, 0.0217271671295166, 0.02167807960510254, 0.02169856071472168, 0.02186444854736328, 0.021735424041748046, 0.021773311614990236, 0.02168627166748047, 0.021696512222290038, 0.021703680038452147, 0.021986303329467775, 0.021708799362182618, 0.02166579246520996, 0.021720064163208007, 0.021677120208740235, 0.02155411148071289, 0.021630975723266603, 0.02169856071472168, 0.021626880645751953, 0.021615615844726564, 0.021582847595214845, 0.021587968826293946, 0.021599231719970705, 0.021565439224243164, 0.021594112396240234, 0.021613567352294923, 0.021621759414672852, 0.02166476821899414, 0.021719039916992186, 0.02149478340148926, 0.021606399536132814, 0.021545984268188476, 0.021646335601806642, 0.021374975204467773, 0.021332992553710937, 0.021634048461914062, 0.021557247161865235, 0.021560319900512694, 0.02164838409423828, 0.021554176330566405, 0.021594112396240234, 0.021580799102783203, 0.021599231719970705, 0.021242879867553712, 0.021510143280029297, 0.02213478469848633, 0.021752832412719726, 0.02162483215332031, 0.04727500915527344, 0.02166579246520996, 0.021626880645751953, 0.02167193603515625, 0.02166169548034668, 0.0216627197265625, 0.021716991424560548, 0.02170579147338867, 0.021461952209472657, 0.02128486442565918, 0.021529600143432616, 0.02206003189086914, 0.02286796760559082, 0.02183475112915039, 0.021819391250610352, 0.021808128356933593, 0.02170675277709961, 0.022013952255249023, 0.02170675277709961, 0.021758975982666014, 0.02305740737915039, 0.02206719970703125, 0.02168115234375, 0.02166067123413086, 0.021833728790283204, 0.02182143974304199, 0.02161664009094238, 0.021728256225585937, 0.021613567352294923, 0.02165043258666992, 0.02165350341796875, 0.021585920333862304, 0.02162483215332031, 0.021659648895263672, 0.021629951477050782, 0.02161664009094238, 0.021735424041748046, 0.021716991424560548, 0.021615615844726564, 0.02162483215332031, 0.021685279846191407, 0.02166371154785156, 0.021753856658935547, 0.021287935256958008, 0.021428224563598632, 0.021704704284667968, 0.021696512222290038, 0.02170982360839844, 0.02165247917175293, 0.021732351303100587, 0.02166886329650879, 0.02166067123413086, 0.021639167785644533, 0.02165862464904785, 0.021574655532836915, 0.02125619125366211, 0.021191680908203125, 0.021719039916992186, 0.02189107131958008, 0.021703744888305666, 0.021666751861572266, 0.02169241523742676, 0.02162380790710449, 0.04730060958862305, 0.021544960021972655, 0.021600255966186522, 0.02168217658996582, 0.021764095306396485, 0.0216494083404541, 0.021539840698242187, 0.0212992000579834, 0.021643264770507813, 0.02168012809753418, 0.021758975982666014, 0.021597183227539063, 0.02191155242919922, 0.02165760040283203, 0.021621759414672852, 0.020761632919311525, 0.02070729637145996, 0.02067865562438965, 0.02063155174255371, 0.020485120773315428, 0.020616191864013672, 0.02084249687194824, 0.02085478401184082, 0.020752384185791017, 0.020727807998657227, 0.020707359313964845, 0.02053731155395508, 0.02044108772277832, 0.020666368484497072, 0.020642816543579103, 0.020758527755737305, 0.022282272338867187, 0.021949407577514648, 0.021773311614990236, 0.02191257667541504, 0.02166476821899414, 0.02164531135559082, 0.021766143798828123, 0.021658687591552733, 0.021703615188598632, 0.02167296028137207, 0.0216627197265625, 0.02172211265563965, 0.021843967437744142, 0.021582847595214845, 0.021948415756225585, 0.022079488754272462, 0.021769216537475586, 0.02169139289855957, 0.02162073516845703, 0.021571584701538086, 0.021634048461914062, 0.020761600494384767, 0.020706304550170897, 0.022038528442382813, 0.02170572853088379, 0.021552127838134767, 0.02171392059326172, 0.02162483215332031, 0.02135865592956543, 0.02154284858703613, 0.02167705535888672, 0.02170163154602051, 0.047067134857177735, 0.021300224304199217, 0.02149478340148926, 0.021643264770507813, 0.021735424041748046, 0.021740543365478517, 0.021935104370117187, 0.022038528442382813, 0.02185830307006836, 0.02185420799255371, 0.021757951736450197, 0.02169856071472168, 0.021605375289916993, 0.02167705535888672, 0.021614591598510743, 0.02166067123413086, 0.021622783660888673, 0.02166988754272461, 0.021612543106079102, 0.02127872085571289, 0.02129100799560547, 0.021505023956298826, 0.02165657615661621, 0.021756927490234376, 0.021582847595214845, 0.021551103591918946, 0.02125721549987793, 0.021323776245117186, 0.020505599975585938, 0.020502527236938475, 0.021238784790039062, 0.02172211265563965, 0.021606399536132814, 0.02164019203186035, 0.021626880645751953, 0.02162892723083496, 0.021579776763916016, 0.0215285758972168, 0.021303295135498047, 0.0212992000579834, 0.021404672622680664, 0.02129408073425293, 0.021227519989013673, 0.021195775985717775, 0.02124595260620117, 0.021283840179443358, 0.02146611213684082, 0.020545536041259766, 0.020725759506225586, 0.02081996726989746, 0.020787200927734374, 0.02085478401184082, 0.020770816802978515, 0.021344255447387696, 0.021255168914794922, 0.021527584075927735, 0.021614559173583986, 0.02163199996948242, 0.02082508850097656, 0.020502527236938475, 0.020558847427368163, 0.020516864776611327, 0.020529151916503906, 0.04635340881347656, 0.02064076805114746, 0.02051481628417969, 0.02087014389038086, 0.02129305648803711, 0.02129715156555176, 0.02125823974609375, 0.021410816192626952, 0.02127257537841797, 0.021217279434204102, 0.021368831634521485, 0.021337087631225587, 0.02125823974609375, 0.021606399536132814, 0.021355520248413085, 0.021415935516357423, 0.021411840438842773, 0.021349376678466796, 0.021351423263549805, 0.021385215759277345, 0.021575679779052736, 0.021622783660888673, 0.02162892723083496, 0.021597183227539063, 0.021757951736450197, 0.021638143539428712, 0.021578752517700195, 0.021529600143432616, 0.021634048461914062, 0.021576704025268553, 0.021160959243774414, 0.02205286407470703, 0.022487039566040038, 0.02199760055541992, 0.02176304054260254, 0.021687295913696288, 0.02187775993347168, 0.021558271408081055, 0.02138630485534668, 0.02153875160217285, 0.021774335861206053, 0.02161664009094238, 0.021599231719970705, 0.02161664009094238, 0.02167500877380371, 0.021760000228881835, 0.021448703765869142, 0.021431360244750976, 0.02122438430786133, 0.021341184616088867, 0.021259263992309572, 0.021309440612792968, 0.021385215759277345, 0.021378047943115236, 0.021227519989013673, 0.021817344665527344, 0.02167807960510254, 0.021601280212402343, 0.021593088150024413, 0.021914623260498048, 0.02168012809753418, 0.02168524742126465, 0.02170572853088379, 0.04642406463623047, 0.020714496612548827, 0.021590015411376954, 0.021465087890625, 0.021255199432373046, 0.02131043243408203, 0.021429248809814453, 0.021712896347045898, 0.02162380790710449, 0.021761024475097656, 0.021607423782348634, 0.021803007125854493, 0.02162380790710449, 0.02170982360839844, 0.02185830307006836, 0.021618688583374023, 0.021902336120605468, 0.02168832015991211, 0.02210918426513672, 0.021775360107421874, 0.02167091178894043, 0.021766176223754884, 0.021707744598388673, 0.021780479431152345, 0.021679103851318358, 0.021849088668823242, 0.021768192291259765, 0.021638143539428712, 0.021607423782348634, 0.021770240783691407, 0.021577728271484374, 0.021634048461914062, 0.021583871841430666, 0.021517311096191406, 0.021639167785644533, 0.021760000228881835, 0.021809152603149414, 0.021707775115966797, 0.021604352951049805, 0.021685279846191407, 0.021841888427734376, 0.023136255264282226, 0.0224901123046875, 0.02183065605163574, 0.021625856399536132, 0.02165350341796875, 0.021679103851318358, 0.021708799362182618, 0.02188287925720215, 0.02168832015991211, 0.021749759674072267, 0.021695487976074217, 0.02168422317504883, 0.021617664337158202, 0.021627904891967774, 0.02165247917175293, 0.021627904891967774, 0.021727231979370116, 0.02166374397277832, 0.021758975982666014, 0.021627904891967774, 0.021646335601806642, 0.02187161636352539, 0.047702014923095705, 0.02162073516845703, 0.021441535949707033, 0.021651456832885742, 0.02145382308959961, 0.02128998374938965, 0.02126950454711914, 0.021746688842773438, 0.02168934440612793, 0.02169241523742676, 0.021793792724609375, 0.02179177665710449, 0.02134320068359375, 0.021591039657592775, 0.021721088409423828, 0.021639167785644533, 0.02172313690185547, 0.02164019203186035, 0.021786624908447266, 0.021832704544067383, 0.021588991165161133, 0.021791744232177734, 0.021642240524291992, 0.021799936294555664, 0.021735424041748046, 0.02168115234375, 0.021630975723266603, 0.021578752517700195, 0.021629951477050782, 0.021608448028564452, 0.021544960021972655, 0.021646335601806642, 0.02168524742126465, 0.021801984786987305, 0.021745664596557617, 0.021732351303100587, 0.02165760040283203, 0.021729280471801758, 0.021763071060180664, 0.02165247917175293, 0.021757951736450197, 0.02200067138671875, 0.021629919052124025, 0.021695487976074217, 0.021929983139038087, 0.021704704284667968, 0.021313535690307618, 0.021433343887329103, 0.022417407989501953, 0.021347328186035155, 0.02130636787414551, 0.02127359962463379, 0.021309440612792968, 0.021569536209106444, 0.021598207473754884, 0.021720064163208007, 0.02161664009094238, 0.021622783660888673, 0.02165350341796875, 0.021626880645751953, 0.021634048461914062, 0.021604352951049805, 0.021538816452026367, 0.047666175842285156, 0.02165862464904785, 0.021625856399536132, 0.021679103851318358, 0.021808128356933593, 0.021865472793579102, 0.0223242244720459, 0.021772287368774415, 0.02167705535888672, 0.02163199996948242, 0.021736448287963867, 0.021707775115966797, 0.021590015411376954, 0.021401599884033205, 0.021191680908203125, 0.021325824737548828, 0.02126028823852539, 0.021234687805175782, 0.0212674560546875, 0.02129305648803711, 0.02128179168701172, 0.02130534362792969, 0.021377023696899415, 0.021315584182739256, 0.021785600662231445, 0.021644287109375, 0.02164121627807617, 0.02164838409423828, 0.02169753646850586, 0.021795839309692384, 0.021518335342407227, 0.021583871841430666, 0.021603328704833984, 0.02166169548034668, 0.02166783905029297, 0.02166886329650879, 0.021642240524291992, 0.021625856399536132, 0.021627904891967774, 0.021405696868896484, 0.02129305648803711, 0.021584896087646483, 0.02169856071472168, 0.021592063903808592, 0.021549055099487305, 0.021555200576782226, 0.021621759414672852, 0.02206105613708496, 0.021763071060180664, 0.02168627166748047, 0.021575679779052736, 0.021772287368774415, 0.021651456832885742, 0.021324800491333007, 0.02143539237976074, 0.02172211265563965, 0.021578752517700195, 0.021618688583374023, 0.021593088150024413, 0.02162892723083496, 0.02165862464904785, 0.021638143539428712, 0.021578815460205077, 0.04645574569702148, 0.02072985649108887, 0.020703231811523438, 0.020739072799682616, 0.020509695053100584, 0.02067967987060547, 0.020715520858764647, 0.020724735260009765, 0.021173248291015623, 0.02171801567077637, 0.021639167785644533, 0.022068288803100584, 0.020679616928100587, 0.020755456924438476, 0.020771839141845702, 0.02072985649108887, 0.020822015762329102, 0.020722688674926756, 0.021150720596313476, 0.02127257537841797, 0.021249055862426758, 0.021355487823486327, 0.021634048461914062, 0.021737472534179687, 0.021551103591918946, 0.021951488494873047, 0.021811199188232423, 0.021914623260498048, 0.021712896347045898, 0.021749759674072267, 0.02166374397277832, 0.021737472534179687, 0.02168627166748047, 0.021894144058227538, 0.021794815063476563, 0.021625856399536132, 0.02208563232421875, 0.02253926467895508, 0.022468608856201173, 0.022047744750976563, 0.021712896347045898, 0.02163711929321289, 0.021619712829589844, 0.021669919967651368, 0.021667808532714845, 0.021808128356933593, 0.021602304458618164, 0.021711872100830077, 0.021587968826293946, 0.021703680038452147, 0.021828607559204103, 0.021721088409423828, 0.02170982360839844, 0.02168012809753418, 0.02166681671142578, 0.021951488494873047, 0.02170982360839844, 0.0218603515625, 0.021755903244018555, 0.021795839309692384, 0.02170982360839844, 0.02168524742126465, 0.021710847854614256, 0.04654079818725586, 0.02084351921081543, 0.020822015762329102, 0.02085785675048828, 0.02081177520751953, 0.02078726387023926, 0.020810688018798828, 0.020793344497680662, 0.020746240615844725, 0.021136383056640624, 0.021687295913696288, 0.021779455184936524, 0.02167398452758789, 0.021699583053588867, 0.021703680038452147, 0.02164735984802246, 0.021785600662231445, 0.021613567352294923, 0.021703680038452147, 0.02163609504699707, 0.021638143539428712, 0.021391359329223633, 0.02125312042236328, 0.021317632675170898, 0.02124185562133789, 0.02127462387084961, 0.02124595260620117, 0.02120806312561035, 0.02122137641906738, 0.021396480560302734, 0.021627904891967774, 0.021555200576782226, 0.021604352951049805, 0.021344255447387696, 0.02119987106323242, 0.021365760803222656, 0.021603328704833984, 0.021835775375366212, 0.021774335861206053, 0.02186137580871582, 0.021754880905151368, 0.021635072708129883, 0.021609472274780273, 0.02165555191040039, 0.02164838409423828, 0.021791744232177734, 0.022615039825439453, 0.021902336120605468, 0.021704704284667968, 0.02163302421569824, 0.02169856071472168, 0.02166681671142578, 0.021606399536132814, 0.021580799102783203, 0.02164019203186035, 0.02143129539489746, 0.021245983123779295, 0.021485536575317384, 0.0216494083404541, 0.02164739227294922, 0.021622751235961916, 0.021736448287963867, 0.02184601593017578]",tokens/s,45.599458576959975,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,google/recurrentgemma-2b,google/recurrentgemma-2b,cuda,0,42,,,True,,,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.223-212.873.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.0,,,,1.21.4,,,,0.12.0,,,recurrent_gemma,MB,1609.699328,7432.830976,0.0,6784.28672,5879.090688,s,10,3.628090881347656,0.3628090881347656,0.008630217165898083,0.3601696624755859,0.3731070434570313,0.3738421112060547,0.37443016540527346,"[0.3727458190917969, 0.3729436950683594, 0.35648641967773437, 0.3706185607910156, 0.37457717895507814, 0.36381423950195313, 0.35652508544921874, 0.35297344970703126, 0.35305783081054687, 0.3543486022949219]",tokens/s,705.6052573437981,kWh,4.192325053807568e-06,2.297189590526233e-06,7.999891457379092e-06,1.4489406101712894e-05,tokens/kWh,17668080.95534961,MB,1652.154368,7434.928128,0.0,6786.383872,5879.093248,s,10,20.26380212402344,2.026380212402344,0.010793219586010728,2.0263109130859376,2.0368320312500003,2.0436820068359376,2.0491619873046876,"[2.02282177734375, 2.024660888671875, 2.0279609375, 2.050531982421875, 2.0280584716796874, 2.016791015625, 2.031745361328125, 2.01314111328125, 2.01278076171875, 2.035309814453125]",tokens/s,31.08992064490766,kWh,2.387496194515076e-05,1.3084500371183901e-05,6.631733466301983e-05,0.00010327679697935449,tokens/kWh,610011.1723313223,,s,630,20.262420440673825,0.03216257212805369,0.0005742437513800482,0.031851520538330076,0.03311728630065918,0.03322188892364502,0.03388601364135743,"[0.03252121734619141, 0.0317706241607666, 0.03183718490600586, 0.03184025573730469, 0.031751167297363284, 0.031780864715576174, 0.03184537506103516, 0.03180748748779297, 0.03203788757324219, 0.03182080078125, 0.03174092864990234, 0.031784959793090824, 0.03178598403930664, 0.03189043235778809, 0.03181465530395508, 0.031920127868652344, 0.03180748748779297, 0.03177267265319824, 0.031866880416870115, 0.03187404823303223, 0.03181875228881836, 0.031784959793090824, 0.032628734588623046, 0.032366592407226565, 0.031886335372924804, 0.03180646324157715, 0.03167129516601563, 0.03180851173400879, 0.03175526428222656, 0.03179520034790039, 0.031908863067626955, 0.03180441665649414, 0.031732736587524416, 0.03182899284362793, 0.031893503189086916, 0.03182284736633301, 0.03278643035888672, 0.031764480590820314, 0.03186073684692383, 0.03199078369140625, 0.031747072219848634, 0.03289190292358399, 0.03174604797363281, 0.03159040069580078, 0.031959039688110355, 0.03304140853881836, 0.0317573127746582, 0.03172659111022949, 0.031855615615844726, 0.03172454452514648, 0.03282124710083008, 0.03195699119567871, 0.03178598403930664, 0.03217407989501953, 0.03377971267700195, 0.03373567962646484, 0.03315507125854492, 0.03310079956054687, 0.03311206436157227, 0.03309465789794922, 0.033140735626220705, 0.031970304489135744, 0.032833534240722655, 0.03208806228637695, 0.03185663986206055, 0.031870975494384765, 0.032898048400878906, 0.031855615615844726, 0.031866880416870115, 0.03178803253173828, 0.031784959793090824, 0.03179110336303711, 0.031900672912597655, 0.03182694435119629, 0.03180544090270996, 0.032702465057373044, 0.03216998291015625, 0.0332861442565918, 0.033058815002441407, 0.03301171112060547, 0.03313356781005859, 0.033277950286865234, 0.033263614654541016, 0.03325030517578125, 0.031662080764770506, 0.03174502372741699, 0.03177881622314453, 0.03177881622314453, 0.03174092864990234, 0.03174297523498535, 0.03182592010498047, 0.031735807418823245, 0.03177984046936035, 0.03172761535644531, 0.031784959793090824, 0.03167948722839355, 0.031720447540283206, 0.03266355133056641, 0.03175014305114746, 0.03181465530395508, 0.031850496292114255, 0.03181363105773926, 0.031719423294067385, 0.03181465530395508, 0.031821823120117186, 0.031718399047851564, 0.03279052734375, 0.03267071914672852, 0.032086017608642575, 0.03213926315307617, 0.03292671966552734, 0.033081344604492184, 0.03315814590454102, 0.033086463928222655, 0.03211775970458984, 0.03176959991455078, 0.03172863960266113, 0.03168870353698731, 0.03175014305114746, 0.03167948722839355, 0.033119232177734374, 0.03176038360595703, 0.03172761535644531, 0.03179110336303711, 0.031579135894775394, 0.03170816040039062, 0.032031742095947266, 0.03196723175048828, 0.03263078308105469, 0.03198054313659668, 0.033051647186279294, 0.033073150634765625, 0.03299327850341797, 0.033185791015625, 0.0318474235534668, 0.03341516876220703, 0.03320115280151367, 0.03171327972412109, 0.0316866569519043, 0.03177779197692871, 0.03167436790466309, 0.031697919845581055, 0.03311206436157227, 0.03307417678833008, 0.03189657592773437, 0.03179007911682129, 0.03168972778320313, 0.031697919845581055, 0.0317255687713623, 0.03183718490600586, 0.032031742095947266, 0.033132545471191405, 0.031834112167358396, 0.03178188705444336, 0.03176652717590332, 0.03175424003601074, 0.03181363105773926, 0.03183718490600586, 0.03168972778320313, 0.033235969543457033, 0.032438270568847655, 0.03197952079772949, 0.03173785591125488, 0.03174297523498535, 0.03179827117919922, 0.03174195289611816, 0.032399360656738284, 0.031720447540283206, 0.032538623809814454, 0.03175628852844238, 0.0317071361541748, 0.032331775665283204, 0.0332042236328125, 0.032968704223632815, 0.033104896545410156, 0.033102848052978515, 0.03316326522827148, 0.031710208892822264, 0.03168563270568848, 0.031697919845581055, 0.03171327972412109, 0.031736831665039066, 0.031697919845581055, 0.03188121604919433, 0.032026622772216795, 0.03312844848632813, 0.03181056022644043, 0.0318791675567627, 0.03179520034790039, 0.03428147125244141, 0.033413120269775394, 0.03352166366577149, 0.03335270309448242, 0.033070079803466795, 0.03322880172729492, 0.031716352462768556, 0.03182796859741211, 0.03177574348449707, 0.031644672393798826, 0.03163955116271973, 0.03176652717590332, 0.031628288269042966, 0.031764480590820314, 0.032233470916748046, 0.03306496047973633, 0.03300249481201172, 0.031784959793090824, 0.031718399047851564, 0.03172761535644531, 0.031834112167358396, 0.03297792053222656, 0.03407462310791016, 0.03316428756713867, 0.0331673583984375, 0.03199590492248535, 0.03187609672546387, 0.03191705513000488, 0.03176755142211914, 0.03182080078125, 0.03180544090270996, 0.03169587135314941, 0.03321343994140625, 0.033999870300292966, 0.03318067169189453, 0.032882686614990234, 0.032963584899902344, 0.032998401641845705, 0.033181697845458984, 0.0331673583984375, 0.033165313720703124, 0.03301990509033203, 0.03311513519287109, 0.03303628921508789, 0.032939006805419925, 0.03309568023681641, 0.034473983764648435, 0.03358003234863281, 0.03331891250610351, 0.03180646324157715, 0.031870975494384765, 0.03189555168151856, 0.0318474235534668, 0.03180441665649414, 0.03189760017395019, 0.03216998291015625, 0.032525310516357424, 0.031927295684814457, 0.031925247192382815, 0.031850496292114255, 0.0318525447845459, 0.03235532760620117, 0.033070079803466795, 0.033426433563232424, 0.031954944610595705, 0.03299942398071289, 0.03318067169189453, 0.03183103942871094, 0.03182489585876465, 0.03179212760925293, 0.03313663864135742, 0.0318023681640625, 0.031838207244873046, 0.03181363105773926, 0.03173990440368652, 0.03243212890625, 0.033960960388183595, 0.033188865661621096, 0.03316940689086914, 0.03314483261108398, 0.03187814331054688, 0.031927295684814457, 0.03190784072875977, 0.03340697479248047, 0.03312025451660156, 0.03181875228881836, 0.03177369689941406, 0.03180031967163086, 0.031714303970336914, 0.0317890567779541, 0.031736831665039066, 0.03180646324157715, 0.03172352027893066, 0.03172352027893066, 0.031716352462768556, 0.03191398429870605, 0.031497215270996096, 0.03170918464660644, 0.03204608154296875, 0.03181465530395508, 0.03176959991455078, 0.03180851173400879, 0.0317573127746582, 0.031699968338012696, 0.03174195289611816, 0.03178598403930664, 0.03177984046936035, 0.031835136413574217, 0.03273932647705078, 0.03333631896972656, 0.03297177505493164, 0.03174502372741699, 0.03304755020141602, 0.03311820983886719, 0.033152000427246094, 0.0325294075012207, 0.031801343917846676, 0.03172352027893066, 0.03168767929077149, 0.03199590492248535, 0.03188121604919433, 0.031784959793090824, 0.0327014389038086, 0.03180646324157715, 0.031887359619140625, 0.03177574348449707, 0.03196211242675781, 0.03189145660400391, 0.032494590759277346, 0.032611328125, 0.03205222320556641, 0.03225600051879883, 0.03202764892578125, 0.03220479965209961, 0.03345305633544922, 0.033051647186279294, 0.033173503875732424, 0.032942081451416014, 0.03288371276855469, 0.032846847534179685, 0.03273113632202149, 0.03184230422973633, 0.031749120712280275, 0.031748096466064454, 0.03182284736633301, 0.03184537506103516, 0.0317573127746582, 0.03179417610168457, 0.031749120712280275, 0.03175833511352539, 0.031801343917846676, 0.031735807418823245, 0.031851520538330076, 0.03196723175048828, 0.03195084762573242, 0.0318525447845459, 0.0318474235534668, 0.03178803253173828, 0.03171126365661621, 0.03172041511535645, 0.03182080078125, 0.03253760147094727, 0.03213926315307617, 0.031936511993408204, 0.03181260871887207, 0.031886335372924804, 0.031838207244873046, 0.03177574348449707, 0.03180339241027832, 0.03187507247924805, 0.031747072219848634, 0.031833087921142575, 0.03183718490600586, 0.03176345634460449, 0.031715328216552735, 0.03175424003601074, 0.03180953598022461, 0.031835136413574217, 0.03181670379638672, 0.031936511993408204, 0.031784959793090824, 0.03180748748779297, 0.031937536239624024, 0.03182592010498047, 0.03182592010498047, 0.03180851173400879, 0.03177779197692871, 0.03176345634460449, 0.03177471923828125, 0.03208806228637695, 0.03225907135009766, 0.03290726470947265, 0.03274444961547852, 0.03172966384887695, 0.031851520538330076, 0.03179315185546875, 0.03176140785217285, 0.031866880416870115, 0.03216998291015625, 0.031921152114868165, 0.031838207244873046, 0.0317706241607666, 0.031720447540283206, 0.031959039688110355, 0.0318474235534668, 0.031938560485839845, 0.032045055389404296, 0.03174092864990234, 0.031765504837036135, 0.03226521682739258, 0.031898624420166014, 0.03185868835449219, 0.03182592010498047, 0.031959039688110355, 0.03259392166137695, 0.03343155288696289, 0.033307647705078124, 0.03299532699584961, 0.032956447601318356, 0.03284988784790039, 0.03227545547485351, 0.03180748748779297, 0.031922176361083986, 0.03174604797363281, 0.03178598403930664, 0.03183001518249512, 0.03184639930725098, 0.03177881622314453, 0.031715328216552735, 0.031899648666381834, 0.0317706241607666, 0.031834112167358396, 0.03175935935974121, 0.03173990440368652, 0.031871999740600586, 0.032996353149414064, 0.03297382354736328, 0.03285708618164063, 0.03275775909423828, 0.03279564666748047, 0.03314176177978516, 0.031893503189086916, 0.031877151489257814, 0.032516063690185545, 0.03295334243774414, 0.032914432525634765, 0.032950271606445314, 0.03391897583007813, 0.03316633605957031, 0.03236249542236328, 0.031884288787841795, 0.032396289825439455, 0.032039936065673826, 0.03188531112670898, 0.0317890567779541, 0.031838207244873046, 0.03184537506103516, 0.03179315185546875, 0.031853567123413085, 0.03177779197692871, 0.031751167297363284, 0.03170508766174317, 0.03183206367492676, 0.03180339241027832, 0.03181260871887207, 0.03237887954711914, 0.03227545547485351, 0.03210649490356445, 0.03263488006591797, 0.03203071975708008, 0.031870975494384765, 0.03186278343200684, 0.03182080078125, 0.03208294296264649, 0.03190681648254395, 0.03175526428222656, 0.03181363105773926, 0.03181260871887207, 0.0317573127746582, 0.03176345634460449, 0.031867904663085936, 0.03176652717590332, 0.03170201683044434, 0.031752191543579104, 0.0315545597076416, 0.03167027282714844, 0.031942655563354495, 0.032304126739501955, 0.03360358428955078, 0.031849472045898435, 0.03189760017395019, 0.03177369689941406, 0.03168870353698731, 0.031682559967041016, 0.03173990440368652, 0.031643648147583005, 0.03186278343200684, 0.033014785766601565, 0.03339571380615235, 0.03311820983886719, 0.032299007415771484, 0.032345088958740234, 0.03172249603271484, 0.031730688095092774, 0.03177267265319824, 0.0316866569519043, 0.03170611190795898, 0.0316866569519043, 0.03165286445617676, 0.03179724884033203, 0.03169484710693359, 0.03186073684692383, 0.03184537506103516, 0.031833087921142575, 0.031939584732055666, 0.032031742095947266, 0.031866880416870115, 0.03186483192443847, 0.03180544090270996, 0.0318474235534668, 0.031703039169311525, 0.03258060836791992, 0.03402547073364258, 0.032161792755126956, 0.03185766410827637, 0.03280486297607422, 0.03250483322143555, 0.031784959793090824, 0.031768575668334964, 0.03172966384887695, 0.03171225547790527, 0.03177881622314453, 0.03170611190795898, 0.03178291130065918, 0.03177164840698242, 0.031733760833740236, 0.031753215789794925, 0.03181363105773926, 0.031939584732055666, 0.032643070220947264, 0.03171327972412109, 0.03170918464660644, 0.032118785858154295, 0.031801343917846676, 0.031693824768066405, 0.03175526428222656, 0.0317255687713623, 0.031660032272338864, 0.03174297523498535, 0.031703039169311525, 0.03282124710083008, 0.03287551879882813, 0.03173990440368652, 0.03183206367492676, 0.033097728729248044, 0.032571392059326174, 0.03226419067382812, 0.03184127998352051, 0.03170099258422852, 0.03168460845947266, 0.03172249603271484, 0.031715328216552735, 0.03168358421325684, 0.03168358421325684, 0.03176959991455078, 0.031783935546875, 0.031665151596069335, 0.031697919845581055, 0.032099327087402346, 0.03183103942871094, 0.03178700828552246, 0.031751167297363284, 0.0317890567779541, 0.031731712341308595, 0.03146240043640137, 0.0318156795501709, 0.031923200607299806, 0.03170611190795898, 0.032546817779541014, 0.031867904663085936, 0.03186380767822266, 0.031753215789794925, 0.03182080078125, 0.03208806228637695, 0.0325591049194336, 0.03183923149108887, 0.0322979850769043, 0.03271475219726563, 0.032882686614990234, 0.032080894470214845, 0.03226419067382812, 0.0318474235534668, 0.03173785591125488, 0.032794624328613284, 0.03264716720581055, 0.03246284866333008, 0.032228351593017575, 0.03179622459411621, 0.03221401596069336, 0.03337420654296875, 0.031835136413574217, 0.03263488006591797, 0.032882720947265624, 0.032016353607177736, 0.03205017471313477, 0.03268915176391601, 0.032266239166259765, 0.03229695892333984, 0.03358617782592774, 0.032080894470214845, 0.03183616065979004, 0.032217086791992186, 0.03196928024291992, 0.0321003532409668, 0.03269222259521484, 0.033192958831787106, 0.03213619232177734, 0.03177164840698242, 0.03180441665649414, 0.031834112167358396, 0.032876544952392575, 0.03250688171386719, 0.03179007911682129, 0.03174092864990234, 0.0321525764465332, 0.03184435272216797, 0.03177574348449707, 0.03264614486694336, 0.031954944610595705, 0.031733760833740236, 0.03227852630615234, 0.03196006393432617, 0.033805313110351565, 0.03238195037841797, 0.03213721466064453, 0.032906238555908206, 0.03251302337646484, 0.03184025573730469, 0.032863231658935545, 0.03277004623413086, 0.03311718368530273]",tokens/s,31.092040649564638,,,,,,MB,999.477248,7390.887936,0.0,6761.218048,6730.975744,s,1,6.99474755859375,6.99474755859375,0.0,6.99474755859375,6.99474755859375,6.99474755859375,6.99474755859375,[6.99474755859375],,kWh,2.747102002083719e-06,1.4654193532931818e-06,6.663060885994665e-06,1.0875582241371565e-05, bfloat16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/gpt-j-6b,EleutherAI/gpt-j-6b,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1266.7904,13227.261952,0.0,12580.814848,12440.741376,s,10,0.7647395248413086,0.07647395248413086,0.0011402058770447509,0.07606838607788086,0.07667963256835937,0.07828215332031249,0.079564169921875,"[0.07988467407226563, 0.07614636993408203, 0.0760401611328125, 0.07606716918945312, 0.07632351684570313, 0.07598838043212891, 0.07611820983886719, 0.0760499496459961, 0.0760514907836914, 0.07606960296630859]",tokens/s,3347.545035718177,kWh,9.010245189421908e-07,4.937203536883316e-07,5.12003355730301e-06,6.514778429933533e-06,tokens/kWh,39295273.469893254,MB,1266.7904,13227.261952,0.0,12580.814848,12543.514112,s,10,22.3466162109375,2.23466162109375,0.0028811882239978067,2.233171142578125,2.2379523681640627,2.239874865722656,2.2414128637695314,"[2.24179736328125, 2.236292236328125, 2.233044677734375, 2.232322265625, 2.232956787109375, 2.23306884765625, 2.237525146484375, 2.2332734375, 2.232357177734375, 2.233978271484375]",tokens/s,28.19218775913142,kWh,2.634037407384778e-05,1.443535125393282e-05,0.00011122697618689687,0.00015200270151467747,tokens/kWh,414466.3178497303,,s,629,22.99535057830811,0.03655858597505263,0.009095728955772675,0.03544575881958008,0.035636379241943364,0.03577099609375,0.11189739501953125,"[0.03591167831420899, 0.0359659538269043, 0.035760128021240234, 0.03568537521362305, 0.03593318557739258, 0.0358287353515625, 0.03571507263183594, 0.03598950576782227, 0.03600076675415039, 0.0359659538269043, 0.0357498893737793, 0.03569155120849609, 0.03587990570068359, 0.035806209564208984, 0.035625984191894534, 0.03537408065795898, 0.035337215423583986, 0.03540582275390625, 0.03543552017211914, 0.03538739013671875, 0.03537715148925781, 0.03544780731201172, 0.03531673431396484, 0.03538739013671875, 0.035394561767578124, 0.0354150390625, 0.03540172958374024, 0.03539148712158203, 0.03583078384399414, 0.03559628677368164, 0.03538739013671875, 0.03539763259887695, 0.03545292663574219, 0.035438625335693356, 0.03543548965454101, 0.035418113708496096, 0.035460094451904296, 0.035487743377685545, 0.035432449340820314, 0.035476478576660156, 0.035492862701416016, 0.03549388885498047, 0.03545395278930664, 0.035530784606933596, 0.03570991897583008, 0.03560038375854492, 0.03551027297973633, 0.03551232147216797, 0.03552979278564453, 0.03550815963745117, 0.03549491119384766, 0.03553792190551758, 0.035514366149902346, 0.035555328369140625, 0.035530750274658206, 0.035560447692871096, 0.035520511627197264, 0.035579902648925785, 0.03556454467773437, 0.03554508972167969, 0.03559423828125, 0.03564441680908203, 0.11201026916503906, 0.035310558319091793, 0.03524710464477539, 0.0353361930847168, 0.03527884674072266, 0.03528192138671875, 0.03524710464477539, 0.035410945892333984, 0.035643390655517575, 0.03552870559692383, 0.03543552017211914, 0.03535769653320313, 0.035345409393310545, 0.035329025268554685, 0.03536793518066406, 0.03534438323974609, 0.035366912841796876, 0.03539148712158203, 0.035373088836669925, 0.03541705703735352, 0.035416065216064455, 0.03538534545898438, 0.03543756866455078, 0.035416065216064455, 0.03539046478271484, 0.03538739013671875, 0.03539251327514648, 0.03539148712158203, 0.03541299057006836, 0.03542323303222656, 0.03547443389892578, 0.03545600128173828, 0.03543040084838867, 0.0354150390625, 0.035399681091308595, 0.035538944244384765, 0.03549184036254883, 0.03549184036254883, 0.036168704986572264, 0.03577040100097656, 0.0355972785949707, 0.035451904296875, 0.03549593734741211, 0.03546931076049804, 0.03550310516357422, 0.03553792190551758, 0.0355788803100586, 0.035519489288330076, 0.0355491828918457, 0.03550515365600586, 0.03553996658325195, 0.0355153923034668, 0.03555132675170898, 0.03554703903198242, 0.03556454467773437, 0.035547134399414065, 0.035593215942382815, 0.03570278549194336, 0.03632128143310547, 0.03584921646118164, 0.03561983871459961, 0.03558195114135742, 0.03560857772827149, 0.1119467544555664, 0.03524607849121094, 0.035225601196289064, 0.035299327850341795, 0.03527577590942383, 0.0352542724609375, 0.03529830551147461, 0.03530649566650391, 0.035329025268554685, 0.035312641143798826, 0.035372032165527346, 0.035291168212890626, 0.035358688354492185, 0.03531161499023437, 0.03535564804077149, 0.03549388885498047, 0.035310592651367184, 0.03532799911499023, 0.0353331184387207, 0.035631103515625, 0.03546316909790039, 0.03540172958374024, 0.03538739013671875, 0.035448833465576174, 0.03540787124633789, 0.03547859191894531, 0.03548262405395508, 0.035441600799560546, 0.03540582275390625, 0.03539763259887695, 0.03545600128173828, 0.03544268798828125, 0.03541196823120117, 0.035419136047363284, 0.03539763259887695, 0.03545804977416992, 0.035451904296875, 0.03548057556152344, 0.035509246826171875, 0.03546419143676758, 0.035454975128173825, 0.03544985580444336, 0.03536896133422852, 0.03545907211303711, 0.03545702362060547, 0.03547238540649414, 0.035501056671142575, 0.03556147384643555, 0.03546214294433594, 0.03551641464233399, 0.03562188720703125, 0.03566387176513672, 0.03545292663574219, 0.03552665710449219, 0.035510303497314454, 0.03566691207885742, 0.035558399200439454, 0.035506175994873046, 0.035533824920654294, 0.035533824920654294, 0.035585025787353515, 0.03558297729492187, 0.03559423828125, 0.11191705322265624, 0.03525120162963867, 0.03523174285888672, 0.03526348876953125, 0.03526860809326172, 0.035266559600830076, 0.03525939178466797, 0.03530342483520508, 0.035297279357910154, 0.03532287979125977, 0.03529011154174805, 0.03531366348266601, 0.03542736053466797, 0.035363807678222656, 0.0352911376953125, 0.035315711975097655, 0.03527888107299805, 0.03534844970703125, 0.03550310516357422, 0.03536793518066406, 0.035329025268554685, 0.03537100982666016, 0.03539353561401367, 0.035342334747314456, 0.03550515365600586, 0.03550515365600586, 0.0353966064453125, 0.03539046478271484, 0.035373054504394534, 0.03543654251098633, 0.035356670379638674, 0.03543142318725586, 0.03544063949584961, 0.03546623992919922, 0.03543961715698242, 0.03540377426147461, 0.03546828842163086, 0.03543756866455078, 0.03548876953125, 0.03544678497314453, 0.03548160171508789, 0.035492862701416016, 0.03552358245849609, 0.03545395278930664, 0.03548672103881836, 0.03551027297973633, 0.03562700653076172, 0.03549593734741211, 0.03549491119384766, 0.035520511627197264, 0.03553177642822265, 0.03550515365600586, 0.035517440795898435, 0.03548876953125, 0.035479583740234376, 0.03556143951416016, 0.035527679443359376, 0.03552153778076172, 0.03559423828125, 0.035506175994873046, 0.03556454467773437, 0.03557580947875977, 0.03559936141967773, 0.11289907073974609, 0.03529216003417969, 0.03528499221801758, 0.03527065658569336, 0.03526348876953125, 0.03531366348266601, 0.035286014556884765, 0.03531270217895508, 0.03535558319091797, 0.035422206878662106, 0.0353331184387207, 0.0352911376953125, 0.03533926391601563, 0.03526144027709961, 0.03535564804077149, 0.03534438323974609, 0.035272705078125, 0.035315711975097655, 0.035399681091308595, 0.0353361930847168, 0.03536896133422852, 0.035394561767578124, 0.035341377258300784, 0.035444671630859376, 0.03538534545898438, 0.03538534545898438, 0.03537919998168945, 0.03544063949584961, 0.03539353561401367, 0.03540889739990234, 0.03567923355102539, 0.03535769653320313, 0.03544575881958008, 0.03553792190551758, 0.03551232147216797, 0.03545395278930664, 0.0354068489074707, 0.035451904296875, 0.03543142318725586, 0.03543654251098633, 0.03541401672363281, 0.03544063949584961, 0.03547340774536133, 0.03546623992919922, 0.035410945892333984, 0.03542425537109375, 0.03550310516357422, 0.03546214294433594, 0.03544780731201172, 0.03543961715698242, 0.03548672103881836, 0.03577139282226562, 0.035692543029785154, 0.03556249618530274, 0.03550515365600586, 0.03554611206054688, 0.03546726226806641, 0.03558399963378906, 0.035555328369140625, 0.03552358245849609, 0.035520511627197264, 0.035538944244384765, 0.035536895751953124, 0.11190169525146484, 0.035404800415039066, 0.03523276901245117, 0.03529523086547852, 0.03526758575439453, 0.03527884674072266, 0.03523481750488281, 0.03527276611328125, 0.03529107284545899, 0.03530649566650391, 0.03528499221801758, 0.03529318237304688, 0.03530547332763672, 0.03529830551147461, 0.03531673431396484, 0.035335166931152344, 0.03527475357055664, 0.03531980895996094, 0.03531673431396484, 0.03534131240844727, 0.0353361930847168, 0.035348480224609374, 0.03530547332763672, 0.0353546257019043, 0.03539251327514648, 0.03538431930541992, 0.03537411117553711, 0.0353730239868164, 0.035378177642822264, 0.035517440795898435, 0.03595980834960937, 0.035501056671142575, 0.03542835235595703, 0.03540889739990234, 0.035448833465576174, 0.035416065216064455, 0.03545811080932617, 0.035469249725341795, 0.03542323303222656, 0.035453025817871096, 0.03544259262084961, 0.035504127502441404, 0.035489791870117186, 0.03546726226806641, 0.03544371032714844, 0.035538944244384765, 0.03548364639282227, 0.035525630950927735, 0.03549087905883789, 0.03557062530517578, 0.0355153923034668, 0.03553996658325195, 0.035519489288330076, 0.035519489288330076, 0.03550822448730469, 0.03556147384643555, 0.035579902648925785, 0.03558195114135742, 0.03564031982421875, 0.03582463836669922, 0.03583078384399414, 0.03565158462524414, 0.035555328369140625, 0.11201331329345703, 0.0352542724609375, 0.035214336395263675, 0.0352542724609375, 0.03535564804077149, 0.035250175476074216, 0.03532908630371094, 0.035390399932861326, 0.03531673431396484, 0.035451904296875, 0.03538022232055664, 0.035296257019042966, 0.035350528717041016, 0.035310592651367184, 0.035448833465576174, 0.03822387313842773, 0.03563212966918945, 0.035323902130126955, 0.03539865493774414, 0.03528396987915039, 0.035320831298828126, 0.03534950256347656, 0.03537715148925781, 0.035394561767578124, 0.03532799911499023, 0.035329025268554685, 0.03537919998168945, 0.03547545623779297, 0.03542425537109375, 0.03544780731201172, 0.03543142318725586, 0.03546828842163086, 0.03543654251098633, 0.035402751922607424, 0.03539046478271484, 0.035394561767578124, 0.03546316909790039, 0.03548364639282227, 0.03542425537109375, 0.03539558410644531, 0.03554816055297851, 0.035998783111572265, 0.03545491027832031, 0.035545120239257814, 0.03551023864746094, 0.03546419143676758, 0.035501056671142575, 0.03547343826293945, 0.03556655883789062, 0.0356577262878418, 0.03550515365600586, 0.03565055847167969, 0.03559628677368164, 0.03554611206054688, 0.035499073028564455, 0.03550918579101563, 0.03551846313476562, 0.035517440795898435, 0.03581235122680664, 0.036141056060791016, 0.035707904815673826, 0.035705856323242184, 0.03563724899291992, 0.1119078369140625, 0.03528806304931641, 0.035280895233154294, 0.03546316909790039, 0.035302398681640625, 0.03532492828369141, 0.03523276901245117, 0.03534643173217773, 0.03533107376098633, 0.03532287979125977, 0.035299327850341795, 0.03534643173217773, 0.035366943359375, 0.03536278533935547, 0.035451904296875, 0.035343360900878903, 0.03538227081298828, 0.03539251327514648, 0.03532492828369141, 0.035297279357910154, 0.035422206878662106, 0.03535564804077149, 0.035394561767578124, 0.03535564804077149, 0.03536588668823242, 0.035490814208984374, 0.03540172958374024, 0.03539558410644531, 0.03536896133422852, 0.03544063949584961, 0.035398719787597656, 0.03539859390258789, 0.03538534545898438, 0.035418113708496096, 0.03537510299682617, 0.035515457153320315, 0.03543852615356445, 0.035438591003417966, 0.0354334716796875, 0.03546419143676758, 0.03561062240600586, 0.03561164855957031, 0.03546112060546875, 0.035489791870117186, 0.03544985580444336, 0.035471359252929685, 0.03548160171508789, 0.03552771377563477, 0.035517406463623044, 0.03555635070800781, 0.0355153923034668, 0.035514366149902346, 0.03555430221557617, 0.03554304122924805, 0.0357314567565918, 0.03566185760498047, 0.035587039947509766, 0.03547340774536133, 0.03554412841796875, 0.03560748672485352, 0.03553996658325195, 0.03570073699951172, 0.035604480743408204, 0.11188633728027343, 0.035266559600830076, 0.03523587036132812, 0.035308513641357425, 0.03527475357055664, 0.03528806304931641, 0.03526348876953125, 0.035310592651367184, 0.035369983673095705, 0.035323902130126955, 0.03530956649780274, 0.035297279357910154, 0.03529318237304688, 0.035272705078125, 0.03532287979125977, 0.03530137634277344, 0.035286014556884765, 0.035315711975097655, 0.0353361930847168, 0.035389438629150394, 0.035464256286621094, 0.03558291244506836, 0.03560140609741211, 0.03535769653320313, 0.03549388885498047, 0.03536793518066406, 0.035351551055908204, 0.035367969512939454, 0.035418079376220706, 0.03538431930541992, 0.03538534545898438, 0.03536383819580078, 0.03539763259887695, 0.035416065216064455, 0.03548876953125, 0.03540991973876953, 0.035438591003417966, 0.03547955322265625, 0.03540787124633789, 0.03543040084838867, 0.03543756866455078, 0.03543552017211914, 0.035476478576660156, 0.03544575881958008, 0.03544166564941406, 0.035514366149902346, 0.03546214294433594, 0.03543251037597656, 0.03553171157836914, 0.035576831817626955, 0.03550822448730469, 0.03564851379394531, 0.03569561767578125, 0.035550209045410154, 0.0354969596862793, 0.03548982238769531, 0.03555120086669922, 0.035566593170166014, 0.03558707046508789, 0.03558399963378906, 0.03560038375854492, 0.035542015075683595, 0.03561164855957031, 0.11184844970703126, 0.03526041412353516, 0.035236862182617186, 0.035315711975097655, 0.03525632095336914, 0.03527372741699219, 0.035198974609375, 0.03531161499023437, 0.035310592651367184, 0.0353331184387207, 0.03528396987915039, 0.03524300765991211, 0.035329025268554685, 0.03538227081298828, 0.035342334747314456, 0.03532492828369141, 0.0353361930847168, 0.03530342483520508, 0.03537100982666016, 0.03535257720947266, 0.035335166931152344, 0.035337215423583986, 0.03538431930541992, 0.03544575881958008, 0.035364864349365234, 0.035343360900878903, 0.035476478576660156, 0.03539558410644531, 0.03543654251098633, 0.0354150390625, 0.03539251327514648, 0.03538639831542969, 0.03566281509399414, 0.03565465545654297, 0.03549593734741211, 0.035535873413085936, 0.035517440795898435, 0.03548160171508789, 0.03542323303222656, 0.0354785270690918, 0.03544268798828125, 0.03552460861206055, 0.035434497833251956, 0.035471359252929685, 0.03543654251098633, 0.035471359252929685, 0.03546316909790039, 0.035489791870117186, 0.03550310516357422, 0.03555942535400391, 0.035705856323242184, 0.03618406295776367, 0.03579391860961914, 0.03557785415649414, 0.035509246826171875, 0.03556966400146484, 0.03566387176513672, 0.035620864868164064, 0.03554508972167969, 0.035519489288330076, 0.03575507354736328, 0.03563616180419922, 0.03559833526611328]",tokens/s,27.353355534111582,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,stabilityai/stablelm-2-1_6b,stabilityai/stablelm-2-1_6b,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1270.910976,4299.685888,0.0,3653.238784,3408.132096,s,10,0.2341802215576172,0.02341802215576172,0.0002974755822864154,0.023323872566223147,0.02352355136871338,0.0239109112739563,0.024220799198150633,"[0.024298271179199218, 0.023331424713134766, 0.023350719451904298, 0.023249759674072265, 0.02326576042175293, 0.02343747138977051, 0.023316320419311524, 0.023334911346435547, 0.023300479888916016, 0.02329510307312012]",tokens/s,10931.751550034907,kWh,2.777987985201782e-07,1.5216764054777846e-07,1.3773817748309843e-06,1.8073482138989412e-06,tokens/kWh,141643983.17451978,MB,1270.910976,4299.685888,0.0,3653.238784,3474.14784,s,10,12.8284580078125,1.28284580078125,0.02374972515115311,1.28130615234375,1.3215483520507811,1.3223690856933594,1.3230256726074219,"[1.2873050537109374, 1.2818013916015625, 1.2574949951171874, 1.2532855224609376, 1.2692486572265624, 1.3231898193359375, 1.2808109130859375, 1.2578243408203125, 1.29613134765625, 1.321365966796875]",tokens/s,49.109565593645904,kWh,1.4904885215720724e-05,8.16770102005328e-06,3.9106511723368935e-05,6.217909795914294e-05,tokens/kWh,1013202.2185557671,,s,629,13.015847944259642,0.02069292200995174,0.002854091788221458,0.020067327499389647,0.021038489532470703,0.021207449340820313,0.04363136993408203,"[0.021745664596557617, 0.020215808868408205, 0.020060159683227538, 0.02001408004760742, 0.019984384536743165, 0.01991372871398926, 0.019983360290527344, 0.019928064346313477, 0.01987379264831543, 0.019949567794799804, 0.01998028755187988, 0.019973119735717772, 0.01993011283874512, 0.019812351226806642, 0.020019264221191407, 0.019854272842407227, 0.019893280029296877, 0.01991983985900879, 0.019961856842041017, 0.019776512145996093, 0.019936256408691407, 0.019756032943725587, 0.019734527587890623, 0.019952640533447266, 0.01979804801940918, 0.019649503707885742, 0.019636224746704102, 0.021346303939819337, 0.020783103942871094, 0.020749311447143554, 0.02089472007751465, 0.020787200927734374, 0.02067865562438965, 0.020617216110229493, 0.020555776596069338, 0.02063052749633789, 0.020785152435302736, 0.020738048553466795, 0.020744192123413087, 0.020651103973388672, 0.02063043212890625, 0.020768768310546876, 0.021725183486938478, 0.021118976593017577, 0.020783103942871094, 0.020553728103637696, 0.020529312133789064, 0.02051875114440918, 0.020531200408935548, 0.02065407943725586, 0.020583423614501953, 0.02067865562438965, 0.020592639923095703, 0.020633600234985353, 0.020558847427368163, 0.02064691162109375, 0.020653055191040038, 0.02059775924682617, 0.020684799194335936, 0.02063564872741699, 0.020549631118774413, 0.02067353630065918, 0.044382209777832034, 0.020682752609252928, 0.02066022491455078, 0.020503551483154296, 0.02061516761779785, 0.02062950325012207, 0.019954687118530275, 0.019971103668212892, 0.01992188835144043, 0.019942399978637695, 0.0198922233581543, 0.019920896530151368, 0.019803136825561524, 0.019735551834106444, 0.019708927154541016, 0.0198973445892334, 0.019952640533447266, 0.02002739143371582, 0.01987276840209961, 0.01987583923339844, 0.019942399978637695, 0.02000588798522949, 0.01986662483215332, 0.01998028755187988, 0.019954687118530275, 0.020093952178955078, 0.020123647689819335, 0.01988812828063965, 0.020187135696411132, 0.0206561279296875, 0.02066739273071289, 0.01986662483215332, 0.019742719650268553, 0.01983897590637207, 0.02004991912841797, 0.019983360290527344, 0.0198287353515625, 0.019923967361450197, 0.019941375732421874, 0.019999744415283204, 0.01982361602783203, 0.020876287460327148, 0.021183488845825195, 0.020563968658447264, 0.020563968658447264, 0.020624383926391602, 0.020797439575195312, 0.020732927322387695, 0.020773887634277344, 0.020715520858764647, 0.020717567443847656, 0.02079641532897949, 0.02068070411682129, 0.020709375381469726, 0.02068070411682129, 0.020802560806274413, 0.020727807998657227, 0.02145280075073242, 0.021909503936767577, 0.021165056228637694, 0.020758527755737305, 0.020809728622436522, 0.020805631637573242, 0.043633663177490234, 0.01987379264831543, 0.019883007049560548, 0.020031488418579102, 0.019974208831787108, 0.02038470458984375, 0.020025344848632814, 0.020137983322143553, 0.020101119995117187, 0.02005504035949707, 0.02002022361755371, 0.020048896789550782, 0.020025407791137696, 0.020079551696777345, 0.02007040023803711, 0.020115455627441405, 0.02000588798522949, 0.020074495315551756, 0.020127840042114258, 0.0200467529296875, 0.02003660774230957, 0.02002739143371582, 0.019998720169067383, 0.020212736129760742, 0.020157440185546875, 0.019935232162475586, 0.019979263305664064, 0.02002124786376953, 0.020039680480957032, 0.019886079788208007, 0.019989503860473632, 0.019861503601074217, 0.01979903984069824, 0.02000694465637207, 0.019782623291015624, 0.019973119735717772, 0.01985228729248047, 0.019896320343017578, 0.019900415420532228, 0.01983795166015625, 0.02001817512512207, 0.01984716796875, 0.019940351486206053, 0.019818496704101563, 0.01985638427734375, 0.01985443115234375, 0.019885984420776368, 0.01987993621826172, 0.019988576889038087, 0.01981328010559082, 0.01985228729248047, 0.019987455368041994, 0.01984614372253418, 0.019924991607666014, 0.019721216201782226, 0.020283456802368163, 0.019814336776733398, 0.019855392456054687, 0.019803104400634767, 0.01969049644470215, 0.019789920806884766, 0.019769247055053712, 0.019746816635131836, 0.04333158493041992, 0.01997209548950195, 0.02001817512512207, 0.0196945915222168, 0.01969254493713379, 0.01988198471069336, 0.01986457633972168, 0.019869695663452147, 0.019653728485107422, 0.019724191665649413, 0.019720191955566405, 0.019738624572753907, 0.01969049644470215, 0.019704832077026366, 0.019787776947021486, 0.01990656089782715, 0.01964236831665039, 0.019680416107177735, 0.01964531135559082, 0.019972063064575194, 0.019923967361450197, 0.019942399978637695, 0.019932159423828123, 0.02003455924987793, 0.019851264953613282, 0.019920896530151368, 0.019927040100097656, 0.020001792907714845, 0.01997209548950195, 0.020033536911010744, 0.019953664779663087, 0.019792896270751953, 0.019681280136108398, 0.019945472717285157, 0.01986867141723633, 0.019681280136108398, 0.019722240447998047, 0.019770368576049805, 0.019920896530151368, 0.02003455924987793, 0.019948543548583983, 0.019940351486206053, 0.02044927978515625, 0.019939327239990236, 0.019949567794799804, 0.019940383911132814, 0.0199270076751709, 0.02003660774230957, 0.019962879180908204, 0.019979263305664064, 0.01985843276977539, 0.019983360290527344, 0.019940351486206053, 0.019904512405395508, 0.019933183670043944, 0.020065280914306642, 0.01984000015258789, 0.019982336044311523, 0.019997695922851562, 0.019952640533447266, 0.019926015853881835, 0.020039680480957032, 0.019946495056152345, 0.04366233444213867, 0.020079679489135742, 0.020071359634399415, 0.019999744415283204, 0.020051008224487306, 0.020023231506347657, 0.02002943992614746, 0.02003763198852539, 0.019997695922851562, 0.020009983062744142, 0.02006630325317383, 0.02007040023803711, 0.01986867141723633, 0.019826688766479493, 0.019978239059448243, 0.02006630325317383, 0.019811328887939454, 0.01984614372253418, 0.019983360290527344, 0.019974143981933593, 0.020033567428588868, 0.020037599563598633, 0.02003660774230957, 0.02002943992614746, 0.0200898551940918, 0.02004787254333496, 0.020033536911010744, 0.01999667167663574, 0.020282367706298828, 0.019911680221557617, 0.02000387191772461, 0.020131807327270507, 0.02005299186706543, 0.020023359298706054, 0.020051904678344726, 0.01998236846923828, 0.020031455993652345, 0.02003046417236328, 0.02003660774230957, 0.020063232421875, 0.020135936737060548, 0.020084735870361328, 0.019990528106689453, 0.019984384536743165, 0.020059135437011717, 0.020060159683227538, 0.020011007308959963, 0.02012876892089844, 0.02001203155517578, 0.019947519302368166, 0.02005299186706543, 0.019979263305664064, 0.02009702491760254, 0.019969024658203126, 0.02004275131225586, 0.020065280914306642, 0.020376575469970702, 0.02172211265563965, 0.021393407821655275, 0.02124595260620117, 0.020942848205566408, 0.020891647338867187, 0.020988927841186524, 0.04510105514526367, 0.020916223526000977, 0.020926528930664063, 0.020947904586791993, 0.020888576507568358, 0.020970495223999023, 0.021120000839233398, 0.020977664947509765, 0.021032960891723632, 0.020953088760375976, 0.020884479522705078, 0.020926464080810548, 0.021393407821655275, 0.020914176940917968, 0.020915199279785156, 0.02107904052734375, 0.020968479156494142, 0.02096227264404297, 0.020883455276489257, 0.020953088760375976, 0.020938751220703124, 0.020899839401245117, 0.02090291213989258, 0.021133312225341795, 0.021147680282592774, 0.02110870361328125, 0.020958208084106447, 0.020939775466918945, 0.020932607650756836, 0.020970495223999023, 0.021127168655395507, 0.02112512016296387, 0.020987903594970703, 0.020958208084106447, 0.021003263473510742, 0.020988927841186524, 0.020931583404541015, 0.0206059513092041, 0.020609024047851563, 0.020945920944213867, 0.02104729652404785, 0.021313535690307618, 0.021115903854370118, 0.02106368064880371, 0.02098080062866211, 0.02092844772338867, 0.02101043128967285, 0.02104422378540039, 0.0210565128326416, 0.020954111099243163, 0.020989952087402345, 0.021073919296264648, 0.02084249687194824, 0.020896768569946288, 0.02110771179199219, 0.021196800231933592, 0.020963327407836914, 0.02091423988342285, 0.020849599838256835, 0.020975616455078124, 0.021371936798095702, 0.020995040893554688, 0.02101862335205078, 0.0438548469543457, 0.020142080307006836, 0.02010419273376465, 0.01984409523010254, 0.01984409523010254, 0.019984384536743165, 0.020109312057495117, 0.02012876892089844, 0.020737024307250978, 0.021004287719726563, 0.021003328323364257, 0.020971456527709962, 0.020948991775512696, 0.020928512573242186, 0.020600831985473633, 0.02068070411682129, 0.020880384445190428, 0.020953088760375976, 0.02088755226135254, 0.02005504035949707, 0.019975168228149414, 0.020067327499389647, 0.020024320602416993, 0.020115455627441405, 0.020032512664794923, 0.02000486373901367, 0.02004787254333496, 0.019998720169067383, 0.020039680480957032, 0.020050943374633787, 0.01987481689453125, 0.020271104812622072, 0.02029465675354004, 0.02002841567993164, 0.019802112579345704, 0.019793920516967774, 0.020607999801635742, 0.02085478401184082, 0.020904960632324218, 0.020916223526000977, 0.02006227111816406, 0.01998329544067383, 0.02008678436279297, 0.020065280914306642, 0.019986431121826173, 0.0200130558013916, 0.02045337677001953, 0.02003558349609375, 0.019869695663452147, 0.019973119735717772, 0.019720191955566405, 0.019794944763183595, 0.019565568923950196, 0.019907583236694337, 0.01999667167663574, 0.02007756805419922, 0.02106982421875, 0.02084249687194824, 0.02110771179199219, 0.02090598487854004, 0.02102681541442871, 0.021204992294311522, 0.02104422378540039, 0.043676673889160154, 0.02001919937133789, 0.02003868865966797, 0.019924959182739257, 0.019914751052856446, 0.01990656089782715, 0.019909664154052733, 0.020010976791381835, 0.01984614372253418, 0.019778560638427735, 0.019825664520263672, 0.019948543548583983, 0.019908607482910155, 0.01999564743041992, 0.019959808349609375, 0.020083711624145507, 0.020039680480957032, 0.019927040100097656, 0.019912704467773438, 0.019953664779663087, 0.019912704467773438, 0.019956735610961913, 0.019983360290527344, 0.019971071243286134, 0.01987993621826172, 0.02001817512512207, 0.019924991607666014, 0.020002815246582033, 0.019941375732421874, 0.019976192474365235, 0.01986764717102051, 0.020370431900024414, 0.019978239059448243, 0.019935232162475586, 0.019893247604370116, 0.019936256408691407, 0.019908607482910155, 0.019953664779663087, 0.019964927673339843, 0.02002124786376953, 0.019948543548583983, 0.019935232162475586, 0.01982464027404785, 0.019915775299072267, 0.019951616287231445, 0.019944448471069336, 0.019893247604370116, 0.019989503860473632, 0.019998720169067383, 0.020221952438354493, 0.02003455924987793, 0.019917823791503905, 0.01988812828063965, 0.019946495056152345, 0.019891199111938478, 0.020016128540039063, 0.020343807220458983, 0.019942399978637695, 0.019942399978637695, 0.019912704467773438, 0.019932159423828123, 0.019947519302368166, 0.019910655975341796, 0.04362547302246094, 0.02008166313171387, 0.020084735870361328, 0.020109312057495117, 0.02009600067138672, 0.02003046417236328, 0.019975168228149414, 0.02001817512512207, 0.020333568572998048, 0.02005504035949707, 0.019964927673339843, 0.019992576599121094, 0.019933183670043944, 0.019965951919555663, 0.019959808349609375, 0.02007244873046875, 0.020385791778564453, 0.02003046417236328, 0.019973184585571287, 0.01994643211364746, 0.01999564743041992, 0.02004684829711914, 0.019998720169067383, 0.021311487197875977, 0.02208768081665039, 0.021323776245117186, 0.02102783966064453, 0.021223424911499023, 0.021209087371826172, 0.02150297546386719, 0.020987903594970703, 0.02087116813659668, 0.020920320510864256, 0.020726783752441406, 0.02070425605773926, 0.020586496353149415, 0.020773887634277344, 0.02084556770324707, 0.020586496353149415, 0.020749311447143554, 0.020813823699951172, 0.020794368743896483, 0.020611072540283205, 0.020718591690063477, 0.02083020782470703, 0.020853887557983397, 0.02089561653137207, 0.020938751220703124, 0.020946943283081054, 0.02065100860595703, 0.020713632583618163, 0.020960159301757812, 0.02082195281982422, 0.020625408172607423, 0.02061414337158203, 0.020739072799682616, 0.020855808258056642, 0.020703231811523438, 0.02050048065185547, 0.020596736907958983, 0.020686847686767578, 0.02079641532897949, 0.020648128509521486, 0.0447393913269043, 0.02084659194946289, 0.02069708824157715, 0.020653055191040038, 0.02060492706298828, 0.020818944931030273, 0.02071244812011719, 0.020807680130004884, 0.02086809539794922, 0.020943872451782225, 0.020899839401245117, 0.021009408950805664, 0.020891647338867187, 0.02115897560119629, 0.020947904586791993, 0.020979711532592774, 0.020956159591674805, 0.02090598487854004, 0.02084659194946289, 0.020831232070922853, 0.02108723258972168, 0.022380544662475587, 0.02164838409423828, 0.021226495742797852, 0.021082111358642578, 0.021053440093994142, 0.02100223922729492, 0.020875263214111327, 0.020892671585083008, 0.020954111099243163, 0.021037055969238282, 0.020916223526000977, 0.020916223526000977, 0.02081996726989746, 0.02085990333557129, 0.021155839920043946, 0.021579776763916016, 0.021963775634765623, 0.021143552780151367, 0.021143552780151367, 0.021067808151245117, 0.02090595245361328, 0.020690944671630858, 0.020789247512817383, 0.020644863128662108, 0.020374528884887694, 0.020557823181152343, 0.02068889617919922, 0.02086297607421875, 0.02123673629760742, 0.02090598487854004, 0.02091007995605469, 0.02091007995605469, 0.021028863906860353, 0.02088652801513672, 0.020967424392700194, 0.020964351654052735, 0.021045248031616212, 0.020952064514160155, 0.020765695571899414, 0.020781055450439453, 0.020942848205566408, 0.020982784271240236]",tokens/s,48.32570284269545,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,togethercomputer/RedPajama-INCITE-Base-3B-v1,togethercomputer/RedPajama-INCITE-Base-3B-v1,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1277.698048,6833.045504,0.0,6186.5984,6021.011456,s,10,0.40395641326904297,0.0403956413269043,0.0014662159776033686,0.03990025520324707,0.040597081756591796,0.042691099929809566,0.04436631446838379,"[0.04478511810302734, 0.03988927841186524, 0.039835487365722656, 0.03995312118530273, 0.040131744384765626, 0.03986150360107422, 0.039911231994628905, 0.03985647964477539, 0.03975775909423828, 0.03997468948364258]",tokens/s,6337.317383534122,kWh,4.750828041343059e-07,2.6032226769420524e-07,2.4116094705622814e-06,3.1470145423907928e-06,tokens/kWh,81346938.99619426,MB,1277.698048,6833.045504,0.0,6186.5984,6148.783616,s,10,16.775238403320316,1.6775238403320312,0.013405465746111726,1.6762723388671876,1.6954143920898437,1.6963447692871094,1.6970890710449218,"[1.6952076416015625, 1.663684326171875, 1.678879638671875, 1.697275146484375, 1.6736650390625, 1.6575526123046875, 1.6889437255859374, 1.669037841796875, 1.687253662109375, 1.66373876953125]",tokens/s,37.55535300620851,kWh,1.9893780788410882e-05,1.0902027267439918e-05,6.349458214323924e-05,9.429039019909002e-05,tokens/kWh,668148.6826704001,,s,629,17.10622822189332,0.02719591132256488,0.0047852990235817995,0.02655948829650879,0.02738524169921875,0.02764349365234375,0.06645981170654297,"[0.02655948829650879, 0.027853824615478515, 0.02773504066467285, 0.027650047302246093, 0.027608064651489257, 0.026048511505126954, 0.026308671951293945, 0.02602694320678711, 0.02607411193847656, 0.026145792007446288, 0.026059776306152343, 0.02609561538696289, 0.02613657569885254, 0.02609459114074707, 0.02612428855895996, 0.02607411193847656, 0.02614169692993164, 0.026033151626586915, 0.026071039199829102, 0.02633932876586914, 0.02794495964050293, 0.028140607833862304, 0.028198848724365233, 0.027580415725708008, 0.027503616333007814, 0.027399168014526368, 0.027459583282470702, 0.027790336608886718, 0.028617727279663087, 0.028004352569580077, 0.027819007873535157, 0.027633663177490234, 0.02749235153198242, 0.0277258243560791, 0.027543552398681642, 0.027487232208251954, 0.0274913272857666, 0.028992511749267577, 0.027716608047485353, 0.027322399139404298, 0.02717795181274414, 0.027265024185180665, 0.027076608657836915, 0.027390975952148438, 0.02719436836242676, 0.02637414360046387, 0.02634035110473633, 0.026246143341064454, 0.026182655334472657, 0.026193920135498046, 0.026245119094848633, 0.02618880081176758, 0.026224639892578124, 0.026177536010742186, 0.026231807708740236, 0.026207231521606447, 0.026205184936523438, 0.02613043212890625, 0.02627174377441406, 0.02612633514404297, 0.02612428855895996, 0.02611404800415039, 0.06654054260253907, 0.02617241668701172, 0.026067968368530273, 0.0261529598236084, 0.02613657569885254, 0.026078208923339844, 0.025972736358642577, 0.0261529598236084, 0.026047487258911133, 0.026145792007446288, 0.02612019157409668, 0.02615500831604004, 0.026067968368530273, 0.026190912246704102, 0.025851839065551756, 0.026113023757934572, 0.026109952926635743, 0.026149887084960938, 0.026578943252563478, 0.02615705680847168, 0.026229759216308594, 0.02615705680847168, 0.025994239807128908, 0.026208255767822267, 0.026102848052978515, 0.026156991958618165, 0.026102848052978515, 0.026084287643432617, 0.02613248062133789, 0.02615193557739258, 0.02634547233581543, 0.026259456634521484, 0.02612735939025879, 0.026140703201293945, 0.02612220764160156, 0.02608230400085449, 0.02611507225036621, 0.02615603256225586, 0.02612428855895996, 0.02627686309814453, 0.026081279754638673, 0.02614067268371582, 0.026234880447387695, 0.026161151885986327, 0.02617651176452637, 0.02635468864440918, 0.027023359298706053, 0.027150335311889647, 0.027083776473999024, 0.027014144897460936, 0.027034624099731445, 0.02696396827697754, 0.02691481590270996, 0.026950656890869142, 0.026887168884277345, 0.02749951934814453, 0.02696396827697754, 0.02700492858886719, 0.02738380813598633, 0.02715545654296875, 0.02716979217529297, 0.027035648345947266, 0.027181055068969725, 0.06659584045410157, 0.02616012763977051, 0.026081279754638673, 0.026049535751342775, 0.026113023757934572, 0.026032127380371094, 0.026242048263549804, 0.0261079044342041, 0.026064895629882814, 0.026077184677124023, 0.026295295715332033, 0.02709401512145996, 0.028419071197509766, 0.02756403160095215, 0.0271779842376709, 0.027157503128051756, 0.026934272766113283, 0.027025407791137695, 0.02697420883178711, 0.027022335052490236, 0.026851327896118164, 0.027044864654541017, 0.026950656890869142, 0.026233856201171874, 0.026089471817016603, 0.02612735939025879, 0.02611814308166504, 0.02622265625, 0.026555328369140624, 0.026129407882690428, 0.026234880447387695, 0.026534912109375, 0.026228736877441407, 0.026158079147338868, 0.02612019157409668, 0.026184703826904295, 0.026104831695556642, 0.026161151885986327, 0.026102783203125, 0.02631679916381836, 0.02614681625366211, 0.026203136444091796, 0.0269752311706543, 0.02633932876586914, 0.0261212158203125, 0.026252288818359375, 0.02679193687438965, 0.027337728500366212, 0.027073535919189453, 0.027241472244262696, 0.027048959732055664, 0.02699673652648926, 0.027056127548217773, 0.027069440841674806, 0.027089920043945313, 0.027208703994750977, 0.027472896575927733, 0.02713804817199707, 0.02698854446411133, 0.027034624099731445, 0.027343904495239258, 0.027088991165161135, 0.027042688369750975, 0.06685593414306641, 0.026181631088256836, 0.026574848175048828, 0.02621446418762207, 0.026343360900878906, 0.026253311157226563, 0.026636287689208983, 0.028028928756713867, 0.027484159469604492, 0.027199487686157226, 0.027043840408325196, 0.027040767669677734, 0.027015167236328123, 0.027062271118164064, 0.026991615295410155, 0.02712985610961914, 0.02670694351196289, 0.02688204765319824, 0.02707967948913574, 0.027026432037353516, 0.026998783111572267, 0.02691993522644043, 0.027108352661132814, 0.027081727981567383, 0.026842111587524413, 0.02710527992248535, 0.026955808639526367, 0.027133920669555663, 0.027006975173950197, 0.027066368103027344, 0.026887168884277345, 0.02736128044128418, 0.02697216033935547, 0.026886144638061524, 0.02676633644104004, 0.02627993583679199, 0.02611814308166504, 0.02692095947265625, 0.027026432037353516, 0.027465728759765624, 0.02692095947265625, 0.027236352920532225, 0.02719436836242676, 0.027072511672973632, 0.02694963264465332, 0.02691584014892578, 0.02698137664794922, 0.0271329288482666, 0.027046911239624022, 0.02711859130859375, 0.027022335052490236, 0.026977279663085937, 0.02696294403076172, 0.02697932815551758, 0.026648576736450196, 0.02650111961364746, 0.026987520217895508, 0.02701312065124512, 0.026989599227905274, 0.026972127914428712, 0.026928127288818358, 0.027148288726806642, 0.026932287216186523, 0.06759212493896484, 0.027443199157714843, 0.028786687850952147, 0.027482112884521483, 0.026278911590576173, 0.026070016860961914, 0.026850303649902343, 0.02711244773864746, 0.02699673652648926, 0.02690355110168457, 0.026265600204467773, 0.026597375869750976, 0.02616422462463379, 0.02613043212890625, 0.026237951278686524, 0.026388479232788087, 0.02612224006652832, 0.026061824798583984, 0.02609971237182617, 0.02614067268371582, 0.02611609649658203, 0.02673971176147461, 0.027244543075561522, 0.02714726448059082, 0.02697113609313965, 0.026077184677124023, 0.026070016860961914, 0.02613862419128418, 0.026871807098388673, 0.02697420883178711, 0.026846208572387696, 0.026158208847045897, 0.02578112030029297, 0.025757696151733397, 0.026261503219604493, 0.02614271926879883, 0.026213375091552735, 0.025883647918701173, 0.026021888732910156, 0.026224639892578124, 0.026253311157226563, 0.02707967948913574, 0.02695782470703125, 0.026695680618286134, 0.026646528244018555, 0.026970111846923828, 0.027073535919189453, 0.026999807357788085, 0.02607513618469238, 0.026537984848022462, 0.026078208923339844, 0.02610688018798828, 0.026860544204711914, 0.02718003273010254, 0.026992639541625976, 0.027157503128051756, 0.02701312065124512, 0.027098175048828124, 0.026204095840454102, 0.026149887084960938, 0.026100736618041992, 0.02614067268371582, 0.02608742332458496, 0.0665374755859375, 0.026291200637817383, 0.026169343948364256, 0.026052608489990234, 0.02616422462463379, 0.02609868812561035, 0.02611609649658203, 0.026192895889282225, 0.026065919876098635, 0.026104831695556642, 0.026066944122314452, 0.026092544555664062, 0.026950656890869142, 0.027259904861450194, 0.026938400268554687, 0.027206623077392578, 0.026183679580688478, 0.026193920135498046, 0.026051584243774413, 0.02614476776123047, 0.02693734359741211, 0.027085823059082033, 0.0268984317779541, 0.027357183456420898, 0.026194944381713867, 0.026307584762573243, 0.02632601547241211, 0.02613248062133789, 0.026161151885986327, 0.026218496322631835, 0.026162176132202147, 0.026230783462524415, 0.026368000030517577, 0.026201087951660155, 0.02612633514404297, 0.0261529598236084, 0.02614067268371582, 0.02614374351501465, 0.026174463272094727, 0.026343423843383788, 0.026162176132202147, 0.026209280014038085, 0.026161151885986327, 0.026236928939819337, 0.0261212158203125, 0.026195968627929687, 0.026096639633178712, 0.026220544815063477, 0.02616422462463379, 0.02614886474609375, 0.026181631088256836, 0.026292224884033204, 0.02642848014831543, 0.02621536064147949, 0.026080255508422853, 0.02628812789916992, 0.026158079147338868, 0.027038719177246092, 0.027068479537963867, 0.026050495147705077, 0.025865215301513672, 0.02575257682800293, 0.025789440155029295, 0.06621286773681641, 0.025800703048706054, 0.025694208145141603, 0.02593791961669922, 0.02580998420715332, 0.028041151046752928, 0.02679910469055176, 0.026806272506713868, 0.026665023803710938, 0.02665260887145996, 0.026578943252563478, 0.026625024795532228, 0.026651647567749022, 0.02656972885131836, 0.026831903457641602, 0.026852319717407227, 0.027059200286865235, 0.028288000106811522, 0.028413951873779295, 0.027339775085449217, 0.027028480529785157, 0.026602495193481446, 0.026607616424560547, 0.027303936004638672, 0.027130880355834962, 0.02634854316711426, 0.026068992614746093, 0.026092544555664062, 0.026076160430908202, 0.02613657569885254, 0.026642431259155275, 0.027157503128051756, 0.027023391723632814, 0.027475936889648438, 0.02695577621459961, 0.02690662384033203, 0.027258880615234377, 0.02714726448059082, 0.02696294403076172, 0.027019264221191407, 0.027046911239624022, 0.02614067268371582, 0.026484735488891603, 0.026968063354492186, 0.02695884895324707, 0.02693017578125, 0.02692300796508789, 0.02676019287109375, 0.02620419120788574, 0.02695062446594238, 0.026917888641357423, 0.02693734359741211, 0.026917888641357423, 0.026656768798828126, 0.026102783203125, 0.02706329536437988, 0.026968063354492186, 0.027088895797729492, 0.027044864654541017, 0.027052032470703126, 0.02698240089416504, 0.027151424407958983, 0.02703046417236328, 0.06652518463134766, 0.026153984069824218, 0.026275840759277344, 0.026853376388549805, 0.02611814308166504, 0.026081279754638673, 0.026427391052246094, 0.026182655334472657, 0.026034175872802736, 0.02713599967956543, 0.0271278076171875, 0.026978303909301758, 0.02700595283508301, 0.026995712280273438, 0.02696396827697754, 0.027394048690795897, 0.02712166404724121, 0.026265600204467773, 0.02607513618469238, 0.02609459114074707, 0.026031103134155274, 0.026658847808837892, 0.027086816787719726, 0.027077632904052733, 0.02693939208984375, 0.02699673652648926, 0.02694655990600586, 0.02709708786010742, 0.02689638328552246, 0.026944511413574217, 0.02696601676940918, 0.02712678337097168, 0.026884096145629883, 0.027213823318481444, 0.02734284782409668, 0.027029504776000978, 0.02690764808654785, 0.025923583984375, 0.025806848526000976, 0.025986047744750978, 0.026599424362182617, 0.02641209602355957, 0.025870271682739258, 0.02596249580383301, 0.02614681625366211, 0.02591334342956543, 0.025873407363891602, 0.02675712013244629, 0.02677452850341797, 0.02670694351196289, 0.026373119354248048, 0.025790464401245116, 0.025803775787353517, 0.02611814308166504, 0.025787391662597657, 0.025838687896728517, 0.02570025634765625, 0.02576383972167969, 0.025697280883789062, 0.02569830322265625, 0.026408960342407226, 0.026635263442993166, 0.026616832733154298, 0.06629170989990234, 0.026816511154174806, 0.027838464736938476, 0.027184127807617187, 0.026908672332763672, 0.026985471725463867, 0.026983455657958986, 0.027108320236206053, 0.026953727722167968, 0.026870784759521486, 0.02694041633605957, 0.02699673652648926, 0.026266624450683593, 0.025991167068481445, 0.026020864486694335, 0.02615705680847168, 0.02611199951171875, 0.026462207794189452, 0.026033151626586915, 0.026073087692260744, 0.026169343948364256, 0.02631372833251953, 0.025963520050048827, 0.026177536010742186, 0.026202144622802733, 0.026151903152465822, 0.02588876724243164, 0.026195968627929687, 0.02594508743286133, 0.02609049606323242, 0.026089471817016603, 0.026216447830200194, 0.02611712074279785, 0.026057727813720705, 0.02612019157409668, 0.027418624877929686, 0.027226112365722657, 0.026995712280273438, 0.027062271118164064, 0.027158527374267577, 0.027069440841674806, 0.02697318458557129, 0.027040767669677734, 0.027035648345947266, 0.027032575607299804, 0.026205184936523438, 0.026651647567749022, 0.027650047302246093, 0.027554815292358398, 0.02774527931213379, 0.027448320388793947, 0.027617279052734374, 0.027463680267333986, 0.027595775604248047, 0.02733568000793457, 0.027616256713867186, 0.027443199157714843, 0.027535360336303712, 0.026703872680664063, 0.026471424102783202, 0.027470848083496095, 0.02755072021484375, 0.027291648864746092, 0.06661939239501953, 0.02613862419128418, 0.026062847137451172, 0.026185728073120116, 0.02612531280517578, 0.026066944122314452, 0.02608639907836914, 0.026218496322631835, 0.026102783203125, 0.02618880081176758, 0.02693120002746582, 0.02617651176452637, 0.026102783203125, 0.02611712074279785, 0.026081279754638673, 0.026335231781005858, 0.02613043212890625, 0.026171392440795898, 0.026158079147338868, 0.026243104934692382, 0.026123231887817382, 0.02615910339355469, 0.026169343948364256, 0.026242048263549804, 0.0261079044342041, 0.02615500831604004, 0.026085376739501953, 0.0265031681060791, 0.026059776306152343, 0.02613555145263672, 0.02611404800415039, 0.026221567153930665, 0.026193920135498046, 0.02612224006652832, 0.02618880081176758, 0.02611609649658203, 0.02637824058532715, 0.02612838363647461, 0.026018815994262694, 0.02627276802062988, 0.026045440673828125, 0.026088447570800782, 0.026072063446044923, 0.026050559997558592, 0.02611199951171875, 0.026595327377319337, 0.02711142349243164, 0.027081727981567383, 0.02697216033935547, 0.02698041534423828, 0.026919872283935546, 0.026916864395141602, 0.027007999420166014, 0.027059200286865235, 0.02674278450012207, 0.026210304260253905, 0.02672947120666504, 0.028011520385742186, 0.027218975067138673, 0.02719024085998535, 0.026825727462768553, 0.027047935485839843, 0.027024383544921874]",tokens/s,36.770233147888085,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,meta-llama/Meta-Llama-3-8B,meta-llama/Meta-Llama-3-8B,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.1,,,,MB,1263.742976,17201.364992,0.0,16554.917888,16436.726784,s,10,0.890863655090332,0.0890863655090332,0.00034962353177301416,0.08893444824218749,0.08931087188720703,0.08969057312011719,0.08999433410644532,"[0.09007027435302735, 0.08880348968505859, 0.08893727874755859, 0.08889881896972657, 0.0889316177368164, 0.08922649383544921, 0.08911043548583984, 0.08887769317626953, 0.08909187316894532, 0.08891567993164062]",tokens/s,2873.6159404105674,kWh,1.0535098971284362e-06,5.77273418059878e-07,6.912751351238865e-06,8.54353466642718e-06,tokens/kWh,29964178.761512138,MB,1263.742976,17201.364992,0.0,16554.917888,16436.729344,s,10,24.990497314453123,2.499049731445312,0.0012980572105728522,2.498893798828125,2.50089658203125,2.5012145263671877,2.5014688818359376,"[2.501532470703125, 2.499459228515625, 2.500825927734375, 2.497968017578125, 2.4987412109375, 2.49904638671875, 2.4997724609375, 2.4980390625, 2.49777197265625, 2.497340576171875]",tokens/s,25.20958234935336,kWh,2.944295109695625e-05,1.6135472700804864e-05,0.00013050602525036111,0.00017608444904812222,tokens/kWh,357782.8725964477,,s,629,25.75105219650269,0.04093966962878011,0.010627582788789067,0.039654399871826174,0.03975680160522461,0.03981680603027343,0.12905442810058593,"[0.03960108947753906, 0.039521278381347655, 0.03961139297485351, 0.039640064239501956, 0.03968819046020508, 0.03962265777587891, 0.03961958312988281, 0.03955916976928711, 0.03959603118896484, 0.03950899124145508, 0.03959807968139648, 0.0396308479309082, 0.03990937423706055, 0.03995443344116211, 0.03988172912597656, 0.03960934448242188, 0.03968307113647461, 0.03970048141479492, 0.03963702392578125, 0.03968406295776367, 0.039739391326904294, 0.03958169555664062, 0.03974246215820312, 0.03958476638793945, 0.039657470703125, 0.0396728630065918, 0.03965951919555664, 0.039643104553222654, 0.039723007202148435, 0.03961958312988281, 0.0397127685546875, 0.03969228744506836, 0.03966566467285156, 0.039672832489013675, 0.039656448364257815, 0.03964825439453125, 0.03969638442993164, 0.03965235137939453, 0.039947265625, 0.03988787078857422, 0.03977523040771484, 0.03968204879760742, 0.03974860763549805, 0.03972710418701172, 0.03970560073852539, 0.03982643127441406, 0.03975270462036133, 0.03966777420043945, 0.03964307022094726, 0.03969740676879883, 0.03967385482788086, 0.03973017501831055, 0.0397127685546875, 0.0397209587097168, 0.039741439819335936, 0.03969126510620117, 0.03975167846679688, 0.0397496337890625, 0.039769119262695315, 0.03970659255981445, 0.03968204879760742, 0.03971686553955078, 0.12888780212402343, 0.039570430755615234, 0.03951411056518555, 0.03956633758544922, 0.039476287841796874, 0.03960006332397461, 0.0395601921081543, 0.03960940933227539, 0.03952124786376953, 0.039593952178955075, 0.03955507278442383, 0.03957964706420898, 0.039610366821289066, 0.03961446380615234, 0.03950284957885742, 0.03960627365112305, 0.039588863372802735, 0.03962572860717774, 0.03962265777587891, 0.03969843292236328, 0.039626750946044925, 0.03967385482788086, 0.03960934448242188, 0.03970764923095703, 0.03972608184814453, 0.039626750946044925, 0.0396759033203125, 0.03972915267944336, 0.03962777709960937, 0.03965542221069336, 0.039769088745117184, 0.03966156768798828, 0.03974246215820312, 0.03970457458496094, 0.03970969772338867, 0.03967180633544922, 0.03972198486328125, 0.03969126510620117, 0.03966668701171875, 0.03967488098144531, 0.039654399871826174, 0.03974662399291992, 0.039645118713378905, 0.039762943267822266, 0.039795711517333986, 0.0396943359375, 0.03968511962890625, 0.03974553680419922, 0.039669761657714846, 0.039777278900146484, 0.03967180633544922, 0.03971072006225586, 0.03964313507080078, 0.03967488098144531, 0.03972710418701172, 0.039847934722900394, 0.03972710418701172, 0.03976499176025391, 0.039831584930419925, 0.03971376037597656, 0.03978854370117187, 0.039702529907226565, 0.039723007202148435, 0.1290997772216797, 0.03962879943847656, 0.03953868865966797, 0.03963391876220703, 0.03960319900512695, 0.039654399871826174, 0.03953664016723633, 0.039626750946044925, 0.03960115051269531, 0.039605247497558595, 0.03953049468994141, 0.039597057342529295, 0.03953049468994141, 0.03964723205566406, 0.03954483032226563, 0.039602176666259765, 0.03952742385864258, 0.03971788787841797, 0.03954483032226563, 0.039636993408203126, 0.039669761657714846, 0.039636993408203126, 0.03960422515869141, 0.03962265777587891, 0.0395786247253418, 0.03968102264404297, 0.03969843292236328, 0.039728126525878905, 0.03968000030517578, 0.03967488098144531, 0.03968716812133789, 0.03974758529663086, 0.03968204879760742, 0.03974041748046875, 0.039672832489013675, 0.039638015747070314, 0.03973017501831055, 0.039656448364257815, 0.039624702453613284, 0.039744510650634765, 0.03959807968139648, 0.0396943359375, 0.039739391326904294, 0.03971072006225586, 0.039702529907226565, 0.03981107330322266, 0.03973529434204102, 0.039806976318359374, 0.03963596725463867, 0.03966668701171875, 0.03971993637084961, 0.03969843292236328, 0.03968511962890625, 0.03979980850219727, 0.03968000030517578, 0.039793663024902344, 0.03981824111938476, 0.03969023895263672, 0.03973222351074219, 0.03975680160522461, 0.039927806854248044, 0.040325119018554685, 0.040123390197753905, 0.12911001586914062, 0.03958784103393555, 0.039577598571777346, 0.039556095123291016, 0.03955814361572266, 0.03956531143188476, 0.0395417594909668, 0.0395601921081543, 0.03946700668334961, 0.039667713165283204, 0.03950694274902344, 0.03960422515869141, 0.03961958312988281, 0.03960115051269531, 0.03953664016723633, 0.039605247497558595, 0.039531520843505856, 0.03958476638793945, 0.03955507278442383, 0.0396124153137207, 0.039613441467285154, 0.03964825439453125, 0.0395489273071289, 0.03962777709960937, 0.03960115051269531, 0.0396308479309082, 0.03960319900512695, 0.039774208068847655, 0.039657470703125, 0.0396759033203125, 0.03965235137939453, 0.0397209587097168, 0.03960422515869141, 0.03963391876220703, 0.039731201171875, 0.03970355224609375, 0.039646209716796874, 0.0396759033203125, 0.03965030288696289, 0.039672832489013675, 0.03960115051269531, 0.039653377532958986, 0.039640064239501956, 0.039714881896972656, 0.039704513549804685, 0.03968204879760742, 0.03964416122436523, 0.03967692947387695, 0.03968511962890625, 0.039657470703125, 0.03964416122436523, 0.03968307113647461, 0.039634944915771485, 0.03968716812133789, 0.03969843292236328, 0.0396943359375, 0.03973017501831055, 0.03972710418701172, 0.03973324966430664, 0.039725055694580076, 0.03978854370117187, 0.03977318572998047, 0.03973734283447266, 0.12907621765136718, 0.039583744049072264, 0.03955712127685547, 0.03956531143188476, 0.03947315216064453, 0.03967180633544922, 0.03974457550048828, 0.03964921569824219, 0.03949772644042969, 0.03958272171020508, 0.03955510330200195, 0.039582687377929686, 0.03953868865966797, 0.03958681488037109, 0.039553024291992187, 0.03955507278442383, 0.0395489273071289, 0.03958272171020508, 0.039588863372802735, 0.039640064239501956, 0.03958681488037109, 0.03961446380615234, 0.039523326873779296, 0.0396308479309082, 0.03957145690917969, 0.03962879943847656, 0.03961753463745117, 0.039605247497558595, 0.039629825592041014, 0.03969331359863281, 0.039836673736572265, 0.03978137588500977, 0.03963904190063477, 0.03965849685668945, 0.03961446380615234, 0.03963391876220703, 0.03965542221069336, 0.039641120910644534, 0.039669727325439455, 0.039669761657714846, 0.03968102264404297, 0.0397127685546875, 0.03970355224609375, 0.039689216613769535, 0.03969331359863281, 0.039792640686035156, 0.03969638442993164, 0.039708671569824217, 0.03970764923095703, 0.039711742401123046, 0.03971583938598633, 0.03971686553955078, 0.03974246215820312, 0.03970150375366211, 0.03975167846679688, 0.03976192092895508, 0.039800830841064457, 0.03977830505371094, 0.03975167846679688, 0.039728126525878905, 0.039754753112792966, 0.03966566467285156, 0.03969747161865234, 0.12940179443359376, 0.03962265777587891, 0.03956838226318359, 0.03960627365112305, 0.039616512298583983, 0.03964313507080078, 0.03951411056518555, 0.03967795181274414, 0.039651329040527344, 0.039613441467285154, 0.039553024291992187, 0.039570430755615234, 0.039534591674804685, 0.039621631622314454, 0.03966873550415039, 0.03967289733886719, 0.039676864624023436, 0.03965542221069336, 0.03954073715209961, 0.03957452774047852, 0.039556129455566406, 0.03960726547241211, 0.03956531143188476, 0.03969740676879883, 0.03958070373535156, 0.03962569427490235, 0.039602176666259765, 0.03963391876220703, 0.039583744049072264, 0.039608318328857424, 0.03961753463745117, 0.03966873550415039, 0.03960115051269531, 0.039640064239501956, 0.03959091186523438, 0.039640064239501956, 0.03964108657836914, 0.039654399871826174, 0.03968102264404297, 0.03969331359863281, 0.03962572860717774, 0.03969228744506836, 0.03971686553955078, 0.039684097290039064, 0.0397127685546875, 0.03970150375366211, 0.03969228744506836, 0.039753791809082034, 0.03971372985839844, 0.03973632049560547, 0.03973529434204102, 0.039667713165283204, 0.03973427200317383, 0.039678974151611326, 0.03974860763549805, 0.03998207855224609, 0.03974553680419922, 0.03971686553955078, 0.039880702972412106, 0.03973529434204102, 0.039771137237548826, 0.0397209587097168, 0.039711742401123046, 0.12928822326660155, 0.039576545715332034, 0.03951513671875, 0.03955916976928711, 0.03950694274902344, 0.039556095123291016, 0.03950387191772461, 0.03962575912475586, 0.03955401611328125, 0.03962060928344727, 0.03956121444702149, 0.03962572860717774, 0.03957555389404297, 0.039564289093017575, 0.03954687881469727, 0.039580673217773435, 0.03954278564453125, 0.039602176666259765, 0.03955814361572266, 0.039640064239501956, 0.0395489273071289, 0.03962572860717774, 0.03951001739501953, 0.03967692947387695, 0.039564289093017575, 0.03964416122436523, 0.03960319900512695, 0.039602176666259765, 0.03961958312988281, 0.03961753463745117, 0.039623680114746096, 0.0396544303894043, 0.039628768920898436, 0.03972198486328125, 0.03975680160522461, 0.039723007202148435, 0.03973324966430664, 0.03974348831176758, 0.03969126510620117, 0.03968307113647461, 0.0396124153137207, 0.03966873550415039, 0.039689216613769535, 0.03970048141479492, 0.039686145782470705, 0.03966156768798828, 0.03962777709960937, 0.03968102264404297, 0.03962060928344727, 0.03964108657836914, 0.03966979217529297, 0.03970966339111328, 0.040124481201171874, 0.040067008972167965, 0.04043468856811523, 0.039927806854248044, 0.039809024810791016, 0.03981721496582031, 0.03973222351074219, 0.03972198486328125, 0.03972198486328125, 0.03975065612792969, 0.03977318572998047, 0.12899839782714845, 0.03954073715209961, 0.03950592041015625, 0.03952435302734375, 0.03949977493286133, 0.03951615905761719, 0.039504894256591795, 0.039591934204101564, 0.0395417594909668, 0.03958476638793945, 0.039580673217773435, 0.03957657623291016, 0.03954687881469727, 0.03958784103393555, 0.03954483032226563, 0.03959603118896484, 0.039613441467285154, 0.039600128173828124, 0.03955814361572266, 0.039626785278320316, 0.03960009765625, 0.039666751861572265, 0.03956934356689453, 0.039607295989990236, 0.03956531143188476, 0.03961958312988281, 0.03958272171020508, 0.03960224151611328, 0.039588863372802735, 0.0396901741027832, 0.03961958312988281, 0.0396492805480957, 0.039593982696533206, 0.03965030288696289, 0.039618560791015625, 0.039634944915771485, 0.0396308479309082, 0.03963289642333984, 0.039626750946044925, 0.03965235137939453, 0.03962879943847656, 0.0396492805480957, 0.039682079315185546, 0.039717857360839846, 0.03970048141479492, 0.03973836898803711, 0.039654399871826174, 0.039734336853027345, 0.03965433502197266, 0.039684097290039064, 0.039772159576416014, 0.03981619262695312, 0.039880702972412106, 0.03973017501831055, 0.03972403335571289, 0.0397209587097168, 0.03973324966430664, 0.03967692947387695, 0.039728126525878905, 0.03970355224609375, 0.0396943359375, 0.03974758529663086, 0.040041473388671874, 0.12932505798339844, 0.039547904968261716, 0.039495681762695314, 0.039537662506103514, 0.039547904968261716, 0.039564289093017575, 0.03955712127685547, 0.03956838226318359, 0.03950592041015625, 0.03958476638793945, 0.03955814361572266, 0.03962777709960937, 0.039504894256591795, 0.039564289093017575, 0.039537662506103514, 0.03958784103393555, 0.039547904968261716, 0.039613441467285154, 0.039556095123291016, 0.03957452774047852, 0.03953664016723633, 0.039706623077392575, 0.039863296508789066, 0.03963603210449219, 0.039559104919433596, 0.03961548614501953, 0.03958272171020508, 0.039599105834960936, 0.03960115051269531, 0.03964108657836914, 0.03964108657836914, 0.03965030288696289, 0.039626750946044925, 0.03964416122436523, 0.03962777709960937, 0.03966156768798828, 0.03966259384155273, 0.03973529434204102, 0.039613441467285154, 0.03968819046020508, 0.03964518356323242, 0.03965951919555664, 0.0396943359375, 0.03973632049560547, 0.03972608184814453, 0.03964723205566406, 0.03969740676879883, 0.03972915267944336, 0.03967488098144531, 0.039711742401123046, 0.039661598205566403, 0.03968713760375977, 0.03971481704711914, 0.03971072006225586, 0.03974655914306641, 0.0397916145324707, 0.03973836898803711, 0.03968819046020508, 0.03971072006225586, 0.03966668701171875, 0.03969945526123047, 0.03970560073852539, 0.03975167846679688, 0.12911001586914062, 0.039570430755615234, 0.03951001739501953, 0.03953664016723633, 0.03953561782836914, 0.03956326293945313, 0.03956531143188476, 0.039626750946044925, 0.0395601921081543, 0.03961548614501953, 0.039518207550048826, 0.039572479248046875, 0.03952230453491211, 0.039613441467285154, 0.03954995346069336, 0.03964416122436523, 0.03958784103393555, 0.039597057342529295, 0.0395601921081543, 0.03956326293945313, 0.039550975799560545, 0.039605247497558595, 0.03956326293945313, 0.03961139297485351, 0.039556095123291016, 0.03960627365112305, 0.039600128173828124, 0.039616512298583983, 0.03959091186523438, 0.039618560791015625, 0.039572479248046875, 0.03960319900512695, 0.03959091186523438, 0.03963187026977539, 0.03960627365112305, 0.039667713165283204, 0.039624702453613284, 0.039656448364257815, 0.0396308479309082, 0.03966463851928711, 0.03964108657836914, 0.03965235137939453, 0.03965849685668945, 0.03971379089355469, 0.0396759033203125, 0.03965849685668945, 0.039656448364257815, 0.03974553680419922, 0.03966156768798828, 0.03971686553955078, 0.03975884628295898, 0.03972608184814453, 0.03969331359863281, 0.03969023895263672, 0.03978137588500977, 0.039711742401123046, 0.039795711517333986, 0.03971072006225586, 0.0397209587097168, 0.039624702453613284, 0.03974246215820312, 0.039695358276367186, 0.03975987243652344]",tokens/s,24.426186363189693,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen2-beta-14B,Qwen/Qwen2-beta-14B,cuda,0,42,,,True,,,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",qwen2,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-32B,Qwen/Qwen1.5-32B,cuda,0,42,,,True,,,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",qwen2,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,0,0,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/0/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66932595-4af43d5848c4a553352769d1;3ab87e0f-b25b-45e4-a606-9900fb3c128d) Repository Not Found for url: https://huggingface.co/0/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: 0 is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/gpt-neo-125m,EleutherAI/gpt-neo-125m,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1207.689216,1084.751872,0.0,438.304768,353.979904,s,17,0.1775642251968384,0.010444954423343436,0.00041006191246230724,0.010357760429382324,0.010782880020141602,0.011423027610778808,0.011592356224060059,"[0.01163468837738037, 0.011370112419128418, 0.010391391754150391, 0.01036076831817627, 0.01028329563140869, 0.010322560310363769, 0.010367615699768067, 0.01034067153930664, 0.010280287742614747, 0.01028012752532959, 0.010357760429382324, 0.01034438419342041, 0.01037059211730957, 0.010366592407226562, 0.010364128112792968, 0.009800288200378418, 0.010328960418701173]",tokens/s,24509.44155657256,kWh,1.1949881906255474e-07,6.547973377102067e-08,2.910551704913972e-07,4.7603372332497257e-07,tokens/kWh,537777025.988634,MB,1207.689216,1084.751872,0.0,438.304768,363.872768,s,17,10.183319030761718,0.5990187665153952,0.007991533929472642,0.5971812133789063,0.599423583984375,0.6060201171875,0.6255276562500001,"[0.630404541015625, 0.597092041015625, 0.597417724609375, 0.5979285888671875, 0.5971812133789063, 0.5999240112304688, 0.5970659790039062, 0.5967833251953125, 0.5980140380859374, 0.59759423828125, 0.5965198364257812, 0.5990899658203125, 0.5978482055664063, 0.5964898071289062, 0.5953594360351563, 0.5926844482421875, 0.595921630859375]",tokens/s,105.17199714206426,kWh,6.682860514682202e-06,3.6619119296741874e-06,1.1577596299731033e-05,2.192236874408742e-05,tokens/kWh,2873777.041862387,,s,1070,10.338569228172302,0.00966221423193673,0.0013272748192884368,0.009465855598449707,0.009597030067443849,0.009966205024719238,0.02029690891265869,"[0.010437631607055664, 0.010578944206237792, 0.010528767585754394, 0.010050559997558594, 0.009961471557617188, 0.010003456115722656, 0.009928704261779785, 0.009888768196105957, 0.009944064140319824, 0.01013759994506836, 0.010068991661071777, 0.009967616081237793, 0.009994239807128906, 0.01009663963317871, 0.010127360343933106, 0.010090496063232422, 0.01002086353302002, 0.010019840240478516, 0.010027008056640625, 0.009927680015563965, 0.009962495803833007, 0.00992460823059082, 0.009980928421020508, 0.009963520050048828, 0.010156031608581542, 0.010176511764526367, 0.01004748821258545, 0.009904128074645996, 0.01012019157409668, 0.010148863792419433, 0.010065919876098632, 0.009891839981079101, 0.0100065279006958, 0.01015500831604004, 0.009936896324157715, 0.00991436767578125, 0.010004480361938477, 0.009866304397583008, 0.009964480400085449, 0.010094592094421387, 0.010203136444091796, 0.01011513614654541, 0.010010560035705567, 0.009942015647888184, 0.010085375785827636, 0.009850879669189454, 0.009788415908813476, 0.010160127639770507, 0.009851903915405273, 0.009558015823364258, 0.009540608406066894, 0.009775103569030762, 0.010049535751342773, 0.009944064140319824, 0.010030079841613769, 0.009971712112426758, 0.00972390365600586, 0.009871359825134277, 0.009644031524658203, 0.009713664054870605, 0.009495552062988282, 0.009676799774169922, 0.02169753646850586, 0.009498623847961426, 0.009517056465148926, 0.009443327903747559, 0.009510944366455078, 0.0094934720993042, 0.009451519966125489, 0.009449472427368164, 0.009451519966125489, 0.009424896240234374, 0.009492480278015136, 0.009474047660827637, 0.00946281623840332, 0.009483231544494628, 0.009462783813476563, 0.009500672340393066, 0.009470975875854493, 0.009409536361694336, 0.009450495719909668, 0.009499648094177245, 0.009395199775695801, 0.009438207626342773, 0.009457663536071777, 0.009464832305908203, 0.009466879844665528, 0.00954265594482422, 0.009481216430664062, 0.009449472427368164, 0.009441280364990234, 0.009445376396179199, 0.009457663536071777, 0.009449472427368164, 0.009485376358032227, 0.009478079795837402, 0.009442303657531738, 0.009439231872558594, 0.009476096153259277, 0.009462783813476563, 0.00943513584136963, 0.009485312461853027, 0.009481216430664062, 0.00948639965057373, 0.009424832344055176, 0.009505855560302734, 0.009447360038757324, 0.009417728424072265, 0.009549823760986328, 0.009450495719909668, 0.009499648094177245, 0.009458687782287598, 0.009474047660827637, 0.009469951629638672, 0.009441280364990234, 0.009449472427368164, 0.009477120399475097, 0.009444352149963378, 0.009463808059692384, 0.009451552391052246, 0.009483231544494628, 0.009432064056396485, 0.00951296043395996, 0.009469951629638672, 0.009475071907043458, 0.020369407653808593, 0.009447423934936524, 0.00949350357055664, 0.009455615997314454, 0.009437184333801269, 0.009487360000610352, 0.009489407539367676, 0.009457663536071777, 0.009537535667419434, 0.009475071907043458, 0.009520159721374512, 0.009514975547790527, 0.009541631698608399, 0.009487360000610352, 0.009492480278015136, 0.009498623847961426, 0.009477120399475097, 0.009474047660827637, 0.009453568458557129, 0.009476096153259277, 0.009464832305908203, 0.009433088302612304, 0.00943513584136963, 0.00952627182006836, 0.009468928337097168, 0.009506815910339356, 0.009497599601745605, 0.009495583534240723, 0.00943612766265869, 0.009510911941528321, 0.009477120399475097, 0.00947612762451172, 0.009442272186279297, 0.009470975875854493, 0.009457663536071777, 0.009523200035095216, 0.009391103744506836, 0.009495552062988282, 0.009388031959533692, 0.009456640243530273, 0.009498623847961426, 0.009502719879150391, 0.009490431785583496, 0.009572352409362793, 0.009443327903747559, 0.009445376396179199, 0.00951193618774414, 0.009475071907043458, 0.009486335754394531, 0.009472064018249512, 0.009454527854919433, 0.009498623847961426, 0.009491456031799317, 0.009430015563964844, 0.009450495719909668, 0.009441280364990234, 0.009458687782287598, 0.009459744453430176, 0.009417695999145508, 0.009453568458557129, 0.009485312461853027, 0.009446399688720703, 0.009527296066284179, 0.020470783233642577, 0.009524224281311035, 0.009502719879150391, 0.00949350357055664, 0.009417728424072265, 0.009481216430664062, 0.009480192184448242, 0.009465855598449707, 0.009499648094177245, 0.009459712028503419, 0.009568256378173828, 0.009463808059692384, 0.009564224243164063, 0.009484224319458008, 0.009505791664123535, 0.009479167938232422, 0.009467904090881347, 0.009491456031799317, 0.009537535667419434, 0.00949350357055664, 0.009444352149963378, 0.009491456031799317, 0.009453568458557129, 0.009459712028503419, 0.009446399688720703, 0.009495552062988282, 0.009382911682128906, 0.009578495979309083, 0.009607168197631836, 0.009431039810180664, 0.00949766445159912, 0.00941158390045166, 0.009469887733459473, 0.009382911682128906, 0.009481216430664062, 0.009475071907043458, 0.009404416084289552, 0.009448448181152343, 0.009537535667419434, 0.009460736274719238, 0.00943513584136963, 0.009444352149963378, 0.00942080020904541, 0.009455615997314454, 0.00961945629119873, 0.009494527816772461, 0.009453568458557129, 0.009443327903747559, 0.009454591751098633, 0.009483263969421387, 0.00942080020904541, 0.009498623847961426, 0.009476160049438477, 0.009464768409729004, 0.009640959739685059, 0.009480192184448242, 0.00950169563293457, 0.00942182445526123, 0.009438207626342773, 0.009506879806518554, 0.00960095977783203, 0.009482239723205567, 0.009476096153259277, 0.020271104812622072, 0.009445376396179199, 0.009499648094177245, 0.009458687782287598, 0.009466879844665528, 0.009486335754394531, 0.009509951591491698, 0.009512895584106446, 0.009517056465148926, 0.009447423934936524, 0.00953446388244629, 0.00971987247467041, 0.009644991874694824, 0.009634816169738769, 0.00944438362121582, 0.0094934720993042, 0.009481247901916503, 0.00945251178741455, 0.009510911941528321, 0.009495552062988282, 0.009466879844665528, 0.009475071907043458, 0.009234432220458985, 0.009309184074401856, 0.009243647575378417, 0.009232383728027344, 0.009338879585266113, 0.009430015563964844, 0.009593855857849122, 0.009484288215637206, 0.009480192184448242, 0.009458687782287598, 0.009481216430664062, 0.009464832305908203, 0.009446399688720703, 0.009482239723205567, 0.009439264297485352, 0.009445343971252441, 0.00950374412536621, 0.009474047660827637, 0.009473024368286133, 0.009485312461853027, 0.009532416343688965, 0.009524224281311035, 0.009459712028503419, 0.009434111595153808, 0.009480192184448242, 0.00941875171661377, 0.00942182445526123, 0.009497599601745605, 0.009447423934936524, 0.009502719879150391, 0.009448448181152343, 0.009489407539367676, 0.009466879844665528, 0.009522175788879395, 0.009447423934936524, 0.009495552062988282, 0.00942796802520752, 0.009598976135253906, 0.00951807975769043, 0.00941977596282959, 0.009464832305908203, 0.020348928451538087, 0.009458687782287598, 0.009441280364990234, 0.009548800468444824, 0.00950374412536621, 0.00951807975769043, 0.009457663536071777, 0.009458687782287598, 0.009446399688720703, 0.009517056465148926, 0.009482239723205567, 0.009648127555847168, 0.009488384246826171, 0.009487360000610352, 0.009432064056396485, 0.009472000122070312, 0.009513024330139161, 0.009577407836914062, 0.00941158390045166, 0.009476096153259277, 0.009494527816772461, 0.009553919792175293, 0.009456640243530273, 0.009506815910339356, 0.009440256118774413, 0.009444352149963378, 0.00941260814666748, 0.009565183639526367, 0.009452544212341308, 0.009521151542663574, 0.009517056465148926, 0.010577919960021973, 0.009673760414123536, 0.009525216102600098, 0.010172415733337402, 0.009739263534545899, 0.009517056465148926, 0.009500672340393066, 0.009523200035095216, 0.00952012825012207, 0.009444352149963378, 0.009485312461853027, 0.00941977596282959, 0.009389056205749511, 0.009390080451965332, 0.009445376396179199, 0.00941055965423584, 0.009465855598449707, 0.00950169563293457, 0.00950374412536621, 0.009438207626342773, 0.009443327903747559, 0.009475071907043458, 0.009466879844665528, 0.009486335754394531, 0.009490431785583496, 0.009516032218933105, 0.009424896240234374, 0.009504768371582031, 0.009473024368286133, 0.009479167938232422, 0.009437184333801269, 0.009396224021911622, 0.020342784881591795, 0.009452544212341308, 0.0094269437789917, 0.009461759567260742, 0.009467904090881347, 0.009457663536071777, 0.0094269437789917, 0.00945257568359375, 0.009399264335632325, 0.009447423934936524, 0.009430015563964844, 0.009446399688720703, 0.009455648422241211, 0.009457632064819336, 0.00979251194000244, 0.009538559913635255, 0.009437184333801269, 0.009572352409362793, 0.00942080020904541, 0.009388031959533692, 0.009453568458557129, 0.009414655685424805, 0.009423871994018555, 0.009567232131958007, 0.009483263969421387, 0.009414655685424805, 0.009434111595153808, 0.00941260814666748, 0.009447423934936524, 0.009470975875854493, 0.009459712028503419, 0.009516032218933105, 0.009440256118774413, 0.00941055965423584, 0.009492480278015136, 0.00941977596282959, 0.009455648422241211, 0.009454560279846192, 0.009532416343688965, 0.009482239723205567, 0.009447423934936524, 0.00950374412536621, 0.009499648094177245, 0.009445376396179199, 0.009468928337097168, 0.009782272338867188, 0.009615360260009765, 0.009543680191040039, 0.009477120399475097, 0.009472000122070312, 0.009455615997314454, 0.00942182445526123, 0.009439231872558594, 0.009522208213806152, 0.009443296432495116, 0.009465855598449707, 0.009439231872558594, 0.009408512115478516, 0.009482239723205567, 0.00939724826812744, 0.009415679931640625, 0.00942182445526123, 0.009437184333801269, 0.020327423095703127, 0.009489407539367676, 0.009466879844665528, 0.009424896240234374, 0.009423871994018555, 0.009446399688720703, 0.009478143692016602, 0.009484288215637206, 0.009491456031799317, 0.009430015563964844, 0.009507840156555175, 0.00941875171661377, 0.009547776222229003, 0.009475071907043458, 0.009470975875854493, 0.009468928337097168, 0.009545727729797364, 0.009488384246826171, 0.009545727729797364, 0.009431039810180664, 0.009446399688720703, 0.009444352149963378, 0.00941055965423584, 0.009443327903747559, 0.009454591751098633, 0.009478143692016602, 0.009433088302612304, 0.009425919532775879, 0.00943513584136963, 0.009465855598449707, 0.009494527816772461, 0.009457663536071777, 0.009468928337097168, 0.009417728424072265, 0.009395199775695801, 0.009452544212341308, 0.009416735649108887, 0.009453536033630371, 0.00940544033050537, 0.009559040069580077, 0.009438207626342773, 0.00941158390045166, 0.009416704177856445, 0.00953651237487793, 0.009442303657531738, 0.009483263969421387, 0.00943513584136963, 0.009477120399475097, 0.009540608406066894, 0.009529408454895019, 0.009396160125732422, 0.009462783813476563, 0.009474047660827637, 0.009466879844665528, 0.00942796802520752, 0.009457695960998535, 0.009521120071411133, 0.009463808059692384, 0.009445376396179199, 0.009604096412658691, 0.00942182445526123, 0.009564160346984863, 0.009431039810180664, 0.020282367706298828, 0.00950374412536621, 0.009488384246826171, 0.009472000122070312, 0.009484288215637206, 0.00951296043395996, 0.009431039810180664, 0.0095098876953125, 0.009469951629638672, 0.00950169563293457, 0.00951807975769043, 0.009566207885742188, 0.009432064056396485, 0.009569279670715332, 0.009543680191040039, 0.009531392097473144, 0.009487360000610352, 0.00952627182006836, 0.009505791664123535, 0.009569279670715332, 0.009474047660827637, 0.009531392097473144, 0.00961023998260498, 0.009508864402770996, 0.00951193618774414, 0.009494527816772461, 0.009462783813476563, 0.009401344299316406, 0.009478143692016602, 0.009463808059692384, 0.009425919532775879, 0.009569279670715332, 0.009480192184448242, 0.00962662410736084, 0.009472000122070312, 0.009457663536071777, 0.009498623847961426, 0.009482239723205567, 0.00942796802520752, 0.009462847709655761, 0.009452544212341308, 0.00943404769897461, 0.00942899227142334, 0.009437184333801269, 0.00942182445526123, 0.009481216430664062, 0.009439231872558594, 0.009456640243530273, 0.009455615997314454, 0.009475071907043458, 0.009507840156555175, 0.00943616008758545, 0.009423871994018555, 0.009461759567260742, 0.009567232131958007, 0.009441280364990234, 0.009467904090881347, 0.009464832305908203, 0.009416735649108887, 0.00958460807800293, 0.009461759567260742, 0.009399328231811523, 0.009469920158386231, 0.02023219108581543, 0.00942908763885498, 0.009558943748474122, 0.009455648422241211, 0.009418720245361329, 0.009456640243530273, 0.009552895545959473, 0.009441280364990234, 0.009449472427368164, 0.009450495719909668, 0.00950169563293457, 0.009537535667419434, 0.009417728424072265, 0.009468928337097168, 0.009488384246826171, 0.009487360000610352, 0.009529439926147461, 0.009444255828857422, 0.009448448181152343, 0.009439231872558594, 0.009496576309204101, 0.009555968284606933, 0.009477120399475097, 0.009560064315795898, 0.009537535667419434, 0.009498623847961426, 0.009517056465148926, 0.009852928161621094, 0.009539584159851074, 0.0094136323928833, 0.009441280364990234, 0.009458687782287598, 0.009452544212341308, 0.00941875171661377, 0.009434111595153808, 0.009451519966125489, 0.009441280364990234, 0.009490431785583496, 0.00952627182006836, 0.009447423934936524, 0.00941158390045166, 0.009433088302612304, 0.009468928337097168, 0.009478143692016602, 0.009417728424072265, 0.009438207626342773, 0.00941977596282959, 0.009438207626342773, 0.00950374412536621, 0.009476096153259277, 0.009393152236938476, 0.009480256080627442, 0.009565152168273926, 0.009457632064819336, 0.009439231872558594, 0.009499648094177245, 0.009469951629638672, 0.009476096153259277, 0.009433088302612304, 0.009522175788879395, 0.009448448181152343, 0.009540608406066894, 0.009447423934936524, 0.020348928451538087, 0.009424896240234374, 0.009468928337097168, 0.00950169563293457, 0.009432064056396485, 0.009464832305908203, 0.009448448181152343, 0.009477120399475097, 0.009453568458557129, 0.009444352149963378, 0.009486335754394531, 0.009454591751098633, 0.009430015563964844, 0.009506815910339356, 0.009445376396179199, 0.009500672340393066, 0.009473024368286133, 0.009423871994018555, 0.009478143692016602, 0.009416704177856445, 0.009438207626342773, 0.009438207626342773, 0.009449472427368164, 0.009423871994018555, 0.009587712287902832, 0.009502719879150391, 0.009559040069580077, 0.009424896240234374, 0.009478143692016602, 0.009491456031799317, 0.009434111595153808, 0.009469951629638672, 0.009478143692016602, 0.009398271560668945, 0.009452544212341308, 0.009476096153259277, 0.00942899227142334, 0.009446399688720703, 0.009456640243530273, 0.009440256118774413, 0.009443327903747559, 0.009463808059692384, 0.009392160415649414, 0.009455583572387695, 0.009492480278015136, 0.009492480278015136, 0.009598976135253906, 0.009521151542663574, 0.009415679931640625, 0.009468928337097168, 0.009453568458557129, 0.009472000122070312, 0.009464832305908203, 0.009433088302612304, 0.009517056465148926, 0.009433088302612304, 0.009529343605041504, 0.009442303657531738, 0.009414655685424805, 0.00940339183807373, 0.009439231872558594, 0.009372672080993653, 0.009453568458557129, 0.020380672454833985, 0.009483263969421387, 0.00980684757232666, 0.009596927642822266, 0.009891839981079101, 0.0101212158203125, 0.00951807975769043, 0.009530367851257325, 0.009448448181152343, 0.00941260814666748, 0.009549823760986328, 0.009447423934936524, 0.009491519927978516, 0.009491392135620117, 0.009445376396179199, 0.009532416343688965, 0.00942905616760254, 0.009396160125732422, 0.009422847747802734, 0.009401344299316406, 0.009563136100769042, 0.009552895545959473, 0.009425919532775879, 0.009476096153259277, 0.009457663536071777, 0.009404416084289552, 0.00942899227142334, 0.009443327903747559, 0.009455615997314454, 0.009440256118774413, 0.00943513584136963, 0.009554944038391112, 0.009558015823364258, 0.00942182445526123, 0.009453568458557129, 0.009495552062988282, 0.00943513584136963, 0.009479167938232422, 0.009414655685424805, 0.009578495979309083, 0.009553919792175293, 0.00942796802520752, 0.009443327903747559, 0.009466879844665528, 0.00942899227142334, 0.009653247833251954, 0.009462783813476563, 0.009488384246826171, 0.00951296043395996, 0.009516032218933105, 0.009469951629638672, 0.009494527816772461, 0.009439231872558594, 0.009464832305908203, 0.009433088302612304, 0.009469951629638672, 0.009450495719909668, 0.009623552322387695, 0.009473024368286133, 0.00955084800720215, 0.009442336082458496, 0.009506784439086913, 0.009422847747802734, 0.020344831466674804, 0.009442303657531738, 0.00951296043395996, 0.009517056465148926, 0.009461759567260742, 0.009547776222229003, 0.009546751976013184, 0.009488384246826171, 0.009440256118774413, 0.009465855598449707, 0.009497599601745605, 0.009434111595153808, 0.009444352149963378, 0.009483263969421387, 0.009442303657531738, 0.00950374412536621, 0.00953446388244629, 0.009467904090881347, 0.009453568458557129, 0.009454591751098633, 0.009478143692016602, 0.009481280326843262, 0.009440192222595215, 0.00941977596282959, 0.0094136323928833, 0.00942796802520752, 0.009459712028503419, 0.009430015563964844, 0.009433088302612304, 0.009434111595153808, 0.009369600296020507, 0.009508864402770996, 0.009498623847961426, 0.009439231872558594, 0.009482239723205567, 0.009451519966125489, 0.009446399688720703, 0.0094136323928833, 0.009449472427368164, 0.009457663536071777, 0.009478143692016602, 0.009486335754394531, 0.009432064056396485, 0.00951193618774414, 0.00943513584136963, 0.009551872253417968, 0.009456640243530273, 0.0094269437789917, 0.009244671821594238, 0.009285632133483887, 0.009755647659301758, 0.009795583724975587, 0.010049535751342773, 0.009558015823364258, 0.009445376396179199, 0.009463808059692384, 0.009504768371582031, 0.009526304244995117, 0.00946070384979248, 0.009548800468444824, 0.009485312461853027, 0.009487360000610352, 0.009443391799926757, 0.02037958335876465, 0.00951193618774414, 0.009464832305908203, 0.009450495719909668, 0.00942796802520752, 0.009460736274719238, 0.009460736274719238, 0.009468928337097168, 0.009454591751098633, 0.009437184333801269, 0.009450495719909668, 0.009444352149963378, 0.009469951629638672, 0.009459712028503419, 0.009443327903747559, 0.009492480278015136, 0.009480192184448242, 0.009432064056396485, 0.009478143692016602, 0.009432064056396485, 0.009487360000610352, 0.009510911941528321, 0.009431039810180664, 0.009559040069580077, 0.009447423934936524, 0.009448448181152343, 0.009448448181152343, 0.009443327903747559, 0.009444352149963378, 0.009454591751098633, 0.009434111595153808, 0.009667584419250488, 0.009476096153259277, 0.00941875171661377, 0.00943513584136963, 0.009462783813476563, 0.009406463623046875, 0.009461759567260742, 0.009430047988891601, 0.009449440002441406, 0.009440256118774413, 0.00952627182006836, 0.00947004795074463, 0.009432991981506348, 0.00952012825012207, 0.009470015525817872, 0.009416640281677246, 0.00951807975769043, 0.009468928337097168, 0.009448479652404786, 0.009405407905578613, 0.009443391799926757, 0.009424832344055176, 0.009472000122070312, 0.009431039810180664, 0.009416704177856445, 0.00943513584136963, 0.009423871994018555, 0.009468928337097168, 0.00958364772796631, 0.009396191596984863, 0.00942905616760254, 0.009448384284973145, 0.020290559768676757, 0.00943616008758545, 0.009430015563964844, 0.009565183639526367, 0.009425919532775879, 0.009459712028503419, 0.009417728424072265, 0.009458687782287598, 0.009531392097473144, 0.009396224021911622, 0.009458687782287598, 0.00943513584136963, 0.009416704177856445, 0.009462783813476563, 0.009406463623046875, 0.009445376396179199, 0.00942182445526123, 0.00941055965423584, 0.009441280364990234, 0.009583616256713867, 0.00941977596282959, 0.009438207626342773, 0.009431039810180664, 0.009453568458557129, 0.009439231872558594, 0.009433088302612304, 0.00941158390045166, 0.009432064056396485, 0.009417759895324706, 0.009417695999145508, 0.009338879585266113, 0.009241600036621094, 0.00922214412689209, 0.009263104438781738, 0.0094136323928833, 0.009434111595153808, 0.009450495719909668, 0.00951296043395996, 0.009462783813476563, 0.00951193618774414, 0.009465920448303222, 0.00951084804534912, 0.009467904090881347, 0.00940544033050537, 0.00941977596282959, 0.009449472427368164, 0.00952012825012207, 0.009461759567260742, 0.009417728424072265, 0.009437184333801269, 0.009556991577148437, 0.009464832305908203, 0.009450495719909668, 0.009488384246826171, 0.009490431785583496, 0.009451519966125489, 0.00941977596282959, 0.00943513584136963, 0.009527296066284179, 0.009433088302612304, 0.009561087608337402, 0.009423871994018555, 0.009456640243530273, 0.019224576950073242, 0.008946687698364257, 0.009009152412414552, 0.008961024284362793, 0.008932352066040039, 0.008953856468200684, 0.008957951545715333, 0.00892416000366211, 0.009243647575378417, 0.009431039810180664, 0.009489407539367676, 0.009456640243530273, 0.009461759567260742, 0.009459712028503419, 0.009453568458557129, 0.009558015823364258, 0.009445376396179199, 0.009409536361694336, 0.0094269437789917, 0.00942080020904541, 0.009447423934936524, 0.009474080085754394, 0.009432031631469726, 0.009488384246826171, 0.009437184333801269, 0.009417728424072265, 0.009442303657531738, 0.009510911941528321, 0.009531392097473144, 0.009452544212341308, 0.00942796802520752, 0.009431039810180664, 0.009797632217407226, 0.009575424194335937, 0.00952012825012207, 0.009449472427368164, 0.009546751976013184, 0.009497599601745605, 0.009465855598449707, 0.009469951629638672, 0.009381952285766602, 0.009400256156921387, 0.00951296043395996, 0.009494527816772461, 0.009440256118774413, 0.009486335754394531, 0.009433088302612304, 0.009476096153259277, 0.009469951629638672, 0.009454591751098633, 0.009472000122070312, 0.009483263969421387, 0.009445376396179199, 0.009568256378173828, 0.009454591751098633, 0.009385984420776367, 0.009425919532775879, 0.009445376396179199, 0.009457663536071777, 0.009524224281311035, 0.009381888389587402, 0.009510911941528321, 0.009417728424072265, 0.020311040878295897, 0.0094136962890625, 0.009492416381835938, 0.009468928337097168, 0.009406463623046875, 0.009529343605041504, 0.009395199775695801, 0.009439295768737792, 0.009384896278381348, 0.009449472427368164, 0.009473024368286133, 0.009479167938232422, 0.00952627182006836, 0.009479167938232422, 0.009448448181152343, 0.009450495719909668, 0.009468928337097168, 0.009423871994018555, 0.009393152236938476, 0.009442303657531738, 0.009393152236938476, 0.009441280364990234, 0.009404416084289552, 0.009533439636230469, 0.00953446388244629, 0.0095098876953125, 0.009415679931640625, 0.009454591751098633, 0.009433088302612304, 0.009438207626342773, 0.009481216430664062, 0.009444352149963378, 0.009467904090881347, 0.009453568458557129, 0.009443327903747559, 0.009449472427368164, 0.00943616008758545, 0.00935529613494873, 0.00943824005126953, 0.009484224319458008, 0.009380864143371583, 0.009480192184448242, 0.009380864143371583, 0.009581567764282227, 0.009439231872558594, 0.009532416343688965, 0.009597951889038087, 0.00950380802154541, 0.009456576347351075, 0.00951296043395996, 0.009446399688720703, 0.009496576309204101, 0.009432064056396485, 0.00941055965423584, 0.00942899227142334, 0.00941260814666748, 0.009361408233642577, 0.009443360328674317, 0.009434111595153808, 0.009389023780822753, 0.00945462417602539, 0.00938595199584961, 0.00940339183807373]",tokens/s,103.4959457527528,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,2,2,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/2/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669324c8-746a2a6c06e1fd6a38f78d76;8cc48f04-fed4-4fbd-b13b-e56fd032d33b) Repository Not Found for url: https://huggingface.co/2/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: 2 is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/xglm-564M,,cuda,0,42,,,,,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,d35829e539df8480b726c647eeabf91e41eae047,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,,MB,1388.036096,2403.86048,0.0,1757.413376,1301.782528,s,12,0.20144134140014652,0.016786778450012207,0.0018214050335381856,0.016153103828430174,0.016703455734252932,0.019434478092193598,0.02210406023025513,"[0.022771455764770507, 0.016551008224487306, 0.016201311111450196, 0.0167042236328125, 0.016696544647216798, 0.016167776107788086, 0.016049343109130858, 0.016016319274902345, 0.016138431549072265, 0.01601785659790039, 0.01609654426574707, 0.016030527114868163]",tokens/s,15250.097019050958,kWh,1.921774259858446e-07,1.0530401582644383e-07,6.138130475090691e-07,9.112944893213575e-07,tokens/kWh,280919069.5212517,MB,1388.036096,2403.86048,0.0,1757.413376,1324.7616,s,12,10.558549926757813,0.8798791605631511,0.010784227070228428,0.8786428222656251,0.8903626342773437,0.8984366760253907,0.9055257141113281,"[0.9072979736328125, 0.881484375, 0.8828522338867187, 0.8911865234375, 0.8829476318359375, 0.882841552734375, 0.8734462890625, 0.8682412719726562, 0.8711069946289063, 0.875488037109375, 0.8658557739257813, 0.87580126953125]",tokens/s,71.60074112867723,kWh,1.0095053256607961e-05,5.531636060229464e-06,2.1067384852988094e-05,3.669407416982551e-05,tokens/kWh,1716898.4754439325,,s,755,10.721403898239132,0.014200534964555147,0.0020255489759926147,0.013850624084472657,0.014336614418029785,0.014493081760406494,0.030158765869140625,"[0.01530777645111084, 0.01447935962677002, 0.015130656242370605, 0.014796768188476563, 0.0142807035446167, 0.01449062442779541, 0.014376959800720214, 0.014313471794128419, 0.014429183959960937, 0.014726143836975097, 0.014943231582641601, 0.015244288444519043, 0.014939135551452636, 0.014836735725402832, 0.015038463592529297, 0.015326208114624023, 0.014498815536499024, 0.014320639610290528, 0.01439948844909668, 0.014312447547912598, 0.014470144271850586, 0.01447219181060791, 0.014910464286804198, 0.014941184043884278, 0.014276608467102051, 0.01438822364807129, 0.014529536247253418, 0.014299136161804199, 0.01425715160369873, 0.01417728042602539, 0.014098431587219238, 0.014183423995971679, 0.01399507236480713, 0.013927359580993653, 0.014554112434387208, 0.014814208030700684, 0.014774271965026856, 0.014413824081420898, 0.014178303718566895, 0.014191679954528809, 0.014287808418273925, 0.014317567825317384, 0.014322688102722168, 0.014212096214294433, 0.014258175849914552, 0.014268416404724121, 0.01360588836669922, 0.013647871971130371, 0.013599743843078613, 0.013652992248535157, 0.013706303596496583, 0.013746111869812012, 0.013850624084472657, 0.013767680168151856, 0.015051775932312012, 0.014500864028930664, 0.014235679626464843, 0.01431446361541748, 0.014199808120727539, 0.014208000183105468, 0.013628416061401367, 0.01368166446685791, 0.031204351425170897, 0.014247936248779297, 0.013598719596862792, 0.013565952301025391, 0.013685759544372558, 0.013692928314208984, 0.013604928016662597, 0.013604800224304199, 0.013643775939941406, 0.014348287582397461, 0.014272512435913086, 0.013604864120483399, 0.014402591705322265, 0.014265312194824218, 0.014206975936889648, 0.014162943840026856, 0.014297087669372559, 0.014167039871215821, 0.01365503978729248, 0.013606911659240722, 0.014262271881103515, 0.014229503631591797, 0.014305279731750489, 0.014249983787536622, 0.014173184394836426, 0.01420083236694336, 0.014223360061645507, 0.014201855659484864, 0.014272512435913086, 0.014284799575805664, 0.014422016143798828, 0.01416806411743164, 0.014263296127319336, 0.0135731201171875, 0.01365503978729248, 0.013587455749511718, 0.013649920463562011, 0.013632512092590332, 0.01364684772491455, 0.013639679908752441, 0.013790207862854004, 0.014238719940185546, 0.01425715160369873, 0.014249983787536622, 0.014197759628295899, 0.014195712089538574, 0.01376972770690918, 0.013666303634643554, 0.013640704154968262, 0.013647871971130371, 0.013629440307617188, 0.014131199836730958, 0.014272512435913086, 0.014150655746459961, 0.01417728042602539, 0.014129152297973633, 0.013595647811889648, 0.013619199752807617, 0.013720576286315917, 0.014217215538024902, 0.01416806411743164, 0.014461952209472656, 0.013668352127075196, 0.030331903457641602, 0.013631487846374512, 0.013629440307617188, 0.013670399665832519, 0.013591551780700683, 0.014339072227478027, 0.014220288276672363, 0.014690303802490234, 0.015069184303283691, 0.014460927963256836, 0.014255104064941406, 0.014277631759643555, 0.01417625617980957, 0.014187520027160644, 0.013637696266174316, 0.013637568473815918, 0.014225407600402832, 0.014327839851379394, 0.01365500831604004, 0.013674495697021484, 0.013644800186157227, 0.013644800186157227, 0.013643775939941406, 0.014126079559326172, 0.01430835247039795, 0.01399295997619629, 0.014062591552734375, 0.013521920204162598, 0.013940735816955567, 0.013748224258422852, 0.013991968154907226, 0.01404412841796875, 0.014261247634887696, 0.014225407600402832, 0.014316543579101563, 0.014173215866088867, 0.014258144378662109, 0.014189567565917969, 0.014412799835205077, 0.014314496040344238, 0.014173184394836426, 0.01419878387451172, 0.014213120460510254, 0.014160896301269531, 0.01365401554107666, 0.013683712005615235, 0.013677568435668945, 0.013714431762695312, 0.013792256355285644, 0.014287872314453125, 0.014243840217590332, 0.01421827220916748, 0.01370518398284912, 0.01368883228302002, 0.013612031936645508, 0.013674495697021484, 0.013647871971130371, 0.013618176460266113, 0.013612031936645508, 0.014309375762939454, 0.014212096214294433, 0.014179327964782714, 0.014226431846618653, 0.0314204158782959, 0.014199808120727539, 0.014220288276672363, 0.014219264030456542, 0.014256128311157227, 0.01428889560699463, 0.014225407600402832, 0.01435750389099121, 0.014276608467102051, 0.014249983787536622, 0.014249983787536622, 0.014217215538024902, 0.014322688102722168, 0.014245887756347657, 0.014218239784240723, 0.01438310432434082, 0.01417420768737793, 0.013648896217346192, 0.014229503631591797, 0.014286848068237304, 0.014244864463806153, 0.013723648071289063, 0.01367142391204834, 0.014216192245483398, 0.014236672401428223, 0.014240768432617188, 0.014216192245483398, 0.013658111572265624, 0.013657088279724122, 0.013592576026916504, 0.013749247550964355, 0.013691904067993164, 0.01364684772491455, 0.013683712005615235, 0.013666303634643554, 0.013718591690063477, 0.0141659517288208, 0.014210047721862793, 0.014279680252075195, 0.014277631759643555, 0.014208000183105468, 0.014228480339050293, 0.01425920009613037, 0.014231552124023437, 0.014268416404724121, 0.014223360061645507, 0.014389247894287109, 0.014271488189697265, 0.014260224342346191, 0.014178303718566895, 0.014152704238891601, 0.014183423995971679, 0.01417625617980957, 0.014181376457214356, 0.014199808120727539, 0.014248959541320801, 0.014325759887695312, 0.014284799575805664, 0.014260224342346191, 0.014287872314453125, 0.014338047981262207, 0.014202879905700684, 0.014384127616882325, 0.031308799743652346, 0.014306303977966308, 0.01426534366607666, 0.013439999580383302, 0.013478912353515626, 0.013599743843078613, 0.013618176460266113, 0.013650943756103515, 0.013716480255126954, 0.013647871971130371, 0.013629440307617188, 0.014221311569213867, 0.01417420768737793, 0.01406156826019287, 0.014043168067932129, 0.013950943946838379, 0.014063648223876953, 0.013977567672729492, 0.013945856094360352, 0.01419878387451172, 0.013645824432373046, 0.013700096130371094, 0.013670399665832519, 0.014195712089538574, 0.014187520027160644, 0.014510080337524415, 0.014345215797424317, 0.014197759628295899, 0.01366220760345459, 0.013649920463562011, 0.014160896301269531, 0.014219264030456542, 0.014231552124023437, 0.014477312088012695, 0.013919232368469238, 0.014194687843322755, 0.014226431846618653, 0.014301183700561524, 0.014233599662780762, 0.014213120460510254, 0.014244864463806153, 0.014254079818725587, 0.014317567825317384, 0.014449664115905762, 0.014379008293151856, 0.014263296127319336, 0.0142807035446167, 0.014256128311157227, 0.013820927619934082, 0.0136878080368042, 0.014114815711975098, 0.01409945583343506, 0.013516799926757812, 0.013850624084472657, 0.014088191986083985, 0.014215167999267577, 0.01427558422088623, 0.014180352210998535, 0.013632512092590332, 0.013623295783996582, 0.013634559631347656, 0.013656064033508301, 0.013607935905456543, 0.030935039520263673, 0.014208000183105468, 0.014180352210998535, 0.014253055572509766, 0.01417420768737793, 0.013635583877563476, 0.013657088279724122, 0.015430656433105469, 0.015126527786254883, 0.014407679557800293, 0.01425920009613037, 0.014338047981262207, 0.014346240043640136, 0.014218239784240723, 0.01425920009613037, 0.014226495742797851, 0.014233535766601562, 0.013649920463562011, 0.013616127967834473, 0.01363865566253662, 0.014310400009155273, 0.014270463943481445, 0.014194687843322755, 0.014149632453918457, 0.014183423995971679, 0.013806591987609864, 0.013742079734802246, 0.013614080429077148, 0.013703167915344238, 0.013717503547668456, 0.014044159889221192, 0.014831616401672363, 0.014458880424499512, 0.014185471534729004, 0.014240768432617188, 0.013615103721618652, 0.013802495956420899, 0.014393343925476074, 0.01428377628326416, 0.013690879821777344, 0.013684736251831055, 0.013622271537780761, 0.013670399665832519, 0.01368172836303711, 0.013761471748352051, 0.013664256095886231, 0.013685759544372558, 0.013701120376586913, 0.01365401554107666, 0.01363865566253662, 0.013682687759399414, 0.013670399665832519, 0.01360588836669922, 0.013599743843078613, 0.013617183685302734, 0.013633503913879394, 0.013599743843078613, 0.013612031936645508, 0.014169088363647461, 0.014277631759643555, 0.014226431846618653, 0.014316543579101563, 0.014204928398132324, 0.03015782356262207, 0.01360588836669922, 0.013601792335510255, 0.013599743843078613, 0.013565952301025391, 0.013614080429077148, 0.013577280044555664, 0.013657024383544922, 0.01358233642578125, 0.013652992248535157, 0.01375641632080078, 0.013717503547668456, 0.01368064022064209, 0.013667327880859375, 0.013674495697021484, 0.013629440307617188, 0.013698047637939453, 0.013602815628051757, 0.014050304412841797, 0.01427558422088623, 0.014402560234069824, 0.014203904151916504, 0.014337023735046387, 0.014303232192993164, 0.014318592071533203, 0.014615551948547363, 0.014318592071533203, 0.014310400009155273, 0.014237728118896485, 0.014213088035583497, 0.013505536079406738, 0.013540351867675781, 0.013527039527893067, 0.013573151588439941, 0.013600735664367676, 0.013645824432373046, 0.013622271537780761, 0.013652992248535157, 0.013608960151672364, 0.014220319747924804, 0.014303199768066406, 0.014193663597106934, 0.014064640045166015, 0.013477888107299805, 0.013487104415893555, 0.013794303894042969, 0.013661184310913087, 0.013634559631347656, 0.01364684772491455, 0.013668352127075196, 0.013629440307617188, 0.01362435245513916, 0.014179295539855957, 0.014215167999267577, 0.014258175849914552, 0.014204928398132324, 0.014169088363647461, 0.014298111915588378, 0.013817855834960938, 0.01368883228302002, 0.013667327880859375, 0.013576191902160644, 0.01416806411743164, 0.030083072662353515, 0.013622271537780761, 0.014060544013977052, 0.014247936248779297, 0.014435327529907227, 0.014224384307861328, 0.014233599662780762, 0.014221343994140625, 0.014329824447631836, 0.013616127967834473, 0.01363865566253662, 0.013666303634643554, 0.013587455749511718, 0.013658111572265624, 0.013622271537780761, 0.013698047637939453, 0.01386291217803955, 0.013703167915344238, 0.01375334358215332, 0.01366220760345459, 0.013701120376586913, 0.01359769630432129, 0.013666303634643554, 0.01359984016418457, 0.013728672027587891, 0.013637632369995116, 0.013912128448486329, 0.014220224380493165, 0.013942784309387207, 0.014271488189697265, 0.014307328224182129, 0.014325759887695312, 0.014211071968078613, 0.01364684772491455, 0.013686783790588379, 0.013742079734802246, 0.013689855575561523, 0.013633536338806153, 0.013645824432373046, 0.013659135818481445, 0.013619199752807617, 0.013651968002319336, 0.013618176460266113, 0.013648896217346192, 0.013620223999023438, 0.013603839874267578, 0.013676544189453126, 0.01364684772491455, 0.013727744102478028, 0.013616127967834473, 0.013603839874267578, 0.01356390380859375, 0.013496319770812988, 0.013639712333679199, 0.013682656288146973, 0.013608991622924805, 0.01363145637512207, 0.013659135818481445, 0.013632512092590332, 0.013635583877563476, 0.013735936164855958, 0.013604864120483399, 0.013619199752807617, 0.030159872055053712, 0.013609984397888183, 0.013615103721618652, 0.013629440307617188, 0.013584383964538574, 0.013620223999023438, 0.013579263687133789, 0.013433856010437012, 0.013420543670654296, 0.013407232284545899, 0.0135731201171875, 0.013987839698791504, 0.01365503978729248, 0.013612031936645508, 0.013619263648986817, 0.013640640258789062, 0.013607935905456543, 0.01387724781036377, 0.01372879981994629, 0.01372873592376709, 0.014499839782714843, 0.014344191551208496, 0.014252032279968262, 0.01425312042236328, 0.014243776321411133, 0.014188544273376465, 0.014292991638183594, 0.014255104064941406, 0.013705216407775878, 0.013643775939941406, 0.013640704154968262, 0.013613056182861329, 0.013643839836120605, 0.013640640258789062, 0.013670399665832519, 0.01387007999420166, 0.01420803165435791, 0.014275551795959473, 0.014336000442504883, 0.01428889560699463, 0.01426534366607666, 0.014190655708312988, 0.014327743530273437, 0.01419878387451172, 0.013657088279724122, 0.013631487846374512, 0.013620223999023438, 0.013670399665832519, 0.013636608123779297, 0.013714431762695312, 0.01366323184967041, 0.013694975852966309, 0.013689855575561523, 0.013676544189453126, 0.013649951934814454, 0.013602784156799316, 0.01375436782836914, 0.013659135818481445, 0.01363865566253662, 0.013717503547668456, 0.014311488151550294, 0.01368057632446289, 0.013658111572265624, 0.030123008728027343, 0.01368166446685791, 0.013651968002319336, 0.013615103721618652, 0.013749247550964355, 0.013619199752807617, 0.013647871971130371, 0.013658111572265624, 0.013707263946533203, 0.013661184310913087, 0.01365401554107666, 0.013629440307617188, 0.013656064033508301, 0.013661184310913087, 0.013643775939941406, 0.013637632369995116, 0.013649920463562011, 0.01367347240447998, 0.013657088279724122, 0.01370419216156006, 0.013734911918640137, 0.01367244815826416, 0.01368064022064209, 0.013628416061401367, 0.013707263946533203, 0.01366323184967041, 0.01366528034210205, 0.013628416061401367, 0.014185471534729004, 0.014222335815429688, 0.014215167999267577, 0.014255104064941406, 0.01417420768737793, 0.014271488189697265, 0.013685759544372558, 0.013706239700317382, 0.013648896217346192, 0.013668352127075196, 0.013624320030212403, 0.013608960151672364, 0.013578240394592284, 0.013747200012207032, 0.01365503978729248, 0.013722623825073242, 0.013987839698791504, 0.014213120460510254, 0.01425100803375244, 0.014137344360351562, 0.014299136161804199, 0.014337023735046387, 0.01425920009613037, 0.014270463943481445, 0.014346240043640136, 0.014315520286560059, 0.014189567565917969, 0.014203935623168946, 0.014187487602233886, 0.014215167999267577, 0.01425920009613037, 0.014254079818725587, 0.014341119766235352, 0.01415167999267578, 0.01430016040802002, 0.030241792678833007, 0.013626367568969726, 0.01365401554107666, 0.013703167915344238, 0.014215167999267577, 0.014222335815429688, 0.01425715160369873, 0.01445683193206787, 0.013668352127075196, 0.013657088279724122, 0.013627391815185547, 0.01367244815826416, 0.013740032196044923, 0.013615103721618652, 0.013641728401184081, 0.013612031936645508, 0.013742079734802246, 0.013636608123779297, 0.013692928314208984, 0.014064640045166015, 0.01364684772491455, 0.013651968002319336, 0.013629440307617188, 0.01363865566253662, 0.013596672058105469, 0.01359769630432129, 0.013606911659240722, 0.01358028793334961, 0.01387110424041748, 0.014248959541320801, 0.014155776023864745, 0.013717503547668456, 0.013712384223937989, 0.013628416061401367, 0.01363046360015869, 0.013611007690429687, 0.013658176422119141, 0.013610943794250489, 0.01359769630432129, 0.013645824432373046, 0.013634559631347656, 0.013583359718322753, 0.013633536338806153, 0.013720576286315917, 0.013658111572265624, 0.013586432456970214, 0.013724672317504882, 0.01365401554107666, 0.013603839874267578, 0.013724672317504882, 0.013636608123779297, 0.01368064022064209, 0.013641728401184081, 0.01367244815826416, 0.013617152214050294, 0.013708288192749024, 0.01363046360015869, 0.013660160064697266, 0.01417625617980957, 0.014216192245483398, 0.013643775939941406, 0.013763584136962891, 0.013627391815185547, 0.030679040908813477, 0.013649920463562011, 0.013645824432373046, 0.013636608123779297, 0.013667360305786132, 0.013649888038635254, 0.013593600273132325, 0.013640704154968262, 0.013733887672424316, 0.014215167999267577, 0.014227456092834472, 0.014219264030456542, 0.013622271537780761, 0.013590527534484862, 0.013858816146850587, 0.01365503978729248, 0.013595647811889648, 0.01359769630432129, 0.013619199752807617, 0.013619199752807617, 0.013639679908752441, 0.01426636791229248, 0.014191616058349609, 0.014276608467102051, 0.014207039833068848, 0.01432262420654297, 0.014256128311157227, 0.014217215538024902, 0.014189567565917969, 0.014227456092834472, 0.014179327964782714, 0.014205951690673829, 0.014188544273376465, 0.014179327964782714, 0.013637632369995116, 0.013684736251831055, 0.013616127967834473, 0.013583359718322753, 0.013660160064697266, 0.013599743843078613, 0.01363046360015869, 0.013608960151672364, 0.013620223999023438, 0.013674495697021484, 0.014307328224182129, 0.014224448204040528, 0.014213055610656737, 0.014164992332458496, 0.014185471534729004, 0.014183423995971679, 0.01419878387451172, 0.014219264030456542, 0.014228480339050293, 0.01416806411743164, 0.014319616317749024, 0.013636608123779297, 0.013627391815185547, 0.013645888328552246, 0.013614015579223633, 0.013584383964538574, 0.013618176460266113, 0.013590527534484862, 0.013558783531188966]",tokens/s,70.41988224359311,,,main,False,False,,,,,,,,,,,,,,,,,,,,,,, bfloat16-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-14B,Qwen/Qwen-14B,cuda,0,42,,,True,,,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",qwen,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,huggyllama/llama-7b,huggyllama/llama-7b,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.1,,,,MB,1252.605952,14529.593344,0.0,13883.14624,13705.186304,s,10,0.9879193649291993,0.09879193649291992,0.0002512661472968626,0.09873307418823242,0.09890953521728515,0.09920227279663087,0.09943646286010742,"[0.09949501037597656, 0.098844482421875, 0.0985115203857422, 0.09878438568115235, 0.09873267364501953, 0.09880620574951172, 0.09872038269042968, 0.09862966156005859, 0.09866156768798828, 0.09873347473144531]",tokens/s,2591.3046052938403,kWh,1.1688001729823925e-06,6.402306186400357e-07,6.192654736254908e-06,8.001685527877335e-06,tokens/kWh,31993259.308693547,MB,1252.605952,14529.593344,0.0,13883.14624,13856.22528,s,10,22.93839282226563,2.293839282226563,0.001022246015489629,2.2933924560546872,2.2952232177734375,2.295579504394531,2.295864533691406,"[2.29338037109375, 2.29463623046875, 2.2931396484375, 2.293404541015625, 2.292788818359375, 2.29299072265625, 2.294100830078125, 2.29514404296875, 2.292871826171875, 2.295935791015625]",tokens/s,27.464871008245936,kWh,2.6996926405915512e-05,1.4795383739971892e-05,0.000113753996776545,0.0001555463069224324,tokens/kWh,405024.0809086952,,s,629,23.789682689666723,0.03782143511870707,0.011804791302641934,0.03638886260986328,0.0365588493347168,0.036620288848876956,0.1356418212890625,"[0.0362158088684082, 0.03625574493408203, 0.03614617538452149, 0.03632025527954102, 0.03617587280273438, 0.03626393508911133, 0.03632230377197266, 0.03629056167602539, 0.03623321533203125, 0.0362608642578125, 0.03621068954467774, 0.03624345779418945, 0.036299774169921875, 0.03666841506958008, 0.036674560546875, 0.03648614501953125, 0.03627008056640625, 0.0363765754699707, 0.03635200119018555, 0.03639603042602539, 0.036313087463378906, 0.03638886260986328, 0.036324352264404294, 0.03636633682250977, 0.03628953552246094, 0.03641958236694336, 0.03631718444824219, 0.036400127410888675, 0.03633356857299805, 0.03639910507202149, 0.036311038970947264, 0.0363469123840332, 0.03631510543823242, 0.036397056579589845, 0.036337665557861325, 0.03641958236694336, 0.03633868789672851, 0.03642675018310547, 0.03638272094726563, 0.03640422439575195, 0.03632844924926758, 0.03643807983398437, 0.03634272003173828, 0.036393985748291016, 0.03638988876342773, 0.036620288848876956, 0.03661004638671875, 0.036512767791748044, 0.03643904113769531, 0.03642985534667969, 0.036356063842773435, 0.03649539184570313, 0.03638678359985351, 0.036670463562011715, 0.03657113647460938, 0.03648921585083008, 0.03639910507202149, 0.03650764846801758, 0.036375553131103515, 0.03649740982055664, 0.03641139221191406, 0.036580352783203124, 0.13603021240234375, 0.03622195053100586, 0.03626291275024414, 0.03629363250732422, 0.036350975036621096, 0.03625881576538086, 0.03630495834350586, 0.036385726928710935, 0.03644416046142578, 0.03632844924926758, 0.03636019134521484, 0.03621785736083984, 0.03639295959472656, 0.03627724838256836, 0.0363612174987793, 0.03628547286987305, 0.03642569732666016, 0.03630694580078125, 0.03640422439575195, 0.03630080032348633, 0.036431873321533206, 0.03633663940429688, 0.03640422439575195, 0.0362977294921875, 0.03634483337402344, 0.036307968139648435, 0.036501502990722655, 0.0362762222290039, 0.03638169479370117, 0.03633356857299805, 0.03647999954223633, 0.03634688186645508, 0.03637350463867187, 0.03633868789672851, 0.03646262359619141, 0.03634172821044922, 0.03641753768920898, 0.03635302352905274, 0.03654560089111328, 0.03641337585449219, 0.036514816284179685, 0.0364400634765625, 0.03647795104980469, 0.03643904113769531, 0.03655475234985352, 0.03644313430786133, 0.036620288848876956, 0.03646156692504883, 0.036450336456298825, 0.03637961578369141, 0.0365219841003418, 0.03645439910888672, 0.03650048065185547, 0.036410369873046876, 0.03666435241699219, 0.036446174621582034, 0.036575263977050784, 0.03641238403320313, 0.03650969696044922, 0.036408321380615234, 0.03663872146606445, 0.03671449661254883, 0.036563007354736325, 0.13556729125976563, 0.0362762222290039, 0.03630284881591797, 0.036229118347167966, 0.036287551879882814, 0.03623110580444336, 0.036383743286132815, 0.03626598358154297, 0.0362977294921875, 0.03626803207397461, 0.03638886260986328, 0.03626291275024414, 0.036291584014892575, 0.0362874870300293, 0.03635302352905274, 0.036357120513916014, 0.03630182266235352, 0.03627110290527344, 0.03639910507202149, 0.036318206787109376, 0.03638579177856445, 0.03629875183105469, 0.03647180938720703, 0.03640115356445312, 0.036370464324951175, 0.03629974365234375, 0.03641856002807617, 0.036359169006347655, 0.036400127410888675, 0.03635200119018555, 0.03640115356445312, 0.03631513595581055, 0.03639295959472656, 0.03635200119018555, 0.03638272094726563, 0.036318206787109376, 0.03638681411743164, 0.03632128143310547, 0.03639910507202149, 0.03638476943969727, 0.03640217590332031, 0.03635302352905274, 0.03644313430786133, 0.03645951843261719, 0.036450302124023434, 0.03639910507202149, 0.03645849609375, 0.036498432159423826, 0.036511745452880856, 0.03646771240234375, 0.03654553604125976, 0.036441089630126954, 0.03652403259277344, 0.03638169479370117, 0.036485118865966795, 0.036416511535644534, 0.036511745452880856, 0.03641958236694336, 0.03653222274780273, 0.03638886260986328, 0.03654348754882813, 0.036485118865966795, 0.036729854583740236, 0.13590835571289062, 0.036152320861816405, 0.0362158088684082, 0.03620761489868164, 0.036225025177001956, 0.03625574493408203, 0.03640217590332031, 0.03624857711791992, 0.03629056167602539, 0.03624857711791992, 0.036345855712890625, 0.03624038314819336, 0.03643392181396484, 0.03623833465576172, 0.0362874870300293, 0.03629056167602539, 0.03628441619873047, 0.03624345779418945, 0.036383743286132815, 0.03624857711791992, 0.03634175872802734, 0.036359169006347655, 0.03642367935180664, 0.03636633682250977, 0.036348926544189454, 0.0362874870300293, 0.03641856002807617, 0.03627008056640625, 0.03639295959472656, 0.03633049774169922, 0.036393985748291016, 0.036340736389160154, 0.03644316864013672, 0.03637039947509765, 0.0363765754699707, 0.036299774169921875, 0.03642777633666992, 0.036354049682617184, 0.03643904113769531, 0.03643494415283203, 0.03647283172607422, 0.03668070220947266, 0.037223423004150394, 0.03657321548461914, 0.036579296112060546, 0.03645337677001953, 0.03649126434326172, 0.036413440704345705, 0.03647283172607422, 0.0363612174987793, 0.036506622314453126, 0.036441089630126954, 0.036490238189697266, 0.036393985748291016, 0.03644723129272461, 0.036431873321533206, 0.03647590255737305, 0.03640627288818359, 0.03650867080688477, 0.03642367935180664, 0.036547584533691405, 0.03643699264526367, 0.03651686477661133, 0.13570252990722656, 0.03622092819213867, 0.036206592559814454, 0.03616358566284179, 0.03624652862548828, 0.03619635009765625, 0.03626700973510742, 0.036291584014892575, 0.03628953552246094, 0.036208641052246096, 0.036307968139648435, 0.036237342834472656, 0.036375518798828124, 0.036326400756835936, 0.036391937255859375, 0.03624959945678711, 0.03625164794921875, 0.0362690544128418, 0.03629260635375976, 0.03621887969970703, 0.03631411361694336, 0.03622604751586914, 0.03634995269775391, 0.0362762222290039, 0.03638988876342773, 0.036279296875, 0.03639807891845703, 0.03629568099975586, 0.03636019134521484, 0.03628851318359375, 0.036397056579589845, 0.0363694076538086, 0.03644723129272461, 0.03631923294067383, 0.03636019134521484, 0.036413440704345705, 0.03640729522705078, 0.036375553131103515, 0.03644416046142578, 0.03671039962768555, 0.036550655364990234, 0.036354049682617184, 0.036429824829101565, 0.03633152008056641, 0.03645542526245117, 0.036390975952148436, 0.03646150588989258, 0.036435966491699216, 0.03644723129272461, 0.0363765754699707, 0.03647590255737305, 0.03637145614624023, 0.03645849609375, 0.036395008087158204, 0.03648819351196289, 0.03644518280029297, 0.03657932662963867, 0.03674009704589844, 0.03700838470458984, 0.036587520599365236, 0.03657932662963867, 0.036429824829101565, 0.036582462310791014, 0.13591033935546876, 0.036274177551269535, 0.036296703338623046, 0.03617491149902344, 0.03623827362060547, 0.036219905853271485, 0.03631206512451172, 0.0363059196472168, 0.03630080032348633, 0.03623321533203125, 0.036424705505371094, 0.03636326217651367, 0.03638681411743164, 0.03627212905883789, 0.036318206787109376, 0.0362608642578125, 0.036310016632080076, 0.03623628616333008, 0.03633868789672851, 0.036466686248779294, 0.03648921585083008, 0.03634995269775391, 0.03639807891845703, 0.036340736389160154, 0.03644313430786133, 0.03630080032348633, 0.036431873321533206, 0.03633561706542969, 0.03649228668212891, 0.036441089630126954, 0.03638784027099609, 0.0363397102355957, 0.036393985748291016, 0.03635609436035156, 0.03642879867553711, 0.03638988876342773, 0.03642060852050781, 0.0363059196472168, 0.03639603042602539, 0.036383743286132815, 0.03642265701293945, 0.0362874870300293, 0.036421630859375, 0.03631718444824219, 0.03645337677001953, 0.03642060852050781, 0.03647385787963867, 0.036395008087158204, 0.03647385787963867, 0.036350975036621096, 0.03648716735839844, 0.03634175872802734, 0.036618240356445314, 0.0363765754699707, 0.0365291519165039, 0.036446208953857424, 0.0365588493347168, 0.03640729522705078, 0.036582401275634766, 0.03638579177856445, 0.036600894927978515, 0.03641644668579101, 0.03663568115234375, 0.135990234375, 0.0362608642578125, 0.03622604751586914, 0.036149246215820316, 0.036278270721435545, 0.03627008056640625, 0.03632844924926758, 0.03861913681030273, 0.03653734588623047, 0.0362874870300293, 0.03628236770629883, 0.036209663391113284, 0.03628441619873047, 0.03632844924926758, 0.03628543853759766, 0.036318206787109376, 0.036310016632080076, 0.03625983810424805, 0.03637247848510742, 0.03624652862548828, 0.036347904205322266, 0.03623526382446289, 0.03630694580078125, 0.036327423095703124, 0.036307968139648435, 0.036275199890136715, 0.036367359161376955, 0.03625267028808594, 0.036410369873046876, 0.036324352264404294, 0.036380672454833986, 0.036332542419433594, 0.03638988876342773, 0.03630899047851562, 0.03637145614624023, 0.036276256561279294, 0.03638166427612305, 0.036313087463378906, 0.03640934371948242, 0.036357120513916014, 0.036431873321533206, 0.03638169479370117, 0.036431873321533206, 0.03640524673461914, 0.03642777633666992, 0.036441089630126954, 0.03645644760131836, 0.036400127410888675, 0.03646361541748047, 0.03638886260986328, 0.03645849609375, 0.036364288330078126, 0.036514816284179685, 0.03639091110229492, 0.03646771240234375, 0.03649945449829101, 0.03649331283569336, 0.03637964630126953, 0.03653843307495117, 0.03643078231811524, 0.03649740982055664, 0.03644211196899414, 0.036620288848876956, 0.13562265014648436, 0.03616767883300781, 0.03624038314819336, 0.03623833465576172, 0.036313087463378906, 0.03627724838256836, 0.036296703338623046, 0.03633561706542969, 0.036326400756835936, 0.03625062561035156, 0.036342784881591796, 0.036316158294677735, 0.036397056579589845, 0.036280319213867186, 0.0363059196472168, 0.03632844924926758, 0.036416511535644534, 0.03628851318359375, 0.036318206787109376, 0.03624448013305664, 0.03633868789672851, 0.0362608642578125, 0.036350975036621096, 0.036334590911865236, 0.036391937255859375, 0.03628236770629883, 0.03638169479370117, 0.03627212905883789, 0.03632844924926758, 0.03629363250732422, 0.03647078323364258, 0.03645951843261719, 0.03647999954223633, 0.03644518280029297, 0.036468734741210936, 0.036466686248779294, 0.03645439910888672, 0.0364031982421875, 0.03638784027099609, 0.03636025619506836, 0.036504512786865236, 0.03647180938720703, 0.036483070373535154, 0.03644934463500977, 0.03663148880004883, 0.036466686248779294, 0.03727667236328125, 0.03662847900390625, 0.03653324890136719, 0.036441089630126954, 0.0365588493347168, 0.03643904113769531, 0.03647488021850586, 0.036452350616455076, 0.03672883224487305, 0.03656294250488281, 0.036555774688720705, 0.036511745452880856, 0.03660800170898437, 0.036452350616455076, 0.03658137512207031, 0.03649228668212891, 0.03656499099731445, 0.13564927673339844, 0.03618201446533203, 0.03623635101318359, 0.03616044616699219, 0.036245502471923825, 0.0362239990234375, 0.036296703338623046, 0.03638476943969727, 0.036348926544189454, 0.03622809600830078, 0.03628851318359375, 0.036206592559814454, 0.036310016632080076, 0.036307968139648435, 0.03627008056640625, 0.03626598358154297, 0.036395008087158204, 0.036326400756835936, 0.036318206787109376, 0.03632025527954102, 0.036506622314453126, 0.03630899047851562, 0.03634380722045898, 0.03630182266235352, 0.03649126434326172, 0.03635200119018555, 0.03635609436035156, 0.03627008056640625, 0.036395008087158204, 0.0362977294921875, 0.036395008087158204, 0.036362239837646484, 0.036342784881591796, 0.036337665557861325, 0.03643801498413086, 0.03631718444824219, 0.036421630859375, 0.036318206787109376, 0.03638681411743164, 0.03637247848510742, 0.03642265701293945, 0.03629056167602539, 0.03644825744628906, 0.036313087463378906, 0.036424705505371094, 0.0363581428527832, 0.03648819351196289, 0.0365588493347168, 0.03657727813720703, 0.03638886260986328, 0.03649740982055664, 0.03644211196899414, 0.03659571075439453, 0.036468734741210936, 0.036566017150878906, 0.03651686477661133, 0.03659571075439453, 0.036501502990722655, 0.036618240356445314, 0.03649228668212891, 0.0366376953125, 0.036452350616455076, 0.03662847900390625, 0.1357506561279297, 0.036208641052246096, 0.03622809600830078, 0.0362690544128418, 0.03670937728881836, 0.03640422439575195, 0.036393985748291016, 0.0363694076538086, 0.03634175872802734, 0.036327423095703124, 0.036383743286132815, 0.03631718444824219, 0.0363694076538086, 0.03638476943969727, 0.03637350463867187, 0.03638272094726563, 0.03635200119018555, 0.036332542419433594, 0.03638169479370117, 0.036334590911865236, 0.03639295959472656, 0.03629363250732422, 0.03639807891845703, 0.036606975555419925, 0.03640934371948242, 0.036359169006347655, 0.03647078323364258, 0.036334590911865236, 0.03645849609375, 0.03632844924926758, 0.03646464157104492, 0.03679948806762695, 0.036604927062988284, 0.03640217590332031, 0.03642265701293945, 0.03640524673461914, 0.03644416046142578, 0.03640524673461914, 0.036557823181152346, 0.0364769287109375, 0.03648921585083008, 0.036446208953857424, 0.03648716735839844, 0.03645644760131836, 0.0365219841003418, 0.0364400634765625, 0.036512832641601566, 0.03652703857421875, 0.03654553604125976, 0.03638579177856445, 0.03648716735839844, 0.03638988876342773, 0.036587520599365236, 0.03645542526245117, 0.03650969696044922, 0.03640627288818359, 0.03658444976806641, 0.036400127410888675, 0.03651686477661133, 0.036424705505371094, 0.03655168151855469, 0.036393985748291016, 0.03654246520996094]",tokens/s,26.440033194440694,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-180B,tiiuae/falcon-180B,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 403 Client Error: Forbidden for url: https://huggingface.co/tiiuae/falcon-180B/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-66932700-3411bedb4ea922cb16147966;dd5d3a48-6cd3-4fc0-99b7-727a0359a801) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/tiiuae/falcon-180B/resolve/main/config.json. If you are trying to create or update content,make sure you have a token with the `write` role. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1826, in _raise_on_head_call_error raise LocalEntryNotFoundError( huggingface_hub.utils._errors.LocalEntryNotFoundError: An error happened while trying to locate the file on the Hub and we cannot find the requested files in the local cache. Please check your connection and try again or make sure your Internet connection is on. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 445, in cached_file raise EnvironmentError( OSError: We couldn't connect to 'https://huggingface.co' to load this file, couldn't find it in the cached files and it looks like tiiuae/falcon-180B is not the path to a directory containing a file named config.json. Checkout your internet connection or see how to run the library in offline mode at 'https://huggingface.co/docs/transformers/installation#offline-mode'. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/gpt-neo-1.3B,EleutherAI/gpt-neo-1.3B,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1232.2816,3760.717824,0.0,3114.27072,2817.469952,s,10,0.24775356674194338,0.024775356674194338,0.0004406562493069464,0.024642784118652346,0.025366348648071288,0.025608470153808596,0.02580216735839844,"[0.0258505916595459, 0.024791872024536133, 0.025312543869018555, 0.024755136489868164, 0.024431135177612303, 0.024574592590332033, 0.024425695419311524, 0.02445907211303711, 0.024710975646972656, 0.024441951751708983]",tokens/s,10332.848215527245,kWh,2.9084853077403454e-07,1.593704725048549e-07,1.3453709916658509e-06,1.7955899949447402e-06,tokens/kWh,142571522.85362253,MB,1232.2816,3762.814976,0.0,3116.367872,2885.484544,s,10,11.001755371093749,1.100175537109375,0.014737269280413707,1.1002377319335936,1.1102823120117187,1.1229966247558594,1.1331680749511719,"[1.1357109375, 1.100788818359375, 1.1048980712890626, 1.1074569091796875, 1.0963321533203125, 1.0871802978515626, 1.07984130859375, 1.0854166259765625, 1.104443603515625, 1.0996866455078125]",tokens/s,57.26358919552743,kWh,1.2713598707538792e-05,6.964930030423382e-06,3.619640904173379e-05,5.587493777969596e-05,tokens/kWh,1127518.0340852777,,s,629,11.205391384124745,0.01781461269336209,0.0029990830027826505,0.017555456161499023,0.01802670783996582,0.018119667434692385,0.04176973937988281,"[0.01833990478515625, 0.01713145637512207, 0.01681203269958496, 0.016891904830932617, 0.016869375228881836, 0.017743871688842772, 0.01808697509765625, 0.018006975173950196, 0.01804595184326172, 0.018027551651000978, 0.01811964797973633, 0.018177024841308592, 0.018036735534667968, 0.01801625633239746, 0.018077695846557617, 0.01805619239807129, 0.01801420783996582, 0.018134048461914062, 0.01807459259033203, 0.01804902458190918, 0.018113536834716795, 0.017991680145263672, 0.01818726348876953, 0.01768448066711426, 0.017854463577270507, 0.018025472640991212, 0.018059328079223634, 0.01807251167297363, 0.018268159866333008, 0.018070528030395508, 0.01840947151184082, 0.018551807403564453, 0.01845043182373047, 0.018531328201293946, 0.018226175308227538, 0.01806540870666504, 0.018110464096069336, 0.01807257652282715, 0.018104320526123048, 0.018053119659423827, 0.01803059196472168, 0.01819545555114746, 0.017977344512939454, 0.018120704650878908, 0.018086912155151368, 0.018025472640991212, 0.018120704650878908, 0.018308095932006836, 0.01802137565612793, 0.01802342414855957, 0.01802649688720703, 0.018098175048828127, 0.01807263946533203, 0.018018239974975585, 0.018117631912231445, 0.018001920700073244, 0.018098175048828127, 0.017975296020507812, 0.018144256591796876, 0.01801318359375, 0.01802239990234375, 0.0180316162109375, 0.04327936172485351, 0.017993728637695314, 0.018104320526123048, 0.017987648010253907, 0.01809401512145996, 0.018010175704956055, 0.018068416595458985, 0.018168832778930662, 0.018119680404663087, 0.01801420783996582, 0.018069503784179687, 0.01811564826965332, 0.017996736526489258, 0.018036735534667968, 0.017757183074951173, 0.017474559783935546, 0.017597440719604493, 0.017584127426147463, 0.017571840286254883, 0.017526784896850587, 0.017546239852905272, 0.017540096282958984, 0.01760972785949707, 0.017879039764404296, 0.017562623977661132, 0.017564672470092774, 0.017507328033447265, 0.017527807235717775, 0.017570816040039062, 0.017622047424316407, 0.017581024169921876, 0.01762611198425293, 0.017507328033447265, 0.017527807235717775, 0.017573888778686524, 0.01781760025024414, 0.017912832260131836, 0.017672191619873046, 0.017698816299438477, 0.017762304306030274, 0.017655807495117186, 0.0168724479675293, 0.016923648834228516, 0.016865280151367186, 0.01681612777709961, 0.01683967971801758, 0.016945152282714843, 0.016923648834228516, 0.016931840896606445, 0.01681817626953125, 0.016858112335205077, 0.01683558464050293, 0.01683967971801758, 0.016891904830932617, 0.016906240463256835, 0.016722944259643553, 0.016689151763916017, 0.016737279891967775, 0.0169564151763916, 0.016871423721313478, 0.016876544952392578, 0.017147903442382813, 0.01741004753112793, 0.043858943939208986, 0.01765990447998047, 0.0176363525390625, 0.01768550491333008, 0.017538047790527343, 0.017488895416259767, 0.017483776092529296, 0.017632255554199217, 0.017343488693237305, 0.01724006462097168, 0.017192960739135742, 0.01744691276550293, 0.017490943908691405, 0.017514495849609374, 0.01762918472290039, 0.017584127426147463, 0.017528831481933595, 0.017502208709716797, 0.017528831481933595, 0.017531904220581054, 0.017538047790527343, 0.017537023544311522, 0.017535999298095704, 0.017705984115600586, 0.017573888778686524, 0.017512447357177736, 0.01757801628112793, 0.017504224777221678, 0.017548351287841796, 0.01751340866088867, 0.017523712158203125, 0.017604608535766602, 0.017472511291503907, 0.01758515167236328, 0.017529855728149413, 0.0174704647064209, 0.017512447357177736, 0.017538047790527343, 0.017510400772094727, 0.01747865676879883, 0.017673215866088866, 0.017533952713012696, 0.017508352279663086, 0.017497087478637697, 0.017554431915283202, 0.017540096282958984, 0.017537023544311522, 0.017674240112304687, 0.017547264099121093, 0.017617919921875, 0.017480703353881837, 0.017276927947998046, 0.017469440460205078, 0.017474559783935546, 0.017530879974365234, 0.01756159973144531, 0.017726463317871095, 0.01761484718322754, 0.017567743301391603, 0.017482751846313475, 0.01759436798095703, 0.017367040634155274, 0.01722777557373047, 0.04272947311401367, 0.018311168670654295, 0.0176680965423584, 0.017513471603393553, 0.017510400772094727, 0.017525760650634766, 0.017451007843017577, 0.01763532829284668, 0.017554431915283202, 0.017589248657226563, 0.017559551239013673, 0.0175994873046875, 0.017560575485229494, 0.017573888778686524, 0.017520736694335938, 0.01761782455444336, 0.017488895416259767, 0.01758515167236328, 0.01758720016479492, 0.017525760650634766, 0.017491968154907226, 0.017510400772094727, 0.01728102493286133, 0.017313791275024415, 0.017583103179931642, 0.017333248138427734, 0.01721139144897461, 0.01745305633544922, 0.017528831481933595, 0.017565696716308594, 0.01762099266052246, 0.01765273666381836, 0.017582080841064454, 0.017622016906738282, 0.017608768463134767, 0.017642431259155274, 0.0176680965423584, 0.017555456161499023, 0.017567743301391603, 0.01762816047668457, 0.01764556884765625, 0.017572864532470703, 0.017547264099121093, 0.017622016906738282, 0.01761689567565918, 0.017649663925170898, 0.0175032958984375, 0.016844736099243165, 0.016973823547363282, 0.01835212707519531, 0.017640447616577147, 0.01760665512084961, 0.01766912078857422, 0.01765376091003418, 0.01759334373474121, 0.01768448066711426, 0.01761075210571289, 0.017617919921875, 0.017624063491821287, 0.017566719055175782, 0.017738752365112305, 0.01768550491333008, 0.017697792053222656, 0.04177203369140625, 0.01683046340942383, 0.016922624588012695, 0.01739366340637207, 0.01761689567565918, 0.017649663925170898, 0.01758720016479492, 0.017713151931762695, 0.01759334373474121, 0.017661951065063478, 0.01759846305847168, 0.01756159973144531, 0.017572864532470703, 0.01760665512084961, 0.01765376091003418, 0.0176312313079834, 0.017671167373657228, 0.017625087738037108, 0.01763430404663086, 0.017604608535766602, 0.017565696716308594, 0.01764454460144043, 0.01762611198425293, 0.01756979179382324, 0.017632255554199217, 0.01759231948852539, 0.01760870361328125, 0.017300479888916014, 0.01747148895263672, 0.01756368064880371, 0.01769059181213379, 0.017674240112304687, 0.017562623977661132, 0.017612800598144532, 0.01763430404663086, 0.017527807235717775, 0.01762816047668457, 0.017547264099121093, 0.017573888778686524, 0.01764249610900879, 0.01761894416809082, 0.017562623977661132, 0.017632255554199217, 0.017665023803710937, 0.0175994873046875, 0.01762918472290039, 0.01760051155090332, 0.016907264709472656, 0.016891904830932617, 0.016908287048339844, 0.016859136581420898, 0.016890880584716796, 0.016939008712768554, 0.01682022476196289, 0.016882688522338866, 0.0170250244140625, 0.016905216217041014, 0.016887807846069337, 0.016884735107421875, 0.01681510353088379, 0.01682431983947754, 0.016869375228881836, 0.017480703353881837, 0.0421776008605957, 0.017000383377075195, 0.016833536148071288, 0.017121280670166016, 0.016661504745483398, 0.016688127517700196, 0.01677107238769531, 0.016903200149536134, 0.016771039962768554, 0.017077247619628907, 0.01737113571166992, 0.017571840286254883, 0.017555456161499023, 0.017515520095825195, 0.017544191360473634, 0.01760972785949707, 0.017554431915283202, 0.01755340766906738, 0.017552480697631836, 0.017490848541259766, 0.017511423110961915, 0.01757900810241699, 0.017427455902099608, 0.01761587142944336, 0.017589248657226563, 0.01732204818725586, 0.017381311416625977, 0.017305599212646485, 0.01720217514038086, 0.017325056076049804, 0.017299455642700197, 0.01762099266052246, 0.01756064033508301, 0.017665983200073242, 0.017622016906738282, 0.01777459144592285, 0.01761075210571289, 0.017559551239013673, 0.01757900810241699, 0.017628223419189452, 0.01756153678894043, 0.01686534309387207, 0.01690719985961914, 0.01680281639099121, 0.016761856079101564, 0.016660480499267577, 0.01680691146850586, 0.016974847793579103, 0.01683967971801758, 0.016837631225585938, 0.0168089599609375, 0.01685606384277344, 0.01686425590515137, 0.016950271606445313, 0.01686016082763672, 0.016869375228881836, 0.01663488006591797, 0.016880640029907225, 0.017047552108764647, 0.018149375915527344, 0.01801215934753418, 0.017719295501708983, 0.017688575744628905, 0.042221569061279295, 0.017573888778686524, 0.01762611198425293, 0.01757695960998535, 0.017415168762207032, 0.016690176010131837, 0.016648191452026367, 0.016746496200561522, 0.0168407039642334, 0.016862272262573242, 0.01670137596130371, 0.016653375625610353, 0.016815040588378908, 0.01682022476196289, 0.01681407928466797, 0.016954368591308593, 0.016846847534179688, 0.016902143478393555, 0.016870399475097657, 0.01696870422363281, 0.016933887481689454, 0.016978944778442383, 0.016892927169799805, 0.016927743911743166, 0.01681407928466797, 0.01663801574707031, 0.016739263534545898, 0.01740287971496582, 0.01734758377075195, 0.017705984115600586, 0.01756876754760742, 0.017879039764404296, 0.017460224151611328, 0.01757695960998535, 0.017558528900146485, 0.017519615173339845, 0.017743871688842772, 0.018181119918823242, 0.017929216384887696, 0.017808448791503905, 0.017609664916992188, 0.017507328033447265, 0.01757798385620117, 0.017533952713012696, 0.01753091239929199, 0.017538015365600585, 0.016867328643798828, 0.01702604866027832, 0.016919551849365236, 0.016871423721313478, 0.016854015350341797, 0.016829439163208008, 0.016987136840820313, 0.0167956485748291, 0.016883712768554687, 0.016958463668823243, 0.016877567291259766, 0.016852991104125976, 0.016865280151367186, 0.01680588722229004, 0.01682841682434082, 0.01680691146850586, 0.016882688522338866, 0.04173209762573242, 0.01681920051574707, 0.016859136581420898, 0.01679974365234375, 0.01681715202331543, 0.01680588722229004, 0.016792575836181642, 0.016951295852661134, 0.01681510353088379, 0.01660825538635254, 0.01685606384277344, 0.016995328903198242, 0.016882688522338866, 0.016930816650390625, 0.01680384063720703, 0.016930816650390625, 0.016854015350341797, 0.01684889602661133, 0.01681100845336914, 0.016759807586669923, 0.01680691146850586, 0.01679974365234375, 0.0167956485748291, 0.017571840286254883, 0.01765171241760254, 0.017552383422851564, 0.017542144775390626, 0.017649663925170898, 0.01768448066711426, 0.017590272903442384, 0.017671167373657228, 0.01698508834838867, 0.016945152282714843, 0.017073152542114257, 0.01702092742919922, 0.016917503356933594, 0.016895999908447267, 0.016897024154663084, 0.01685196876525879, 0.016926719665527345, 0.016845823287963867, 0.01700044822692871, 0.016882688522338866, 0.018183168411254884, 0.018489343643188477, 0.017822719573974608, 0.017597440719604493, 0.017763328552246094, 0.01758720016479492, 0.01757900810241699, 0.017558528900146485, 0.017570816040039062, 0.017317888259887695, 0.017744895935058593, 0.017635360717773437, 0.017713119506835937, 0.017530879974365234, 0.01762303924560547, 0.01757798385620117, 0.017622016906738282, 0.01765273666381836, 0.017535999298095704, 0.017556480407714844, 0.04275305557250977, 0.01749500846862793, 0.017538047790527343, 0.017514495849609374, 0.017533952713012696, 0.017510400772094727, 0.017497087478637697, 0.01762816047668457, 0.017539072036743163, 0.017605632781982423, 0.01740390396118164, 0.01723289680480957, 0.01738956832885742, 0.017539072036743163, 0.017541120529174805, 0.01760665512084961, 0.01764761543273926, 0.017690624237060547, 0.017535999298095704, 0.017532928466796875, 0.017528831481933595, 0.01758720016479492, 0.017556480407714844, 0.01756159973144531, 0.017541120529174805, 0.017564672470092774, 0.017712192535400392, 0.017583040237426757, 0.017564672470092774, 0.017508352279663086, 0.017506303787231444, 0.01763532829284668, 0.017498111724853514, 0.017732608795166017, 0.01757596778869629, 0.01752572822570801, 0.017551359176635743, 0.017438720703125, 0.01723391914367676, 0.017693695068359376, 0.01752272033691406, 0.017508319854736328, 0.017541120529174805, 0.01758515167236328, 0.017527807235717775, 0.017540096282958984, 0.017556480407714844, 0.017521663665771483, 0.017321983337402345, 0.017325056076049804, 0.01776639938354492, 0.017516544342041016, 0.017625087738037108, 0.017532928466796875, 0.017368064880371094, 0.017376256942749024, 0.01721958351135254, 0.017451007843017577, 0.017528831481933595, 0.017466367721557616, 0.01749504089355469, 0.017543167114257813, 0.017517568588256836, 0.041763839721679685, 0.016850944519042968, 0.016773120880126953, 0.01680691146850586, 0.01676595115661621, 0.01665023994445801, 0.016592927932739258, 0.01667990493774414, 0.01663692855834961, 0.01676288032531738, 0.017385471343994142, 0.017675264358520508, 0.017741823196411134, 0.017557504653930665, 0.01760767936706543, 0.017693695068359376, 0.017584127426147463, 0.0176680965423584, 0.01759129524230957, 0.017574911117553712, 0.017522687911987304, 0.017516544342041016, 0.017505279541015627, 0.01762918472290039, 0.017570816040039062, 0.01762611198425293, 0.017583103179931642, 0.017527807235717775, 0.017528863906860353, 0.017605600357055665, 0.01761484718322754, 0.01763839912414551, 0.017556480407714844, 0.017625087738037108, 0.01758515167236328, 0.017297407150268555, 0.017378303527832033, 0.017525760650634766, 0.01754521560668945, 0.017605632781982423, 0.01765068817138672, 0.017573888778686524, 0.017609760284423827, 0.017612768173217774, 0.017559551239013673, 0.01760358428955078, 0.01757900810241699, 0.017724416732788087, 0.017575935363769533, 0.01760153579711914, 0.017520639419555666, 0.01761075210571289, 0.017581056594848633, 0.01759436798095703, 0.017667072296142578, 0.017661951065063478, 0.017636383056640625, 0.017544160842895506, 0.01762918472290039, 0.01760873603820801, 0.017620960235595704, 0.01749510383605957, 0.01753593635559082]",tokens/s,56.133693008807896,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,mistralai/Mistral-7B-v0.1,mistralai/Mistral-7B-v0.1,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.1,,,,MB,1274.90048,15987.113984,0.0,15340.66688,15114.865664,s,10,0.854706184387207,0.0854706184387207,0.0003098703950784983,0.08537995147705078,0.08564060363769532,0.08599409561157227,0.08627688919067383,"[0.08634758758544922, 0.08527552032470703, 0.08539059448242188, 0.08556204986572266, 0.08528928375244141, 0.08518434906005859, 0.08543116760253906, 0.08536930847167969, 0.0853647689819336, 0.08549155426025391]",tokens/s,2995.1813228488877,kWh,1.011139816708035e-06,5.540558798971829e-07,6.51219656777779e-06,8.077392264383008e-06,tokens/kWh,31693397.02973489,MB,1275.22816,15987.113984,0.0,15340.66688,15134.140416,s,10,23.784769042968755,2.3784769042968756,0.0010452634175498762,2.378458740234375,2.379689135742187,2.379695397949219,2.379700407714844,"[2.378469482421875, 2.378447998046875, 2.37710205078125, 2.37970166015625, 2.376613037109375, 2.378392333984375, 2.377587890625, 2.379660400390625, 2.3791064453125, 2.379687744140625]",tokens/s,26.48753909957517,kWh,2.8018497145838217e-05,1.535513389884307e-05,0.0001203092937782227,0.000163682924822904,tokens/kWh,384890.4830370215,,s,629,24.51449241256715,0.03897375582284123,0.010205007440762926,0.037733375549316404,0.03786137771606445,0.03793366928100586,0.12354338958740235,"[0.038307838439941407, 0.03789619064331055, 0.037819393157958986, 0.03764121627807617, 0.03763916778564453, 0.037684223175048825, 0.03766067123413086, 0.03761971282958984, 0.03770163345336914, 0.037618686676025394, 0.037713920593261716, 0.03765043258666992, 0.037678081512451174, 0.03773440170288086, 0.03772313690185547, 0.03769651031494141, 0.037631999969482424, 0.03768320083618164, 0.03767603302001953, 0.03763097763061524, 0.037644287109375, 0.03765043258666992, 0.03773132705688476, 0.03770880126953125, 0.03770265579223633, 0.037674049377441406, 0.03774457550048828, 0.037735424041748046, 0.03770982360839844, 0.0377784309387207, 0.037713920593261716, 0.03765760040283203, 0.03766988754272461, 0.03766172790527344, 0.03771286392211914, 0.037684223175048825, 0.03768012619018555, 0.03768729782104492, 0.03771596908569336, 0.03777024078369141, 0.03769139099121094, 0.03770163345336914, 0.03779379272460937, 0.037759998321533206, 0.037746688842773435, 0.037736446380615234, 0.03778870391845703, 0.037776351928710934, 0.03774156951904297, 0.037754878997802735, 0.03774771118164062, 0.037797889709472655, 0.03782860946655273, 0.037875743865966795, 0.037792736053466794, 0.037789695739746096, 0.037763072967529294, 0.03785113525390625, 0.03776204681396484, 0.03786342239379883, 0.03774156951904297, 0.03780505752563477, 0.12354969787597657, 0.03761459350585938, 0.037558273315429686, 0.037594112396240234, 0.03758694458007812, 0.03770268630981445, 0.03755721664428711, 0.03768832015991211, 0.03759206390380859, 0.03765248107910156, 0.037566463470458986, 0.037631038665771485, 0.037604286193847654, 0.037655582427978514, 0.03785827255249023, 0.037736446380615234, 0.037645313262939455, 0.037730304718017575, 0.03763814544677734, 0.03774054336547852, 0.03767193603515625, 0.03772518539428711, 0.03768832015991211, 0.03778559875488281, 0.03771596908569336, 0.037700607299804685, 0.037647361755371096, 0.037795841217041014, 0.037703678131103514, 0.037748737335205076, 0.037684223175048825, 0.037806079864501956, 0.03766579055786133, 0.037718017578125, 0.03770470428466797, 0.037716991424560545, 0.037698558807373043, 0.03775897598266602, 0.03767603302001953, 0.03771596908569336, 0.037664768218994144, 0.03778355026245117, 0.03772313690185547, 0.03778252792358398, 0.037716991424560545, 0.03778870391845703, 0.03783983993530273, 0.03783679962158203, 0.0377149429321289, 0.03781222534179687, 0.03793305587768555, 0.03840921783447265, 0.0379422721862793, 0.03784089660644531, 0.03782656097412109, 0.03791872024536133, 0.03787366485595703, 0.037789695739746096, 0.037820415496826174, 0.03782758331298828, 0.03789619064331055, 0.0379156494140625, 0.03785728073120117, 0.1236490249633789, 0.03762998580932617, 0.03774153518676758, 0.0376360969543457, 0.03763507080078125, 0.037662750244140626, 0.03758076858520508, 0.03765555191040039, 0.03764121627807617, 0.03765862274169922, 0.03761151885986328, 0.037615615844726565, 0.037601280212402347, 0.0376545295715332, 0.03765555191040039, 0.037722110748291016, 0.037663745880126956, 0.03771289443969727, 0.037648384094238284, 0.03762790298461914, 0.037678081512451174, 0.03762998580932617, 0.037663711547851565, 0.03775692749023438, 0.037618686676025394, 0.03766579055786133, 0.03769036865234375, 0.03772825622558594, 0.03770675277709961, 0.03771289443969727, 0.03773235321044922, 0.037733375549316404, 0.037667839050292966, 0.037719039916992186, 0.037735424041748046, 0.03769651031494141, 0.037705726623535156, 0.0376627197265625, 0.03769241714477539, 0.037720062255859374, 0.03786444854736328, 0.03833139038085937, 0.03787161636352539, 0.03776921463012695, 0.03775283050537109, 0.03777433776855469, 0.037768192291259765, 0.03784908676147461, 0.03776204681396484, 0.03772825622558594, 0.03780915069580078, 0.037720062255859374, 0.03775283050537109, 0.037736446380615234, 0.03779174423217774, 0.03778867340087891, 0.03779993438720703, 0.037754878997802735, 0.03784703826904297, 0.03779379272460937, 0.03779072189331055, 0.03776716613769531, 0.037792766571044925, 0.12421734619140624, 0.03772518539428711, 0.037617664337158206, 0.03768627166748047, 0.0375551986694336, 0.03764633560180664, 0.03760950469970703, 0.0376954574584961, 0.03773952102661133, 0.037781505584716796, 0.03759001541137695, 0.03767398452758789, 0.037646366119384767, 0.0376954574584961, 0.03762073516845703, 0.037774368286132814, 0.037715007781982425, 0.03788483047485352, 0.037684223175048825, 0.03772108840942383, 0.037730304718017575, 0.03776716613769531, 0.03764121627807617, 0.03775078582763672, 0.03776921463012695, 0.038035457611083984, 0.03793407821655274, 0.03785625457763672, 0.037703678131103514, 0.03771596908569336, 0.03767398452758789, 0.03780198287963867, 0.03769241714477539, 0.03770675277709961, 0.03769139099121094, 0.03768524932861328, 0.037720062255859374, 0.03779379272460937, 0.037978111267089845, 0.03802214431762695, 0.03779072189331055, 0.03784294509887695, 0.03781631851196289, 0.037823486328125, 0.037792766571044925, 0.03774771118164062, 0.03773952102661133, 0.03784601593017578, 0.03777740859985351, 0.03783168029785156, 0.03775897598266602, 0.03781017684936523, 0.03777740859985351, 0.037771263122558595, 0.037730304718017575, 0.03782963180541992, 0.03779891204833984, 0.03778255844116211, 0.03782857513427734, 0.03779379272460937, 0.037784576416015625, 0.037784576416015625, 0.0378787841796875, 0.12352716827392578, 0.03761356735229492, 0.037607425689697264, 0.037617664337158206, 0.03765657424926758, 0.037572608947753904, 0.03760025787353516, 0.03764019012451172, 0.03764633560180664, 0.0376258544921875, 0.037615615844726565, 0.037651454925537106, 0.037631999969482424, 0.03767603302001953, 0.037631999969482424, 0.03771289443969727, 0.03765350341796875, 0.03768320083618164, 0.03768115234375, 0.03768524932861328, 0.03768627166748047, 0.03767603302001953, 0.03768627166748047, 0.03771596908569336, 0.0376627197265625, 0.03768832015991211, 0.03769139099121094, 0.037694465637207034, 0.03772108840942383, 0.03768627166748047, 0.03772825622558594, 0.03773440170288086, 0.037703678131103514, 0.03768320083618164, 0.037746688842773435, 0.037748737335205076, 0.03766988754272461, 0.03771187210083008, 0.03774156951904297, 0.03770982360839844, 0.03771289443969727, 0.03772723388671875, 0.037751808166503906, 0.037722110748291016, 0.037749759674072264, 0.03772825622558594, 0.03778153610229492, 0.037791713714599606, 0.03778252792358398, 0.037743614196777346, 0.03779379272460937, 0.03774259185791016, 0.037773311614990236, 0.037781505584716796, 0.03778867340087891, 0.037833728790283204, 0.037771263122558595, 0.03789311981201172, 0.037897216796875, 0.03784089660644531, 0.037852161407470705, 0.037835777282714846, 0.03783475112915039, 0.12338893127441407, 0.03763507080078125, 0.03769548797607422, 0.03768115234375, 0.03762483215332031, 0.03764940643310547, 0.03759308624267578, 0.037698558807373043, 0.037594112396240234, 0.03765248107910156, 0.037550079345703126, 0.03763302230834961, 0.037617664337158206, 0.03772825622558594, 0.03760025787353516, 0.03771187210083008, 0.037663745880126956, 0.037718017578125, 0.0376360969543457, 0.03778662490844727, 0.037644287109375, 0.037722110748291016, 0.0376627197265625, 0.03773235321044922, 0.03765862274169922, 0.037736446380615234, 0.03768627166748047, 0.03773235321044922, 0.037703678131103514, 0.037720062255859374, 0.03767091369628906, 0.03775283050537109, 0.03767603302001953, 0.03775795364379883, 0.037689342498779296, 0.037751808166503906, 0.03767910385131836, 0.03778867340087891, 0.03777433776855469, 0.03775897598266602, 0.03773132705688476, 0.03777740859985351, 0.03772415924072266, 0.037749759674072264, 0.03779993438720703, 0.03784499359130859, 0.03778867340087891, 0.0378419189453125, 0.037754878997802735, 0.03775795364379883, 0.037748737335205076, 0.037751808166503906, 0.03775385665893555, 0.03786137771606445, 0.03802316665649414, 0.038727680206298826, 0.0378869743347168, 0.0377968635559082, 0.03781119918823242, 0.03787673568725586, 0.0378081283569336, 0.037771263122558595, 0.037855232238769534, 0.12368793487548828, 0.03761151885986328, 0.037636127471923825, 0.03768521499633789, 0.0377149429321289, 0.03774156951904297, 0.03769548797607422, 0.037720062255859374, 0.03761052703857422, 0.037637088775634764, 0.03762995147705078, 0.03769241714477539, 0.037661697387695314, 0.037765121459960936, 0.03775283050537109, 0.03768012619018555, 0.03771289443969727, 0.037684223175048825, 0.03765350341796875, 0.03769036865234375, 0.03767091369628906, 0.037730304718017575, 0.03769036865234375, 0.037763072967529294, 0.037663745880126956, 0.03769343948364258, 0.03773750305175781, 0.037663711547851565, 0.03768115234375, 0.03768524932861328, 0.03775692749023438, 0.03770265579223633, 0.037733375549316404, 0.03772518539428711, 0.03772415924072266, 0.037700607299804685, 0.037718017578125, 0.037738494873046875, 0.037705726623535156, 0.03771289443969727, 0.037713920593261716, 0.037716991424560545, 0.03776208114624023, 0.0377374382019043, 0.03775692749023438, 0.03782553482055664, 0.03807027053833008, 0.03783679962158203, 0.03775897598266602, 0.03775385665893555, 0.03773952102661133, 0.037765121459960936, 0.037789695739746096, 0.03779481506347656, 0.03779993438720703, 0.037822463989257815, 0.03778252792358398, 0.037751808166503906, 0.037817344665527344, 0.03775590515136719, 0.03783679962158203, 0.037768192291259765, 0.03784601593017578, 0.12363468933105469, 0.03762688064575195, 0.03759820938110352, 0.03778559875488281, 0.037760032653808596, 0.037770206451416016, 0.03759516906738281, 0.037709793090820315, 0.037544960021972655, 0.03765248107910156, 0.03764640045166016, 0.03766265487670899, 0.03760947036743164, 0.03775283050537109, 0.037634048461914066, 0.03771289443969727, 0.0378081283569336, 0.037703678131103514, 0.037648384094238284, 0.03773952102661133, 0.03765043258666992, 0.037719039916992186, 0.037690399169921875, 0.037758945465087894, 0.037694465637207034, 0.03768729782104492, 0.037692447662353516, 0.03771286392211914, 0.037716991424560545, 0.03806924819946289, 0.03791155242919922, 0.03783475112915039, 0.03768115234375, 0.03790643310546875, 0.03805491256713867, 0.037765182495117185, 0.03776300811767578, 0.037820415496826174, 0.03770982360839844, 0.03785420989990235, 0.03769343948364258, 0.03770470428466797, 0.03774771118164062, 0.03779072189331055, 0.03783782577514649, 0.037748737335205076, 0.037800960540771485, 0.03794124984741211, 0.03783065414428711, 0.03775078582763672, 0.03786137771606445, 0.037956607818603515, 0.03778867340087891, 0.03776409530639648, 0.0378081283569336, 0.03794739151000977, 0.03783475112915039, 0.037835777282714846, 0.03789823913574219, 0.037814273834228515, 0.037776382446289065, 0.03782860946655273, 0.037792766571044925, 0.12364492797851563, 0.03760947036743164, 0.03760639953613281, 0.03766886520385742, 0.03761459350585938, 0.03762278366088867, 0.037684223175048825, 0.037735424041748046, 0.037716991424560545, 0.037738494873046875, 0.037651454925537106, 0.03764121627807617, 0.03767603302001953, 0.037684223175048825, 0.03770470428466797, 0.03770880126953125, 0.03770982360839844, 0.03787980651855469, 0.03772415924072266, 0.037716991424560545, 0.037678081512451174, 0.037735424041748046, 0.03778867340087891, 0.0377262077331543, 0.03777740859985351, 0.0377968635559082, 0.03772723388671875, 0.03774771118164062, 0.037743614196777346, 0.03772518539428711, 0.03776716613769531, 0.037759998321533206, 0.03771289443969727, 0.037735424041748046, 0.03774156951904297, 0.0378869743347168, 0.037810207366943356, 0.03773436737060547, 0.0377968635559082, 0.037779457092285154, 0.037733375549316404, 0.03774771118164062, 0.03772825622558594, 0.03777433776855469, 0.037773311614990236, 0.037719039916992186, 0.03785932922363281, 0.037803009033203126, 0.03777536010742188, 0.03771289443969727, 0.03780710220336914, 0.03786342239379883, 0.03786649703979492, 0.03774771118164062, 0.03779891204833984, 0.03781324768066406, 0.03778464126586914, 0.03778144073486328, 0.03782355117797852, 0.03788793563842773, 0.03786956787109375, 0.03791974258422852, 0.038019073486328124, 0.12372991943359375, 0.03763814544677734, 0.037661697387695314, 0.03762483215332031, 0.03757567977905273, 0.03766681671142578, 0.03760332870483398, 0.03776204681396484, 0.03760639953613281, 0.03774259185791016, 0.03767091369628906, 0.037648384094238284, 0.03756854248046875, 0.03769443130493164, 0.03762688064575195, 0.03776921463012695, 0.03768627166748047, 0.0377077751159668, 0.037773311614990236, 0.037759998321533206, 0.03769753646850586, 0.03772518539428711, 0.03772108840942383, 0.037889022827148434, 0.03774054336547852, 0.03777024078369141, 0.03779481506347656, 0.037784576416015625, 0.03765657424926758, 0.03771804809570312, 0.037721057891845704, 0.03785932922363281, 0.03767500686645508, 0.037700607299804685, 0.03772825622558594, 0.03777433776855469, 0.03780710220336914, 0.037738494873046875, 0.0377077751159668, 0.03786342239379883, 0.03769343948364258, 0.03779379272460937, 0.03772313690185547, 0.03784601593017578, 0.03787673568725586, 0.037940223693847655, 0.03783475112915039, 0.037996543884277346, 0.037759998321533206, 0.03794841766357422, 0.03783782577514649, 0.0377784309387207, 0.03774566268920899, 0.03776204681396484, 0.037787681579589845, 0.037883872985839846, 0.03802009582519531, 0.03783782577514649, 0.037959678649902344, 0.03810508728027344, 0.03784908676147461, 0.0377968635559082, 0.03782451248168945]",tokens/s,25.658291814255506,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,google/recurrentgemma-7b,google/recurrentgemma-7b,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/google/recurrentgemma-7b/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66947035-354592131c07fd051b497d90;b0375409-8cc0-44ff-a8a8-cb575bf186e5) Repository Not Found for url: https://huggingface.co/google/recurrentgemma-7b/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: google/recurrentgemma-7b is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,mistralai/Mixtral-8x7B-v0.1,mistralai/Mixtral-8x7B-v0.1,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ self.load_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 250, in load_model_with_no_weights self.pretrained_model = self.automodel_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3710, in from_pretrained model = cls(config, *model_args, **model_kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/mixtral/modeling_mixtral.py"", line 1323, in __init__ self.model = MixtralModel(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/mixtral/modeling_mixtral.py"", line 1087, in __init__ [MixtralDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/mixtral/modeling_mixtral.py"", line 1087, in [MixtralDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/mixtral/modeling_mixtral.py"", line 877, in __init__ self.block_sparse_moe = MixtralSparseMoeBlock(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/mixtral/modeling_mixtral.py"", line 824, in __init__ self.experts = nn.ModuleList([MixtralBlockSparseTop2MLP(config) for _ in range(self.num_experts)]) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/mixtral/modeling_mixtral.py"", line 824, in self.experts = nn.ModuleList([MixtralBlockSparseTop2MLP(config) for _ in range(self.num_experts)]) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/mixtral/modeling_mixtral.py"", line 791, in __init__ self.w2 = nn.Linear(self.ffn_dim, self.hidden_dim, bias=False) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/linear.py"", line 98, in __init__ self.weight = Parameter(torch.empty((out_features, in_features), **factory_kwargs)) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_device.py"", line 78, in __torch_function__ return func(*args, **kwargs) torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 112.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,t,t,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/t/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66947396-6460313908a75e12626d5686;103f2a9c-028e-4560-bb48-c706cba0c653) Repository Not Found for url: https://huggingface.co/t/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: t is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-72B,Qwen/Qwen-72B,cuda,0,42,,,True,,,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",qwen,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-30b,facebook/opt-30b,cuda,0,42,,,True,,,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",opt,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-MoE-A2.7B,Qwen/Qwen1.5-MoE-A2.7B,cuda,0,42,,,True,,,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",qwen2_moe,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-1.3b,EleutherAI/pythia-1.3b,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1286.262784,3848.798208,0.0,3202.351104,3105.82784,s,10,0.24081798553466796,0.024081798553466795,0.00038659848964031586,0.02391201686859131,0.024304146957397462,0.024740841484069825,0.025090197105407715,"[0.02517753601074219, 0.024207103729248048, 0.023898752212524414, 0.023821407318115235, 0.024194112777709963, 0.023959680557250975, 0.02386591911315918, 0.023917184829711915, 0.023906848907470704, 0.02386944007873535]",tokens/s,10630.43524060816,kWh,2.84827513946906e-07,1.5607165578747143e-07,1.2483898875999952e-06,1.6892890573343727e-06,tokens/kWh,151543040.48115796,MB,1286.262784,3848.798208,0.0,3202.351104,3188.152832,s,10,12.569388427734376,1.2569388427734376,0.014872383514358697,1.2568008422851562,1.2705527954101563,1.2805973205566406,1.2886329406738282,"[1.290641845703125, 1.2683206787109376, 1.259688232421875, 1.2572567138671875, 1.2520152587890625, 1.2313369140625, 1.26155517578125, 1.2448743896484376, 1.247354248046875, 1.256344970703125]",tokens/s,50.12177033290689,kWh,1.4532639801382539e-05,7.963009467016875e-06,3.82537250474e-05,6.074937431579941e-05,tokens/kWh,1037047.7179320555,,s,629,12.76368588638305,0.020292028436221077,0.002905302237704868,0.020099071502685546,0.020469760513305665,0.020771033477783203,0.04389322860717773,"[0.021151807785034178, 0.019891136169433592, 0.020847616195678712, 0.020914176940917968, 0.020915199279785156, 0.02111692810058594, 0.02111692810058594, 0.020792320251464845, 0.020775936126708985, 0.02079641532897949, 0.02068172836303711, 0.020815872192382814, 0.020719615936279297, 0.020724735260009765, 0.02085171127319336, 0.020757503509521484, 0.02087731170654297, 0.02070528030395508, 0.020705312728881837, 0.020793312072753905, 0.020719615936279297, 0.020801536560058592, 0.020790271759033203, 0.02076367950439453, 0.02073904037475586, 0.02065100860595703, 0.02084147262573242, 0.02065203285217285, 0.020731903076171874, 0.020502559661865233, 0.020415519714355467, 0.02015225601196289, 0.020214784622192384, 0.02018611145019531, 0.02021785545349121, 0.02026393508911133, 0.020148223876953125, 0.02012876892089844, 0.02026393508911133, 0.02018611145019531, 0.020222976684570314, 0.019801088333129883, 0.019523584365844726, 0.019513343811035155, 0.019523584365844726, 0.01947238349914551, 0.020706304550170897, 0.02047385597229004, 0.02043801689147949, 0.020192256927490236, 0.020212736129760742, 0.020229120254516602, 0.020196352005004883, 0.020288511276245116, 0.02051584053039551, 0.020315135955810547, 0.02010419273376465, 0.02039193534851074, 0.02030182456970215, 0.020184064865112306, 0.020719680786132812, 0.020320192337036132, 0.04494438552856445, 0.02020351982116699, 0.02027827262878418, 0.020206592559814454, 0.02023423957824707, 0.020159488677978517, 0.020176895141601564, 0.02027008056640625, 0.02019327926635742, 0.020350976943969725, 0.020204608917236327, 0.020264896392822265, 0.020212736129760742, 0.020196352005004883, 0.020686847686767578, 0.020505599975585938, 0.020063232421875, 0.020143104553222657, 0.020145151138305666, 0.0194652156829834, 0.01945395278930664, 0.019491840362548828, 0.019962879180908204, 0.021910528182983398, 0.020692991256713866, 0.020377599716186523, 0.020299776077270508, 0.02028441619873047, 0.020185087203979494, 0.02021171188354492, 0.02027827262878418, 0.02027827262878418, 0.02022604751586914, 0.02021785545349121, 0.020237312316894532, 0.02026905632019043, 0.020231168746948244, 0.020274175643920898, 0.020180992126464844, 0.02027212715148926, 0.020329471588134765, 0.020324352264404297, 0.020385791778564453, 0.020288511276245116, 0.02024345588684082, 0.020207679748535157, 0.020246463775634764, 0.02050662422180176, 0.020231168746948244, 0.020379648208618165, 0.02028339195251465, 0.020220928192138672, 0.019507200241088866, 0.019505151748657225, 0.01946316719055176, 0.019615743637084963, 0.019482624053955077, 0.019563520431518554, 0.019511295318603517, 0.019519487380981446, 0.019551231384277345, 0.01946828842163086, 0.019519487380981446, 0.04389580917358398, 0.019528703689575197, 0.01942835235595703, 0.019365888595581054, 0.019504127502441407, 0.019555328369140625, 0.019558399200439454, 0.019569664001464843, 0.019843072891235353, 0.02023526382446289, 0.02001203155517578, 0.020419584274291993, 0.020810752868652343, 0.01988198471069336, 0.019573759078979493, 0.01967411231994629, 0.019490816116333007, 0.0196177921295166, 0.01961676788330078, 0.01965977668762207, 0.019656736373901366, 0.019595232009887695, 0.019581951141357423, 0.01962393569946289, 0.01944268798828125, 0.019603456497192383, 0.019483648300170898, 0.01983795166015625, 0.0195020809173584, 0.019555328369140625, 0.01968332862854004, 0.02049843215942383, 0.020368383407592772, 0.020337663650512695, 0.020238336563110353, 0.02022400093078613, 0.020126720428466797, 0.020195392608642577, 0.02016249656677246, 0.02007040023803711, 0.02028339195251465, 0.020200447082519533, 0.02022400093078613, 0.02020351982116699, 0.02026905632019043, 0.020372512817382813, 0.020278240203857423, 0.020368383407592772, 0.020162559509277343, 0.0204769287109375, 0.021015552520751952, 0.020327423095703127, 0.020190208435058594, 0.020032512664794923, 0.019955711364746095, 0.020406272888183592, 0.02030899238586426, 0.020246528625488282, 0.02024550437927246, 0.020190208435058594, 0.020315135955810547, 0.02025881576538086, 0.02020966339111328, 0.04391526412963867, 0.019556352615356445, 0.0194703369140625, 0.01962291145324707, 0.01949388885498047, 0.01964339256286621, 0.02002739143371582, 0.01942323112487793, 0.019357696533203125, 0.019296255111694336, 0.019309568405151366, 0.019731456756591798, 0.020345855712890625, 0.020303871154785155, 0.019999744415283204, 0.02030899238586426, 0.02022809600830078, 0.02022809600830078, 0.02028646469116211, 0.020157440185546875, 0.020344831466674804, 0.020315135955810547, 0.020353023529052734, 0.020099071502685546, 0.019663871765136717, 0.019560447692871095, 0.019535871505737306, 0.019458047866821288, 0.019479551315307618, 0.019519487380981446, 0.019553279876708983, 0.019589120864868165, 0.019518463134765626, 0.019500032424926757, 0.01945088005065918, 0.019501056671142578, 0.019952640533447266, 0.020179967880249023, 0.020221952438354493, 0.02024550437927246, 0.020157440185546875, 0.02009600067138672, 0.02022604751586914, 0.020164608001708984, 0.020365312576293947, 0.020338687896728515, 0.020237312316894532, 0.02027622413635254, 0.02022604751586914, 0.020199424743652345, 0.020343807220458983, 0.020328447341918944, 0.020205568313598633, 0.020162559509277343, 0.020285472869873047, 0.02021993637084961, 0.020214719772338866, 0.020336639404296874, 0.020273151397705077, 0.020164608001708984, 0.02021171188354492, 0.01944473648071289, 0.019750911712646483, 0.044450817108154295, 0.019489791870117186, 0.019546112060546874, 0.01944166374206543, 0.019371007919311522, 0.019351551055908203, 0.019405824661254883, 0.019426368713378907, 0.019482559204101562, 0.01941196823120117, 0.019332096099853514, 0.019368959426879884, 0.01946214485168457, 0.0194969596862793, 0.019611648559570313, 0.019578943252563475, 0.01955526351928711, 0.02004377555847168, 0.02020966339111328, 0.02025062370300293, 0.020132863998413086, 0.020262912750244142, 0.020411392211914063, 0.020392959594726562, 0.020173824310302735, 0.020156415939331054, 0.020149248123168945, 0.020167680740356447, 0.020146175384521483, 0.02025881576538086, 0.020208639144897463, 0.02031001663208008, 0.020305919647216796, 0.020187135696411132, 0.02024038314819336, 0.020195327758789062, 0.020199424743652345, 0.02018611145019531, 0.02023423957824707, 0.020281343460083007, 0.020161535263061522, 0.020141056060791016, 0.020146175384521483, 0.02021785545349121, 0.02018611145019531, 0.02023526382446289, 0.0202106876373291, 0.020241472244262697, 0.020245439529418947, 0.019703807830810546, 0.019560447692871095, 0.019599359512329103, 0.019527679443359376, 0.019488800048828126, 0.019586015701293945, 0.019599359512329103, 0.019535871505737306, 0.01948057556152344, 0.01942118453979492, 0.019373056411743163, 0.019479551315307618, 0.019543039321899415, 0.01982259178161621, 0.044063743591308595, 0.019581951141357423, 0.019532800674438477, 0.019489791870117186, 0.019562496185302734, 0.019557376861572266, 0.019491840362548828, 0.019546112060546874, 0.01942937660217285, 0.019339263916015623, 0.019567615509033204, 0.019539968490600586, 0.019520511627197267, 0.019345407485961915, 0.01927884864807129, 0.01940275192260742, 0.019366912841796875, 0.019557376861572266, 0.019628032684326172, 0.019509248733520508, 0.019334144592285156, 0.01939148712158203, 0.019334144592285156, 0.019377151489257814, 0.019342336654663086, 0.019333120346069335, 0.019319807052612305, 0.019325952529907226, 0.019383359909057617, 0.019322816848754882, 0.019386367797851564, 0.01943244743347168, 0.019507200241088866, 0.019522560119628905, 0.019523584365844726, 0.019558399200439454, 0.019721216201782226, 0.01954921531677246, 0.019512319564819337, 0.019634143829345703, 0.019499008178710937, 0.019516416549682617, 0.019524608612060547, 0.019520511627197267, 0.019526655197143555, 0.019475456237792968, 0.019582975387573243, 0.019564544677734375, 0.019538944244384765, 0.019509248733520508, 0.019536895751953123, 0.019569664001464843, 0.019538944244384765, 0.019531839370727538, 0.01954502487182617, 0.019569664001464843, 0.0194969596862793, 0.01945088005065918, 0.01951955223083496, 0.02041644859313965, 0.02039708709716797, 0.020190176010131837, 0.020192256927490236, 0.04383846282958984, 0.019486719131469727, 0.019507200241088866, 0.01964134407043457, 0.01968332862854004, 0.019500032424926757, 0.019562496185302734, 0.01949286460876465, 0.01942323112487793, 0.01948467254638672, 0.019577856063842772, 0.019374080657958984, 0.019354623794555666, 0.019550207138061524, 0.021597183227539063, 0.02074118423461914, 0.020300735473632814, 0.020200447082519533, 0.02021990394592285, 0.02025369644165039, 0.02019327926635742, 0.020200447082519533, 0.020587520599365236, 0.020340736389160157, 0.02024860763549805, 0.02024239921569824, 0.020220928192138672, 0.02025984001159668, 0.020347904205322266, 0.02005299186706543, 0.019965951919555663, 0.020171775817871093, 0.02023526382446289, 0.020162559509277343, 0.02062848091125488, 0.020468736648559572, 0.021319679260253906, 0.020706304550170897, 0.020364288330078126, 0.02044108772277832, 0.020363264083862305, 0.020254720687866212, 0.020195327758789062, 0.020158464431762696, 0.02012876892089844, 0.019971071243286134, 0.019389440536499023, 0.019570688247680663, 0.019463199615478516, 0.019528671264648436, 0.019507200241088866, 0.019552255630493166, 0.019551231384277345, 0.019548160552978516, 0.019647487640380858, 0.019537919998168944, 0.019539968490600586, 0.020363264083862305, 0.02025881576538086, 0.02023526382446289, 0.02023321533203125, 0.020281343460083007, 0.020231168746948244, 0.044044288635253906, 0.019486719131469727, 0.019364864349365234, 0.019422208786010742, 0.019577856063842772, 0.019527679443359376, 0.019518463134765626, 0.01961881637573242, 0.019942399978637695, 0.01960960006713867, 0.019505151748657225, 0.019591167449951173, 0.019475456237792968, 0.019515392303466796, 0.019458047866821288, 0.019529727935791014, 0.019581951141357423, 0.019540992736816407, 0.019551231384277345, 0.01946009635925293, 0.01945599937438965, 0.019479551315307618, 0.01948876762390137, 0.019594240188598632, 0.019524608612060547, 0.019514368057250975, 0.019438592910766602, 0.019551231384277345, 0.019341312408447265, 0.019386367797851564, 0.019343360900878907, 0.019834880828857423, 0.019714048385620117, 0.019581951141357423, 0.019551231384277345, 0.019507200241088866, 0.019675167083740234, 0.019625951766967773, 0.01946828842163086, 0.019575807571411134, 0.01963827133178711, 0.019748863220214845, 0.01966592025756836, 0.01940787124633789, 0.0193832950592041, 0.0202926082611084, 0.020148223876953125, 0.020154367446899413, 0.020404224395751954, 0.020257791519165038, 0.02016972732543945, 0.020231168746948244, 0.02024448013305664, 0.02031827163696289, 0.020296640396118164, 0.02006425666809082, 0.02012473678588867, 0.020514751434326173, 0.02024140739440918, 0.020642847061157227, 0.020385759353637695, 0.02025881576538086, 0.020265983581542968, 0.04407910537719727, 0.01964134407043457, 0.019686431884765626, 0.019511264801025392, 0.019408895492553712, 0.019358720779418945, 0.019491840362548828, 0.019555328369140625, 0.019561471939086913, 0.01942732810974121, 0.019518463134765626, 0.019543039321899415, 0.019501056671142578, 0.019561471939086913, 0.01958502388000488, 0.019539968490600586, 0.01943142318725586, 0.019536895751953123, 0.019709951400756837, 0.019595264434814453, 0.019430400848388672, 0.019536928176879884, 0.01944265556335449, 0.019369983673095705, 0.01947750473022461, 0.019561471939086913, 0.019528703689575197, 0.019544095993041993, 0.019529695510864257, 0.019568639755249022, 0.019598335266113282, 0.019752960205078125, 0.01973151969909668, 0.019793920516967774, 0.02035500717163086, 0.019987455368041994, 0.020392959594726562, 0.021120000839233398, 0.020279296875, 0.020282367706298828, 0.02023526382446289, 0.020214784622192384, 0.02025267219543457, 0.02017791938781738, 0.020206592559814454, 0.020418560028076172, 0.02027519989013672, 0.020183040618896485, 0.020255775451660157, 0.02030281639099121, 0.020139007568359374, 0.020289535522460937, 0.02018611145019531, 0.019962879180908204, 0.020106239318847655, 0.020050943374633787, 0.019778560638427735, 0.019348480224609374, 0.019486719131469727, 0.01940377616882324, 0.019476512908935546, 0.01951024055480957, 0.019513343811035155, 0.043886592864990234, 0.019506175994873046, 0.019521535873413084, 0.019764223098754884, 0.019554304122924804, 0.019523584365844726, 0.0195020809173584, 0.019595264434814453, 0.019559423446655275, 0.01946316719055176, 0.019500032424926757, 0.01943552017211914, 0.019620864868164063, 0.02089369583129883, 0.02066534423828125, 0.0204400634765625, 0.020248575210571287, 0.02024038314819336, 0.02019327926635742, 0.020279296875, 0.020205568313598633, 0.020212736129760742, 0.02023526382446289, 0.020057088851928712, 0.02004275131225586, 0.019920896530151368, 0.020113407135009767, 0.020142080307006836, 0.02011955261230469, 0.02027724838256836, 0.02026188850402832, 0.020530176162719727, 0.02041548728942871, 0.020206592559814454, 0.02025881576538086, 0.020222976684570314, 0.020190208435058594, 0.020314111709594726, 0.020159488677978517, 0.02045747184753418, 0.020254720687866212, 0.02024140739440918, 0.020153343200683595, 0.020123647689819335, 0.020176895141601564, 0.02032537651062012, 0.02016972732543945, 0.02028441619873047, 0.019542015075683594, 0.019518463134765626, 0.019560447692871095, 0.0194969596862793, 0.01948876762390137, 0.019557376861572266, 0.019524608612060547, 0.019534847259521485, 0.019568639755249022, 0.019487743377685548, 0.0194969596862793, 0.019486719131469727, 0.019589120864868165, 0.019422208786010742, 0.019499008178710937]",tokens/s,49.28043557316377,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-1.4b,EleutherAI/pythia-1.4b,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1275.834368,3848.798208,0.0,3202.351104,3105.82784,s,10,0.24147904014587404,0.024147904014587402,0.0005044599971227972,0.023981328010559085,0.024345826911926268,0.024975265407562255,0.025478816204071043,"[0.02560470390319824, 0.024193920135498048, 0.024107711791992188, 0.02385955238342285, 0.023864992141723634, 0.02388572883605957, 0.0237938232421875, 0.024205951690673827, 0.024017663955688478, 0.023944992065429688]",tokens/s,10601.334171502176,kWh,2.849579786998999e-07,1.5612320748476186e-07,1.239496306187915e-06,1.6805774923725767e-06,tokens/kWh,152328590.11969078,MB,1276.280832,3848.798208,0.0,3202.351104,3188.152832,s,10,12.560411621093749,1.256041162109375,0.02114655036495289,1.2566647338867187,1.2818962890625,1.2891886840820312,1.2950226000976561,"[1.2964810791015624, 1.2592294921875, 1.2549832763671875, 1.2301461181640625, 1.2293675537109374, 1.2498057861328125, 1.27036181640625, 1.2802757568359375, 1.25834619140625, 1.23141455078125]",tokens/s,50.15759188512487,kWh,1.4725982288280166e-05,8.068005656063925e-06,3.874639679361247e-05,6.154038473795654e-05,tokens/kWh,1023718.0067732532,,s,629,12.753896476745602,0.020276464986876956,0.0029344008735951803,0.019969024658203126,0.02045972480773926,0.020908031845092774,0.04386340957641602,"[0.02070528030395508, 0.020354047775268554, 0.020370431900024414, 0.02124492835998535, 0.021164031982421876, 0.02126335906982422, 0.021398527145385742, 0.02124799919128418, 0.021144575119018554, 0.02103910446166992, 0.020414464950561522, 0.020752384185791017, 0.02122444725036621, 0.021200895309448242, 0.021166080474853514, 0.02107494354248047, 0.021120000839233398, 0.02123673629760742, 0.020541471481323244, 0.020411392211914063, 0.02013795280456543, 0.02007756805419922, 0.02003763198852539, 0.020918272018432618, 0.021218303680419923, 0.021421056747436523, 0.019960832595825196, 0.019557376861572266, 0.020304895401000975, 0.020333568572998048, 0.020575231552124023, 0.020436992645263673, 0.020371456146240235, 0.0203786563873291, 0.02093257522583008, 0.021024768829345702, 0.020485120773315428, 0.02048307228088379, 0.02053638458251953, 0.02037446403503418, 0.020382720947265624, 0.020145183563232423, 0.020263904571533202, 0.020380672454833985, 0.020332544326782227, 0.020358144760131838, 0.020419584274291993, 0.02040012741088867, 0.020378623962402344, 0.020348928451538087, 0.020362239837646484, 0.020328447341918944, 0.02041651153564453, 0.020331520080566406, 0.019620864868164063, 0.019512319564819337, 0.0194334716796875, 0.020167680740356447, 0.02040115165710449, 0.02031718444824219, 0.02037555122375488, 0.02045030403137207, 0.044227584838867184, 0.019522560119628905, 0.019489791870117186, 0.019510271072387696, 0.019533824920654298, 0.019515392303466796, 0.019489791870117186, 0.01965465545654297, 0.01949286460876465, 0.020237312316894532, 0.020355072021484375, 0.020084735870361328, 0.02045337677001953, 0.02044108772277832, 0.020374528884887694, 0.020376575469970702, 0.02006118392944336, 0.02030284881591797, 0.020371456146240235, 0.020379648208618165, 0.019475456237792968, 0.019507200241088866, 0.0194467830657959, 0.019582975387573243, 0.01945395278930664, 0.019354623794555666, 0.01927168083190918, 0.019327999114990235, 0.01927884864807129, 0.019384319305419923, 0.0192675838470459, 0.019412992477416992, 0.01926655960083008, 0.01928704071044922, 0.019339263916015623, 0.02007347106933594, 0.020024320602416993, 0.02006118392944336, 0.020023296356201172, 0.020125696182250977, 0.020148223876953125, 0.02008576011657715, 0.020148223876953125, 0.02039193534851074, 0.02042572784423828, 0.020436992645263673, 0.020360191345214843, 0.020487167358398437, 0.020344831466674804, 0.020420608520507814, 0.020410367965698242, 0.020410367965698242, 0.020402175903320312, 0.02044108772277832, 0.020398080825805662, 0.020528127670288086, 0.02045337677001953, 0.020386816024780274, 0.020366336822509764, 0.02044313621520996, 0.020349952697753908, 0.020397056579589845, 0.020379680633544922, 0.04500579071044922, 0.020388864517211915, 0.020336639404296874, 0.02040934371948242, 0.02042163276672363, 0.020540416717529295, 0.02044108772277832, 0.020518911361694335, 0.020380672454833985, 0.019586048126220702, 0.019580928802490235, 0.019500032424926757, 0.01948876762390137, 0.019500032424926757, 0.01947648048400879, 0.01962188720703125, 0.019500032424926757, 0.019522560119628905, 0.019481599807739256, 0.019557376861572266, 0.019562496185302734, 0.019528703689575197, 0.019554304122924804, 0.019587072372436523, 0.019491840362548828, 0.019555328369140625, 0.01948569679260254, 0.019479551315307618, 0.01947648048400879, 0.019490816116333007, 0.019553279876708983, 0.019568639755249022, 0.019526655197143555, 0.019551231384277345, 0.01955740737915039, 0.019505119323730467, 0.019482624053955077, 0.019517440795898438, 0.01946419143676758, 0.019589120864868165, 0.019535871505737306, 0.01946214485168457, 0.019487743377685548, 0.01960140800476074, 0.020320255279541014, 0.020385791778564453, 0.02040934371948242, 0.02046463966369629, 0.020331520080566406, 0.02041548728942871, 0.020386816024780274, 0.020389888763427736, 0.020381696701049806, 0.02040934371948242, 0.020338687896728515, 0.020419584274291993, 0.02043391990661621, 0.020397056579589845, 0.020329471588134765, 0.020376575469970702, 0.020357120513916017, 0.020377599716186523, 0.020331520080566406, 0.04388147354125976, 0.020065343856811524, 0.020249536514282226, 0.02024140739440918, 0.020237312316894532, 0.020303871154785155, 0.02028441619873047, 0.020315135955810547, 0.0194652156829834, 0.019374080657958984, 0.019397632598876953, 0.01945907211303711, 0.019486719131469727, 0.019482624053955077, 0.01923891258239746, 0.019329023361206055, 0.019466239929199217, 0.01946214485168457, 0.019449855804443358, 0.01942732810974121, 0.019514368057250975, 0.0195020809173584, 0.0194969596862793, 0.019505151748657225, 0.0195020809173584, 0.01942527961730957, 0.019579904556274414, 0.01947340774536133, 0.01943654441833496, 0.019503103256225587, 0.01947340774536133, 0.019514368057250975, 0.019575807571411134, 0.0194201602935791, 0.019458047866821288, 0.01947648048400879, 0.01945702362060547, 0.01948467254638672, 0.01944473648071289, 0.01949286460876465, 0.019526655197143555, 0.01940275192260742, 0.019555328369140625, 0.01942835235595703, 0.019475456237792968, 0.019491840362548828, 0.019481599807739256, 0.019544063568115236, 0.019315711975097655, 0.0192675838470459, 0.019260416030883788, 0.019293184280395507, 0.019306495666503908, 0.019286016464233398, 0.019482624053955077, 0.019508224487304687, 0.019314687728881837, 0.019401727676391603, 0.01925017547607422, 0.01926246452331543, 0.01927065658569336, 0.019300384521484373, 0.019288032531738282, 0.04381695938110351, 0.019474431991577147, 0.01946316719055176, 0.01945702362060547, 0.019504127502441407, 0.01944063949584961, 0.0194652156829834, 0.019526655197143555, 0.019503103256225587, 0.01945702362060547, 0.019404800415039062, 0.01944268798828125, 0.019305471420288087, 0.019280895233154297, 0.01938739204406738, 0.019308544158935546, 0.01926553535461426, 0.019194879531860352, 0.019224576950073242, 0.01924608039855957, 0.01924300765991211, 0.01949286460876465, 0.01946112060546875, 0.019560447692871095, 0.01945907211303711, 0.01942118453979492, 0.019353631973266602, 0.01927881622314453, 0.01922764778137207, 0.019333120346069335, 0.01925119972229004, 0.019369983673095705, 0.019478527069091797, 0.019464256286621094, 0.019474367141723632, 0.01943449592590332, 0.01945395278930664, 0.019408895492553712, 0.019486719131469727, 0.02039910316467285, 0.020373504638671876, 0.020363264083862305, 0.020267040252685546, 0.020330463409423828, 0.020291584014892578, 0.02031718444824219, 0.01923072052001953, 0.019321855545043946, 0.019316736221313476, 0.01923788833618164, 0.019353599548339845, 0.019305471420288087, 0.01926246452331543, 0.019300352096557616, 0.019333120346069335, 0.019400703430175782, 0.019339263916015623, 0.01926553535461426, 0.019374080657958984, 0.019530752182006835, 0.019969024658203126, 0.019983360290527344, 0.019554304122924804, 0.04390911865234375, 0.019574783325195313, 0.019544063568115236, 0.019508224487304687, 0.019512319564819337, 0.019563520431518554, 0.019490816116333007, 0.019573759078979493, 0.01950828742980957, 0.0195665283203125, 0.01960038375854492, 0.019553279876708983, 0.019501056671142578, 0.01945497512817383, 0.0194467830657959, 0.019560447692871095, 0.019487743377685548, 0.019515392303466796, 0.019527679443359376, 0.01948876762390137, 0.0195020809173584, 0.019546112060546874, 0.019466239929199217, 0.020892671585083008, 0.020764671325683593, 0.02047283172607422, 0.020455423355102538, 0.020406272888183592, 0.020398080825805662, 0.020281343460083007, 0.020406272888183592, 0.02046463966369629, 0.020344831466674804, 0.020144128799438478, 0.020092927932739257, 0.020039680480957032, 0.02042470359802246, 0.020374528884887694, 0.020371488571166992, 0.020474847793579103, 0.02042572784423828, 0.020403200149536133, 0.020350976943969725, 0.020436992645263673, 0.02030080032348633, 0.019607551574707033, 0.01967513656616211, 0.019617824554443358, 0.019503072738647462, 0.019362815856933592, 0.019405824661254883, 0.019475456237792968, 0.019525632858276368, 0.01949286460876465, 0.019507200241088866, 0.019664928436279296, 0.0196976318359375, 0.019513343811035155, 0.019506175994873046, 0.019555328369140625, 0.019513343811035155, 0.01948467254638672, 0.0194847354888916, 0.043734977722167966, 0.01948876762390137, 0.01947750473022461, 0.0195020809173584, 0.019481599807739256, 0.019458047866821288, 0.019471359252929688, 0.019557376861572266, 0.01944166374206543, 0.01943449592590332, 0.020115455627441405, 0.019501056671142578, 0.019750911712646483, 0.020393983840942383, 0.020279296875, 0.020289535522460937, 0.020175872802734376, 0.020017152786254884, 0.0202106876373291, 0.020343807220458983, 0.02026188850402832, 0.020001792907714845, 0.020238336563110353, 0.020478975296020507, 0.02021171188354492, 0.020242431640625, 0.02027827262878418, 0.02025164794921875, 0.02030899238586426, 0.02025881576538086, 0.02027827262878418, 0.02030182456970215, 0.02027827262878418, 0.02030080032348633, 0.020257791519165038, 0.020246528625488282, 0.02025164794921875, 0.020324352264404297, 0.02030182456970215, 0.020340736389160157, 0.020392959594726562, 0.020292640686035156, 0.020290592193603515, 0.020275136947631837, 0.020315135955810547, 0.020487167358398437, 0.020333568572998048, 0.020340736389160157, 0.02024550437927246, 0.02023628807067871, 0.02024345588684082, 0.020980768203735352, 0.020469728469848632, 0.020322303771972656, 0.02038374328613281, 0.02084966468811035, 0.02032640075683594, 0.020349952697753908, 0.020304895401000975, 0.02030182456970215, 0.02030080032348633, 0.020247552871704103, 0.020346879959106445, 0.045080577850341794, 0.02028646469116211, 0.020323328018188477, 0.020152320861816408, 0.020926464080810548, 0.020344831466674804, 0.02003660774230957, 0.020320255279541014, 0.02023526382446289, 0.020033536911010744, 0.019998720169067383, 0.020141056060791016, 0.020063232421875, 0.0200263671875, 0.019958784103393554, 0.02007040023803711, 0.020269088745117188, 0.020373472213745115, 0.020374528884887694, 0.020279296875, 0.020345855712890625, 0.020313087463378905, 0.020321279525756835, 0.020379648208618165, 0.020351999282836913, 0.02030899238586426, 0.020556800842285155, 0.020509695053100584, 0.020248575210571287, 0.020364288330078126, 0.02045849609375, 0.02043391990661621, 0.02032537651062012, 0.02031001663208008, 0.02030284881591797, 0.02101862335205078, 0.021251071929931642, 0.020554752349853517, 0.020332544326782227, 0.020374528884887694, 0.02032640075683594, 0.020330495834350586, 0.020356096267700196, 0.02031001663208008, 0.02037660789489746, 0.02031920051574707, 0.020372480392456056, 0.02043084716796875, 0.020368383407592772, 0.02042265510559082, 0.020287488937377928, 0.02032640075683594, 0.02039091110229492, 0.020371456146240235, 0.020876287460327148, 0.020567039489746093, 0.02027622413635254, 0.020340736389160157, 0.02040934371948242, 0.02028646469116211, 0.01942835235595703, 0.019503103256225587, 0.01947238349914551, 0.04482867050170898, 0.020380672454833985, 0.020378623962402344, 0.020414464950561522, 0.020526079177856444, 0.02045849609375, 0.02033459281921387, 0.020371456146240235, 0.020264959335327147, 0.020323328018188477, 0.02026700782775879, 0.020314111709594726, 0.020319232940673827, 0.02031718444824219, 0.020335615158081053, 0.020533248901367186, 0.02035916709899902, 0.020351999282836913, 0.02025984001159668, 0.02027008056640625, 0.020229120254516602, 0.020526079177856444, 0.02068889617919922, 0.021384191513061524, 0.020746240615844725, 0.020366336822509764, 0.020341760635375978, 0.020356096267700196, 0.02022809600830078, 0.02020147132873535, 0.02025984001159668, 0.020366336822509764, 0.020331520080566406, 0.020254751205444337, 0.01949590492248535, 0.01943756866455078, 0.01944473648071289, 0.01948057556152344, 0.01946316719055176, 0.01946828842163086, 0.019501056671142578, 0.0194201602935791, 0.019486719131469727, 0.019451904296875, 0.01942835235595703, 0.019400703430175782, 0.019507200241088866, 0.019582975387573243, 0.01945088005065918, 0.019495935440063478, 0.0194201602935791, 0.019478527069091797, 0.019551231384277345, 0.01926246452331543, 0.019342336654663086, 0.019517440795898438, 0.019481599807739256, 0.01945292854309082, 0.019528703689575197, 0.019449855804443358, 0.01949388885498047, 0.019563520431518554, 0.01947750473022461, 0.04393471908569336, 0.019523584365844726, 0.019506175994873046, 0.019482624053955077, 0.01948057556152344, 0.019487808227539063, 0.01951123237609863, 0.019580928802490235, 0.019521535873413084, 0.01946316719055176, 0.01945292854309082, 0.01944268798828125, 0.020059135437011717, 0.020361215591430663, 0.020371456146240235, 0.019590143203735352, 0.0194969596862793, 0.019479551315307618, 0.019545087814331053, 0.019577856063842772, 0.019490816116333007, 0.01947238349914551, 0.01948467254638672, 0.01957683181762695, 0.019494911193847657, 0.019497983932495116, 0.019504127502441407, 0.019557376861572266, 0.0194652156829834, 0.0196177921295166, 0.019575807571411134, 0.019567615509033204, 0.019537919998168944, 0.01946214485168457, 0.01958198356628418, 0.01953481674194336, 0.019504127502441407, 0.01941196823120117, 0.019481599807739256, 0.019607551574707033, 0.01945395278930664, 0.01945599937438965, 0.01944883155822754, 0.01946419143676758, 0.019505151748657225, 0.019615808486938478, 0.019454912185668947, 0.019512319564819337, 0.01946316719055176, 0.01943654441833496, 0.01943142318725586, 0.01946316719055176, 0.01944063949584961, 0.019508224487304687, 0.01946009635925293, 0.019536895751953123, 0.019571712493896484, 0.019436607360839842, 0.019512256622314452, 0.019511295318603517, 0.019533824920654298, 0.01942425537109375, 0.01945702362060547]",tokens/s,49.3182613757973,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,x,x,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/x/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66932460-5881fe5657b64c5403dd4559;46bbe37e-f455-4d29-b2c7-72c4a0f26fed) Repository Not Found for url: https://huggingface.co/x/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: x is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,01-ai/Yi-34B,01-ai/Yi-34B,cuda,0,42,,,True,,,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",llama,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/xglm-4.5B,,cuda,0,42,,,,,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,d35829e539df8480b726c647eeabf91e41eae047,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,,MB,1384.71424,11335.630848,0.0,10689.183744,9357.549568,s,10,0.5806327590942382,0.05806327590942382,0.0006150812332001308,0.05790153503417969,0.05850998077392578,0.05911591796875,0.059600667724609376,"[0.05972185516357422, 0.057774398803710936, 0.057828575134277346, 0.05752217483520508, 0.05797449493408203, 0.05802070236206055, 0.058375328063964844, 0.05752883148193359, 0.05764361572265625, 0.058242782592773434]",tokens/s,4408.983061847024,kWh,6.824863444190677e-07,3.739691475481056e-07,3.1574546332414543e-06,4.213910125208628e-06,tokens/kWh,60751177.02879949,MB,1385.04192,11335.630848,0.0,10689.183744,9490.407424,s,10,19.066519165039065,1.9066519165039064,0.004705715644958097,1.9076243896484375,1.912862060546875,1.91339599609375,1.9138231445312501,"[1.900245361328125, 1.901219482421875, 1.90844189453125, 1.9002161865234375, 1.912743408203125, 1.913929931640625, 1.9088603515625, 1.9046826171875, 1.906806884765625, 1.909373046875]",tokens/s,33.042213659805654,kWh,2.2234201332884884e-05,1.2183572898631048e-05,8.273878075036202e-05,0.00011715655498187796,tokens/kWh,537741.998386219,,s,629,19.556330526351925,0.031091145510893366,0.006896125826306854,0.030336000442504882,0.030623960876464844,0.030774271774291994,0.08797329864501953,"[0.030104576110839845, 0.029843456268310548, 0.029835296630859376, 0.029777887344360352, 0.02991923141479492, 0.02980147171020508, 0.03020185661315918, 0.03019878387451172, 0.030243839263916016, 0.029743104934692382, 0.030219263076782226, 0.02981990432739258, 0.029857791900634766, 0.030270463943481447, 0.030512128829956055, 0.029892608642578124, 0.030327808380126952, 0.030716928482055664, 0.03100160026550293, 0.030443519592285157, 0.030327808380126952, 0.029813760757446288, 0.030055423736572266, 0.030131200790405273, 0.030311424255371092, 0.030295040130615233, 0.030110719680786133, 0.029944831848144532, 0.029944831848144532, 0.02992233657836914, 0.030110687255859376, 0.029852672576904295, 0.030049280166625978, 0.03000422477722168, 0.030019584655761718, 0.02992742347717285, 0.02995712089538574, 0.029921279907226563, 0.030166015625, 0.029885440826416015, 0.029855743408203125, 0.029816831588745117, 0.030068735122680663, 0.030043136596679686, 0.03078963279724121, 0.03032678413391113, 0.030628864288330077, 0.030362688064575195, 0.030394304275512696, 0.030437376022338865, 0.02999398422241211, 0.029932607650756837, 0.030230464935302733, 0.031065088272094726, 0.03082956886291504, 0.030084096908569335, 0.030195711135864257, 0.029970432281494142, 0.030518272399902343, 0.029869056701660155, 0.02999091148376465, 0.02993561553955078, 0.08813670349121094, 0.03056947135925293, 0.030520320892333985, 0.030486528396606444, 0.03040563201904297, 0.03038310432434082, 0.030131200790405273, 0.030854143142700196, 0.02994380760192871, 0.030111743927001954, 0.02997657585144043, 0.029742143630981446, 0.029866943359375, 0.02978099250793457, 0.0299550724029541, 0.030612480163574218, 0.029867008209228517, 0.02989676856994629, 0.029805503845214843, 0.030038015365600586, 0.030102527618408204, 0.030481407165527344, 0.029916160583496092, 0.03002060890197754, 0.0301844482421875, 0.0304005126953125, 0.02991619110107422, 0.029802528381347657, 0.029846464157104492, 0.029869056701660155, 0.029863935470581054, 0.0301711368560791, 0.029875200271606447, 0.02977689552307129, 0.03018342399597168, 0.030505983352661133, 0.029891584396362306, 0.03037900733947754, 0.03061452865600586, 0.030455808639526367, 0.03038310432434082, 0.030494720458984374, 0.029893632888793945, 0.030475263595581056, 0.03037696075439453, 0.030478368759155272, 0.03007177543640137, 0.030255104064941408, 0.030448640823364258, 0.030479360580444335, 0.030500864028930662, 0.029973503112792968, 0.029901824951171874, 0.029929567337036132, 0.03033795166015625, 0.03077120018005371, 0.030537727355957032, 0.030500864028930662, 0.029875200271606447, 0.02998784065246582, 0.02992639923095703, 0.030092287063598632, 0.02998067283630371, 0.08810189056396485, 0.029808639526367187, 0.03017932891845703, 0.030468095779418947, 0.030245887756347657, 0.030645248413085937, 0.03118396759033203, 0.030800800323486328, 0.03034316825866699, 0.030471168518066406, 0.030101503372192383, 0.030484479904174806, 0.03041996765136719, 0.029831167221069335, 0.03035852813720703, 0.030026752471923827, 0.029870080947875976, 0.030243839263916016, 0.030432256698608398, 0.030507007598876954, 0.03123404884338379, 0.03076300811767578, 0.030324735641479493, 0.03057459259033203, 0.03040563201904297, 0.03039948844909668, 0.03034316825866699, 0.030402559280395508, 0.03040460777282715, 0.030482431411743165, 0.029847583770751952, 0.030039072036743164, 0.030203840255737305, 0.030465024948120117, 0.03040358352661133, 0.030417919158935547, 0.030538751602172853, 0.030457855224609375, 0.03038412857055664, 0.030091264724731445, 0.029864992141723633, 0.02986595153808594, 0.029862911224365234, 0.030050304412841795, 0.03042099189758301, 0.030543872833251953, 0.03037491226196289, 0.03075584030151367, 0.02999193572998047, 0.030263328552246095, 0.030429216384887697, 0.03046803283691406, 0.030436351776123048, 0.02994278335571289, 0.029957183837890623, 0.030086080551147462, 0.029920255661010742, 0.029912063598632813, 0.03000217628479004, 0.030202880859375, 0.030437376022338865, 0.029944831848144532, 0.02989779281616211, 0.0878611831665039, 0.03036262321472168, 0.030328832626342773, 0.030444543838500978, 0.03037491226196289, 0.030365695953369142, 0.02981990432739258, 0.030105600357055663, 0.029897727966308595, 0.029875200271606447, 0.030061567306518554, 0.029912063598632813, 0.03035443115234375, 0.030734336853027344, 0.030491647720336915, 0.02999091148376465, 0.029920255661010742, 0.03035136032104492, 0.03036774444580078, 0.029898752212524415, 0.02988857650756836, 0.029796287536621092, 0.02994892883300781, 0.030507007598876954, 0.030398464202880858, 0.029922304153442384, 0.029891584396362306, 0.029921279907226563, 0.029900800704956054, 0.029868032455444334, 0.029894655227661132, 0.029983808517456054, 0.02984441566467285, 0.02997145652770996, 0.03033907127380371, 0.030038015365600586, 0.02993561553955078, 0.030518272399902343, 0.030499839782714845, 0.030574655532836913, 0.030612415313720703, 0.030446592330932616, 0.029875200271606447, 0.030089279174804688, 0.029985727310180663, 0.030504959106445313, 0.03040153694152832, 0.030646272659301758, 0.030414848327636718, 0.029965311050415038, 0.030283775329589844, 0.03056230354309082, 0.030336000442504882, 0.030480384826660156, 0.030030847549438477, 0.030123008728027343, 0.0299683837890625, 0.02997248077392578, 0.03001241683959961, 0.02998374366760254, 0.030039072036743164, 0.029973472595214844, 0.029884416580200194, 0.08879513549804688, 0.03039641571044922, 0.030567424774169922, 0.03075071907043457, 0.03101798439025879, 0.030859264373779297, 0.030496768951416016, 0.030650367736816408, 0.030514175415039063, 0.030479360580444335, 0.030940160751342774, 0.030685184478759765, 0.029908992767333983, 0.03032268714904785, 0.029880319595336914, 0.030213119506835938, 0.029867008209228517, 0.029817855834960938, 0.03022336006164551, 0.03060736083984375, 0.03053260803222656, 0.03001241683959961, 0.029903871536254883, 0.030054399490356445, 0.030239744186401366, 0.030486528396606444, 0.030443519592285157, 0.030539775848388673, 0.030492671966552733, 0.030492671966552733, 0.029879295349121093, 0.03019059181213379, 0.02991923141479492, 0.02991001510620117, 0.02987014389038086, 0.030530496597290038, 0.030031871795654298, 0.029978624343872072, 0.03033087921142578, 0.03080601692199707, 0.030616575241088868, 0.030488576889038086, 0.03061043167114258, 0.030623743057250977, 0.02997555160522461, 0.02995199966430664, 0.03038003158569336, 0.03077631950378418, 0.03059507179260254, 0.029911039352416992, 0.029860864639282225, 0.03035545539855957, 0.03060223960876465, 0.03061043167114258, 0.030437376022338865, 0.030499839782714845, 0.029853696823120116, 0.03020595169067383, 0.030496768951416016, 0.030619647979736327, 0.030612480163574218, 0.030523391723632814, 0.030495744705200195, 0.08832819366455077, 0.02975948715209961, 0.029772800445556642, 0.030340095520019532, 0.030274560928344726, 0.030442495346069336, 0.03040460777282715, 0.030604288101196288, 0.030323711395263672, 0.03036262321472168, 0.030478336334228515, 0.03021414375305176, 0.03040460777282715, 0.030481407165527344, 0.030546943664550782, 0.030555135726928712, 0.030342144012451173, 0.03041996765136719, 0.030427135467529298, 0.030441471099853516, 0.03038515281677246, 0.03017728042602539, 0.030461952209472655, 0.030474239349365235, 0.030315519332885742, 0.02996428871154785, 0.029845504760742186, 0.029924352645874022, 0.0299550724029541, 0.030278656005859376, 0.030254079818725587, 0.031038463592529295, 0.031352832794189454, 0.03059916877746582, 0.030511104583740234, 0.02999193572998047, 0.03037183952331543, 0.03058073616027832, 0.030694400787353516, 0.030964736938476563, 0.030525440216064452, 0.029920255661010742, 0.030345216751098632, 0.030005247116088866, 0.029878271102905272, 0.02993971252441406, 0.030485504150390624, 0.03101388740539551, 0.030500864028930662, 0.030502912521362304, 0.030361600875854492, 0.030468095779418947, 0.03055820846557617, 0.030451711654663087, 0.030440448760986328, 0.0307640323638916, 0.030437376022338865, 0.03053670310974121, 0.03036262321472168, 0.030538751602172853, 0.030461952209472655, 0.029965311050415038, 0.030406656265258788, 0.08879933166503906, 0.029881248474121092, 0.029874176025390626, 0.029879295349121093, 0.030308351516723633, 0.030487552642822265, 0.03038515281677246, 0.030875648498535156, 0.030427135467529298, 0.029978624343872072, 0.03037900733947754, 0.029825023651123047, 0.02993561553955078, 0.030274560928344726, 0.030457855224609375, 0.030670848846435547, 0.029844480514526366, 0.03016908836364746, 0.030278656005859376, 0.03040153694152832, 0.030632959365844727, 0.030482431411743165, 0.030643199920654295, 0.03076198387145996, 0.030515199661254884, 0.030398464202880858, 0.030402559280395508, 0.030526464462280273, 0.029900800704956054, 0.030023712158203125, 0.030428192138671876, 0.030513151168823242, 0.030415807723999023, 0.030093311309814453, 0.030382080078125, 0.030013439178466796, 0.030457855224609375, 0.03052851104736328, 0.030686208724975586, 0.030737407684326173, 0.030629888534545898, 0.03060223960876465, 0.030534656524658203, 0.031015935897827147, 0.02995916748046875, 0.02998271942138672, 0.029997055053710937, 0.030673919677734376, 0.030478336334228515, 0.030445568084716795, 0.03040460777282715, 0.030487552642822265, 0.029946943283081055, 0.030144447326660155, 0.030439424514770507, 0.030111743927001954, 0.029953023910522462, 0.029962240219116212, 0.029888511657714844, 0.030187519073486328, 0.0299550724029541, 0.03000217628479004, 0.029870080947875976, 0.08782848358154297, 0.02974515151977539, 0.030303232192993163, 0.029858848571777345, 0.02981679916381836, 0.030200832366943358, 0.030327808380126952, 0.03058278465270996, 0.03041587257385254, 0.0299182071685791, 0.02982809638977051, 0.031097856521606446, 0.0307957763671875, 0.030418943405151368, 0.03040563201904297, 0.03058585548400879, 0.030513151168823242, 0.0299233283996582, 0.029880319595336914, 0.02979532814025879, 0.029844480514526366, 0.029930496215820314, 0.029898752212524415, 0.030069759368896484, 0.029887487411499023, 0.029890623092651367, 0.030273536682128906, 0.031237056732177734, 0.03061452865600586, 0.030439424514770507, 0.029896703720092774, 0.030144512176513674, 0.029970432281494142, 0.029877248764038085, 0.02994380760192871, 0.029922304153442384, 0.02995712089538574, 0.02996326446533203, 0.03014656066894531, 0.030765056610107422, 0.03038105583190918, 0.030066688537597655, 0.03053670310974121, 0.030018560409545897, 0.030085119247436523, 0.03057459259033203, 0.030450687408447266, 0.03073945617675781, 0.03040358352661133, 0.030489599227905274, 0.03042918395996094, 0.03017420768737793, 0.03054080009460449, 0.02992639923095703, 0.029981760025024413, 0.030638015747070313, 0.030480384826660156, 0.03052441596984863, 0.03022643280029297, 0.030038015365600586, 0.02995097541809082, 0.03020595169067383, 0.030440448760986328, 0.08801689910888671, 0.029865983963012696, 0.029885440826416015, 0.029886463165283202, 0.029708288192749024, 0.02975334358215332, 0.030003200531005858, 0.030520320892333985, 0.029724672317504884, 0.02976153564453125, 0.030283775329589844, 0.030493696212768553, 0.029855743408203125, 0.029892608642578124, 0.02994175910949707, 0.030041088104248048, 0.02998886489868164, 0.0304005126953125, 0.029893632888793945, 0.030418943405151368, 0.030624832153320312, 0.030375871658325195, 0.030349311828613282, 0.030700544357299804, 0.030368831634521483, 0.030441408157348634, 0.030481407165527344, 0.030510080337524413, 0.030058496475219725, 0.030442495346069336, 0.030327808380126952, 0.030559232711791992, 0.030507007598876954, 0.030492671966552733, 0.0303687686920166, 0.030515199661254884, 0.030342144012451173, 0.030476287841796876, 0.029911039352416992, 0.03037491226196289, 0.030441471099853516, 0.03039948844909668, 0.030492671966552733, 0.030588960647583006, 0.030455808639526367, 0.030536672592163087, 0.030432256698608398, 0.030678016662597656, 0.0299233283996582, 0.029922304153442384, 0.03018854331970215, 0.030547967910766603, 0.030462976455688476, 0.030521343231201172, 0.03058892822265625, 0.030127103805541993, 0.03038515281677246, 0.030448640823364258, 0.029944831848144532, 0.030070783615112305, 0.030084096908569335, 0.03019878387451172, 0.030457855224609375, 0.08856575775146484, 0.030151679992675783, 0.03038617515563965, 0.03033907127380371, 0.03018649673461914, 0.02976870346069336, 0.029740095138549805, 0.03007583999633789, 0.030682111740112306, 0.031228927612304686, 0.0307640323638916, 0.030611455917358397, 0.02989673614501953, 0.029823999404907226, 0.02987414360046387, 0.030040063858032227, 0.030227455139160156, 0.03056435203552246, 0.030444543838500978, 0.030546943664550782, 0.03038719940185547, 0.030490623474121094, 0.029868032455444334, 0.03017523193359375, 0.03021107292175293, 0.030465024948120117, 0.030479360580444335, 0.030458879470825196, 0.029938688278198244, 0.02994380760192871, 0.030202880859375, 0.03057561683654785, 0.03038617515563965, 0.03038924789428711, 0.030478336334228515, 0.0305797119140625, 0.030455808639526367, 0.030687231063842774, 0.030456832885742188, 0.03055001640319824, 0.030461952209472655, 0.030423040390014647, 0.030488576889038086, 0.030535680770874023, 0.03056844711303711, 0.03053673553466797, 0.030519296646118164, 0.03048137664794922, 0.030495744705200195, 0.030065664291381834, 0.029949951171875, 0.030034944534301757, 0.03022336006164551, 0.03053260803222656, 0.03038310432434082, 0.030652416229248046, 0.029973503112792968, 0.030414848327636718, 0.02987424087524414, 0.030026687622070312, 0.030011392593383788, 0.03000115203857422, 0.029936639785766602]",tokens/s,32.16349811394473,,,main,False,False,,,,,,,,,,,,,,,,,,,,,,, bfloat16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/xglm-7.5B,,cuda,0,42,,,,,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,d35829e539df8480b726c647eeabf91e41eae047,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,,MB,1412.95616,17796.95616,0.0,17150.509056,15280.037888,s,10,0.7948976974487305,0.07948976974487305,0.0012653188159965402,0.07898975753784179,0.07981931610107422,0.08153316268920899,0.0829042399597168,"[0.08324700927734376, 0.07900713348388672, 0.07892217254638671, 0.07943846130371093, 0.07897238159179687, 0.0793662109375, 0.07895059204101562, 0.0791526107788086, 0.07893456268310547, 0.07890656280517579]",tokens/s,3220.540213182735,kWh,9.356064524759041e-07,5.126679413529126e-07,5.204938547150007e-06,6.653212940978824e-06,tokens/kWh,38477650.16255998,MB,1413.28384,17796.95616,0.0,17150.509056,15412.539904,s,10,23.875788330078127,2.3875788330078125,0.004442179524753045,2.387416259765625,2.390898388671875,2.394737890625,2.3978094921875,"[2.398577392578125, 2.384237060546875, 2.387581787109375, 2.387875, 2.390045166015625, 2.38243115234375, 2.389617431640625, 2.387250732421875, 2.382696044921875, 2.3854765625]",tokens/s,26.38656329543438,kWh,2.8115896507119885e-05,1.5408518680882924e-05,0.00012326024755824757,0.00016678466274625038,tokens/kWh,377732.57422266394,,s,629,24.54879435729982,0.03902828991621591,0.009442083404149369,0.03785830307006836,0.038099301147460934,0.038223308563232423,0.11712037200927734,"[0.03879116821289062, 0.03846451187133789, 0.038316032409667966, 0.03855052947998047, 0.03863654327392578, 0.03841024017333984, 0.03848812866210938, 0.03845727920532226, 0.03841535949707031, 0.03841228866577148, 0.038621185302734375, 0.039378944396972655, 0.03846144104003906, 0.03816755294799805, 0.038133758544921875, 0.03811532974243164, 0.037940223693847655, 0.03806208038330078, 0.03787673568725586, 0.03791462326049805, 0.03779072189331055, 0.03782963180541992, 0.03779891204833984, 0.037855232238769534, 0.03792588806152344, 0.03785830307006836, 0.03791155242919922, 0.03783679962158203, 0.037806079864501956, 0.03789619064331055, 0.03787161636352539, 0.03784703826904297, 0.03787673568725586, 0.03791667175292969, 0.037822463989257815, 0.03786240005493164, 0.037826591491699216, 0.037829662322998045, 0.03784396743774414, 0.03780499267578125, 0.037822463989257815, 0.037875743865966795, 0.03782857513427734, 0.037850112915039064, 0.037817344665527344, 0.03791462326049805, 0.03787980651855469, 0.03787366485595703, 0.03793817520141601, 0.037897216796875, 0.03802009582519531, 0.03790643310546875, 0.03791872024536133, 0.03789516830444336, 0.03791257476806641, 0.03789209747314453, 0.038056961059570314, 0.037862430572509764, 0.03797603225708008, 0.03857612609863281, 0.038348800659179685, 0.03794841766357422, 0.11724697875976563, 0.03767398452758789, 0.037736446380615234, 0.037694465637207034, 0.03770163345336914, 0.037719039916992186, 0.03768012619018555, 0.03780915069580078, 0.03773747253417969, 0.03773235321044922, 0.03780915069580078, 0.03770163345336914, 0.037768192291259765, 0.037773311614990236, 0.03772927856445313, 0.037727294921875, 0.03775379180908203, 0.03767705535888672, 0.03780505752563477, 0.03767705535888672, 0.03780198287963867, 0.03776716613769531, 0.03780710220336914, 0.037820415496826174, 0.03788288116455078, 0.03769139099121094, 0.03790848159790039, 0.037771263122558595, 0.037874687194824216, 0.03779993438720703, 0.0377907829284668, 0.03809785461425781, 0.037910526275634765, 0.03777740859985351, 0.03786444854736328, 0.03778559875488281, 0.037820415496826174, 0.037738494873046875, 0.038233150482177736, 0.03790124893188476, 0.03785420989990235, 0.037852161407470705, 0.0379576301574707, 0.037849151611328125, 0.037889984130859374, 0.037820415496826174, 0.03791667175292969, 0.03791360092163086, 0.03790643310546875, 0.03785420989990235, 0.03813683319091797, 0.03785932922363281, 0.037905406951904294, 0.037905406951904294, 0.038075393676757815, 0.03800166320800781, 0.037967872619628903, 0.03785625457763672, 0.037953536987304685, 0.0378787841796875, 0.037991424560546876, 0.03783475112915039, 0.03790233612060547, 0.11716812896728515, 0.037975040435791016, 0.0380682258605957, 0.037768192291259765, 0.037684223175048825, 0.03775078582763672, 0.037748737335205076, 0.03784089660644531, 0.03779283142089844, 0.037775295257568356, 0.037814273834228515, 0.037736446380615234, 0.03772927856445313, 0.037784576416015625, 0.03782656097412109, 0.03804159927368164, 0.03782451248168945, 0.03782144165039063, 0.038703102111816406, 0.03787571334838867, 0.037822463989257815, 0.03782454299926758, 0.03806108856201172, 0.03798828887939453, 0.038035457611083984, 0.0380313606262207, 0.03804467010498047, 0.03782553482055664, 0.03783679962158203, 0.037771263122558595, 0.03781836700439453, 0.03786649703979492, 0.037838848114013675, 0.0378430061340332, 0.037881790161132814, 0.03784294509887695, 0.037819393157958986, 0.03778047943115234, 0.03787059020996094, 0.03789311981201172, 0.03790028762817383, 0.03777740859985351, 0.03786240005493164, 0.03775692749023438, 0.03799859237670898, 0.03787366485595703, 0.03793100738525391, 0.03793203353881836, 0.038063102722167966, 0.03787059020996094, 0.03793407821655274, 0.03788083267211914, 0.03789823913574219, 0.03779993438720703, 0.03785113525390625, 0.037921791076660154, 0.03792998504638672, 0.03784089660644531, 0.037953536987304685, 0.03790643310546875, 0.03806719970703125, 0.038152191162109376, 0.03799961471557617, 0.11808153533935548, 0.03801599884033203, 0.038019073486328124, 0.03786137771606445, 0.03773235321044922, 0.03768115234375, 0.037713920593261716, 0.037781505584716796, 0.03777740859985351, 0.03772415924072266, 0.037823486328125, 0.037733375549316404, 0.03785318374633789, 0.037720062255859374, 0.037718017578125, 0.03775897598266602, 0.03808153533935547, 0.03778252792358398, 0.037814273834228515, 0.03769142532348633, 0.037809120178222654, 0.03770982360839844, 0.03785830307006836, 0.03783065414428711, 0.037754878997802735, 0.03770675277709961, 0.038004737854003906, 0.03787673568725586, 0.03784806442260742, 0.03779993438720703, 0.037822463989257815, 0.038112255096435545, 0.03819417572021484, 0.03778559875488281, 0.03795251083374023, 0.037920768737792966, 0.037850112915039064, 0.037819393157958986, 0.03806412887573242, 0.03795251083374023, 0.03786342239379883, 0.037822463989257815, 0.03800166320800781, 0.037781505584716796, 0.03794432067871094, 0.03776204681396484, 0.037855232238769534, 0.037928958892822266, 0.03792998504638672, 0.038029312133789066, 0.03814092636108399, 0.03786956787109375, 0.038002689361572264, 0.037852161407470705, 0.03827814483642578, 0.03846451187133789, 0.038114303588867186, 0.037920768737792966, 0.03793407821655274, 0.03785420989990235, 0.037940223693847655, 0.03801497650146484, 0.038002689361572264, 0.11722649383544922, 0.03811635208129883, 0.03811840057373047, 0.037754878997802735, 0.037748737335205076, 0.03775897598266602, 0.037891071319580076, 0.03781631851196289, 0.037738494873046875, 0.037981182098388674, 0.03800678253173828, 0.037719039916992186, 0.03772108840942383, 0.03797507095336914, 0.03786748886108399, 0.037817344665527344, 0.03775590515136719, 0.03781119918823242, 0.03778252792358398, 0.03797196960449219, 0.037926910400390625, 0.038002689361572264, 0.03781324768066406, 0.03782144165039063, 0.03783270263671875, 0.037976062774658204, 0.03792899322509766, 0.03780707168579102, 0.03794944000244141, 0.037806079864501956, 0.03790848159790039, 0.03784908676147461, 0.037907455444335936, 0.037781505584716796, 0.03784703826904297, 0.038002689361572264, 0.038063102722167966, 0.037789695739746096, 0.037889022827148434, 0.03807846450805664, 0.037981182098388674, 0.03793305587768555, 0.03787059020996094, 0.037787647247314454, 0.03810508728027344, 0.03802316665649414, 0.03809692764282226, 0.03802313613891602, 0.03803340911865234, 0.03799039840698242, 0.0381102066040039, 0.0380313606262207, 0.03807852935791016, 0.037883838653564456, 0.038002689361572264, 0.038117374420166016, 0.03794636917114258, 0.0381399040222168, 0.0381736946105957, 0.03807231903076172, 0.03795251083374023, 0.03797196960449219, 0.038152191162109376, 0.11776000213623047, 0.037705726623535156, 0.037765121459960936, 0.03773952102661133, 0.037703678131103514, 0.03777433776855469, 0.03778047943115234, 0.03778559875488281, 0.03775078582763672, 0.0377446403503418, 0.03776716613769531, 0.03771187210083008, 0.037763072967529294, 0.03779481506347656, 0.0377784309387207, 0.037733375549316404, 0.03775078582763672, 0.03773235321044922, 0.03778867340087891, 0.03778566360473633, 0.03771692657470703, 0.03765760040283203, 0.03777433776855469, 0.03777536010742188, 0.037749759674072264, 0.037738494873046875, 0.037792766571044925, 0.03775283050537109, 0.037792800903320316, 0.037729248046875, 0.037787647247314454, 0.0378869743347168, 0.03773952102661133, 0.037720062255859374, 0.03784400177001953, 0.03774972915649414, 0.03783071899414062, 0.03775993728637695, 0.037958656311035156, 0.037868545532226565, 0.0379095344543457, 0.03787772750854492, 0.03797401428222656, 0.037820415496826174, 0.03783987045288086, 0.037746688842773435, 0.03782963180541992, 0.037889022827148434, 0.03794739151000977, 0.037855232238769534, 0.037921791076660154, 0.03779891204833984, 0.03779481506347656, 0.03779481506347656, 0.037969921112060545, 0.03807436752319336, 0.037907455444335936, 0.03785830307006836, 0.037868545532226565, 0.03784601593017578, 0.03790848159790039, 0.037792766571044925, 0.03788390350341797, 0.11712409973144532, 0.03759718322753906, 0.03772313690185547, 0.03768012619018555, 0.03767193603515625, 0.03774259185791016, 0.03763097763061524, 0.0377077751159668, 0.03776921463012695, 0.03807436752319336, 0.037884929656982425, 0.037969921112060545, 0.03805184173583984, 0.037928958892822266, 0.03788595199584961, 0.03794636917114258, 0.03790643310546875, 0.037684223175048825, 0.03787673568725586, 0.03781119918823242, 0.03784806442260742, 0.038353919982910156, 0.0381921272277832, 0.03803955078125, 0.038002689361572264, 0.037959712982177735, 0.03780604934692383, 0.03779891204833984, 0.03801190567016602, 0.038010879516601564, 0.03782556915283203, 0.037986270904541014, 0.03791974258422852, 0.03777433776855469, 0.037822463989257815, 0.03782656097412109, 0.038056961059570314, 0.037926910400390625, 0.038073345184326174, 0.038013950347900394, 0.038130687713623046, 0.03799859237670898, 0.03800371170043945, 0.03789823913574219, 0.03786444854736328, 0.037868545532226565, 0.03786444854736328, 0.03817267227172851, 0.03807436752319336, 0.037907455444335936, 0.03789926528930664, 0.03812454223632813, 0.03796889495849609, 0.03790848159790039, 0.03798323059082031, 0.038128639221191404, 0.03783679962158203, 0.03784806442260742, 0.038152191162109376, 0.03799347305297852, 0.03810713577270508, 0.03789113616943359, 0.0380302734375, 0.11763507080078126, 0.037804031372070314, 0.03779891204833984, 0.03812761688232422, 0.03837542343139649, 0.03808665466308594, 0.03774054336547852, 0.03773440170288086, 0.03779174423217774, 0.03770982360839844, 0.03769036865234375, 0.037720062255859374, 0.037776382446289065, 0.03769651031494141, 0.037776416778564456, 0.03797705459594727, 0.03786240005493164, 0.03782656097412109, 0.03783475112915039, 0.03770675277709961, 0.037751808166503906, 0.03768320083618164, 0.03783475112915039, 0.03783987045288086, 0.03779072189331055, 0.03771084976196289, 0.03786649703979492, 0.03781836700439453, 0.03781631851196289, 0.037749759674072264, 0.037951488494873044, 0.03786140823364258, 0.03784188842773437, 0.03782553482055664, 0.03787980651855469, 0.037797889709472655, 0.03783168029785156, 0.037781505584716796, 0.03788390350341797, 0.037976062774658204, 0.03787059020996094, 0.037852161407470705, 0.037897216796875, 0.037761054992675784, 0.03793507385253906, 0.03784912109375, 0.037898208618164064, 0.037943294525146484, 0.03791155242919922, 0.03782553482055664, 0.03786956787109375, 0.037823486328125, 0.03791462326049805, 0.03792998504638672, 0.03796585464477539, 0.038024158477783204, 0.03801497650146484, 0.038075393676757815, 0.038626304626464845, 0.038330368041992184, 0.0381317138671875, 0.03788390350341797, 0.03796480178833008, 0.11709645080566407, 0.03769753646850586, 0.03772825622558594, 0.03762278366088867, 0.03767193603515625, 0.03777433776855469, 0.03768115234375, 0.03767603302001953, 0.03768835067749023, 0.0377374382019043, 0.03782758331298828, 0.03773952102661133, 0.0377077751159668, 0.03773132705688476, 0.03772723388671875, 0.037754878997802735, 0.037789695739746096, 0.03791360092163086, 0.03775897598266602, 0.037765121459960936, 0.03777740859985351, 0.03776204681396484, 0.0377784309387207, 0.03781836700439453, 0.03779891204833984, 0.03775692749023438, 0.03782860946655273, 0.03776102447509765, 0.03779993438720703, 0.037759998321533206, 0.037820415496826174, 0.03790028762817383, 0.037765121459960936, 0.03770675277709961, 0.03789926528930664, 0.037735424041748046, 0.03783168029785156, 0.037746688842773435, 0.0378152961730957, 0.037923839569091795, 0.03786240005493164, 0.03785932922363281, 0.03794124984741211, 0.037820415496826174, 0.03786444854736328, 0.03782656097412109, 0.03785830307006836, 0.037872638702392575, 0.037889022827148434, 0.03782758331298828, 0.037904384613037106, 0.03792486572265625, 0.03798732757568359, 0.037897216796875, 0.03784806442260742, 0.03804159927368164, 0.03798527908325195, 0.03783987045288086, 0.03789209747314453, 0.03780710220336914, 0.03793203353881836, 0.03791974258422852, 0.03798015975952149, 0.11711078643798828, 0.03786137771606445, 0.03815526580810547, 0.03790233612060547, 0.03775897598266602, 0.03766886520385742, 0.03768729782104492, 0.03782758331298828, 0.037754878997802735, 0.03770675277709961, 0.03784089660644531, 0.03780198287963867, 0.037771297454833985, 0.037704673767089844, 0.03770268630981445, 0.0377845458984375, 0.03793612670898437, 0.0377149429321289, 0.03773747253417969, 0.037713920593261716, 0.03785017776489258, 0.037717952728271484, 0.037767230987548826, 0.03779372787475586, 0.037751808166503906, 0.03773747253417969, 0.037738494873046875, 0.0378787841796875, 0.03785830307006836, 0.037784576416015625, 0.03780505752563477, 0.03810815811157227, 0.03781836700439453, 0.03775385665893555, 0.037819393157958986, 0.03779072189331055, 0.03789823913574219, 0.03786547088623047, 0.03798220825195313, 0.038063102722167966, 0.03788185501098633, 0.03780915069580078, 0.03793612670898437, 0.037889022827148434, 0.03782860946655273, 0.037787647247314454, 0.03802828979492188, 0.03801299285888672, 0.03791455841064453, 0.037996543884277346, 0.03794636917114258, 0.037872638702392575, 0.03784499359130859, 0.03789004898071289, 0.03820854568481445, 0.037983200073242185, 0.03784601593017578, 0.037838848114013675, 0.03816447830200195, 0.037812255859375, 0.03792892837524414, 0.03791462326049805, 0.03815423965454102]",tokens/s,25.622439572596,,,main,False,False,,,,,,,,,,,,,,,,,,,,,,, bfloat16-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-14B,Qwen/Qwen1.5-14B,cuda,0,42,,,True,,,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",qwen2,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,meta-llama/Llama-2-70b-hf,meta-llama/Llama-2-70b-hf,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ self.load_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 250, in load_model_with_no_weights self.pretrained_model = self.automodel_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3710, in from_pretrained model = cls(config, *model_args, **model_kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 1101, in __init__ self.model = LlamaModel(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 884, in __init__ [LlamaDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 884, in [LlamaDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 679, in __init__ self.mlp = LlamaMLP(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 192, in __init__ self.gate_proj = nn.Linear(self.hidden_size, self.intermediate_size, bias=config.mlp_bias) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/linear.py"", line 98, in __init__ self.weight = Parameter(torch.empty((out_features, in_features), **factory_kwargs)) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_device.py"", line 78, in __torch_function__ return func(*args, **kwargs) torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 448.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Deci/DeciCoder-1b,Deci/DeciCoder-1b,cuda,0,42,,,True,,,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",llama,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,meta-llama/Meta-Llama-3-70B,meta-llama/Meta-Llama-3-70B,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ self.load_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 250, in load_model_with_no_weights self.pretrained_model = self.automodel_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3710, in from_pretrained model = cls(config, *model_args, **model_kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 1101, in __init__ self.model = LlamaModel(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 884, in __init__ [LlamaDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 884, in [LlamaDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 679, in __init__ self.mlp = LlamaMLP(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 193, in __init__ self.up_proj = nn.Linear(self.hidden_size, self.intermediate_size, bias=config.mlp_bias) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/linear.py"", line 98, in __init__ self.weight = Parameter(torch.empty((out_features, in_features), **factory_kwargs)) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_device.py"", line 78, in __torch_function__ return func(*args, **kwargs) torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 448.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,r,r,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/r/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669473d3-6fa063ad2591412516235075;bbea7084-0185-418e-bf46-73a8f0e8155d) Repository Not Found for url: https://huggingface.co/r/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: r is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,togethercomputer/RedPajama-INCITE-Base-7B-v0.1,togethercomputer/RedPajama-INCITE-Base-7B-v0.1,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1291.874304,15192.293376,0.0,14545.846272,14291.49696,s,10,0.8603953018188477,0.08603953018188476,0.0014630530252219324,0.08554342651367187,0.0861418746948242,0.08828324050903319,0.08999633316040039,"[0.09042460632324219, 0.08564316558837891, 0.085666015625, 0.08545999908447266, 0.08552323150634765, 0.085552001953125, 0.08550291442871094, 0.08553485107421875, 0.08548777770996094, 0.08560073852539063]",tokens/s,2975.3765444653677,kWh,1.0130481828085492e-06,5.550388602643626e-07,5.368218919452976e-06,6.9363059625258866e-06,tokens/kWh,36907253.13777486,MB,1291.874304,15192.293376,0.0,14545.846272,14489.229824,s,10,24.025206542968753,2.402520654296875,0.004819502401922563,2.40228759765625,2.4069535400390625,2.409331750488281,2.411234318847656,"[2.4117099609375, 2.406345703125, 2.400972412109375, 2.39948828125, 2.4049345703125, 2.394389892578125, 2.403602783203125, 2.397329833984375, 2.406425048828125, 2.400008056640625]",tokens/s,26.222459268903837,kWh,2.8124991333269556e-05,1.5413555086426452e-05,0.0001209556321395457,0.00016449417855924172,tokens/kWh,382992.27700213646,,s,629,24.75707799530028,0.039359424475835125,0.010177661621910281,0.038163455963134765,0.038368665313720705,0.03851306228637695,0.12367187957763673,"[0.038007808685302735, 0.03805081558227539, 0.03806617736816406, 0.03811123275756836, 0.038215679168701173, 0.038354942321777344, 0.03809689712524414, 0.0379504623413086, 0.038145057678222655, 0.03813475036621094, 0.03829043197631836, 0.03809280014038086, 0.03844300842285156, 0.03867750549316406, 0.03863142395019531, 0.03824844741821289, 0.03810508728027344, 0.03807651138305664, 0.038158302307128904, 0.0382729606628418, 0.03817062377929688, 0.038228992462158204, 0.03818188858032227, 0.03820748901367187, 0.0381921272277832, 0.03815526580810547, 0.03822694396972656, 0.03822387313842773, 0.03818393707275391, 0.038158336639404294, 0.03824947357177735, 0.03825459289550781, 0.03820236968994141, 0.038204414367675785, 0.03820236968994141, 0.038231040954589846, 0.038258689880371094, 0.03820646286010742, 0.03890073776245117, 0.03847372817993164, 0.038307838439941407, 0.03860275268554687, 0.03842969512939453, 0.038286334991455076, 0.03836108779907227, 0.03828121566772461, 0.03846656036376953, 0.03833651351928711, 0.03823206329345703, 0.03824127960205078, 0.038258689880371094, 0.03828838348388672, 0.03826892852783203, 0.038234111785888675, 0.03905535888671875, 0.038621185302734375, 0.038284286499023434, 0.038373374938964845, 0.038267902374267575, 0.038247425079345705, 0.038319103240966795, 0.03827609634399414, 0.12385279846191406, 0.03794636917114258, 0.03783987045288086, 0.03799347305297852, 0.03803238296508789, 0.0380948486328125, 0.03801804733276367, 0.038002689361572264, 0.03772623825073242, 0.03802825546264649, 0.03808358383178711, 0.03809382247924804, 0.0380948486328125, 0.038152191162109376, 0.03816550445556641, 0.03829759979248047, 0.0381214714050293, 0.038100990295410156, 0.03812351989746094, 0.038010879516601564, 0.03821670532226563, 0.03813580703735352, 0.03812659072875976, 0.038225921630859375, 0.03810201644897461, 0.03820544052124023, 0.03813180923461914, 0.03818598556518555, 0.038125473022460936, 0.03815628814697265, 0.038188030242919925, 0.038383617401123046, 0.038591487884521485, 0.03849318313598633, 0.03830169677734375, 0.038266880035400394, 0.03826892852783203, 0.03818598556518555, 0.038209537506103515, 0.038368255615234374, 0.038188064575195316, 0.03836412811279297, 0.03817267227172851, 0.038316032409667966, 0.03820851135253906, 0.0382371826171875, 0.03825151824951172, 0.03835801696777344, 0.038231040954589846, 0.038184959411621096, 0.03823308944702149, 0.03832217788696289, 0.03825151824951172, 0.03836723327636719, 0.038353919982910156, 0.03841334533691406, 0.03823721694946289, 0.03832723236083985, 0.038299648284912106, 0.038266880035400394, 0.038217727661132815, 0.03829145431518555, 0.03826278305053711, 0.12428594970703125, 0.03802521514892578, 0.03810611343383789, 0.03805593490600586, 0.03809075164794922, 0.03810713577270508, 0.03806208038330078, 0.03822182464599609, 0.037806079864501956, 0.03784396743774414, 0.03811635208129883, 0.03814297485351562, 0.038010879516601564, 0.03810611343383789, 0.03806617736816406, 0.038234111785888675, 0.03794432067871094, 0.038128639221191404, 0.0381399040222168, 0.03818905639648437, 0.038010879516601564, 0.037743614196777346, 0.037733375549316404, 0.038174720764160154, 0.03807436752319336, 0.038145023345947264, 0.03809382247924804, 0.038100990295410156, 0.03806924819946289, 0.03808873748779297, 0.0382217903137207, 0.038217727661132815, 0.037848224639892576, 0.03791856002807617, 0.038012928009033206, 0.03800284957885742, 0.03811936187744141, 0.038110111236572264, 0.038312961578369144, 0.03821875381469726, 0.03813683319091797, 0.03807231903076172, 0.038414337158203124, 0.037986305236816405, 0.03819724655151367, 0.03821977615356445, 0.038408191680908206, 0.038312961578369144, 0.03791462326049805, 0.038188030242919925, 0.038292510986328125, 0.03823715209960937, 0.038353919982910156, 0.03829043197631836, 0.038351905822753905, 0.0383927993774414, 0.03815423965454102, 0.037997566223144534, 0.037997566223144534, 0.037966846466064456, 0.03801190567016602, 0.037953536987304685, 0.037961727142333986, 0.12367359924316407, 0.03768217468261719, 0.03820748901367187, 0.03838873672485352, 0.03789209747314453, 0.03768320083618164, 0.03803647994995117, 0.03817574310302734, 0.037967872619628903, 0.03774259185791016, 0.03800371170043945, 0.0381102066040039, 0.038056961059570314, 0.03808358383178711, 0.037969921112060545, 0.037766143798828124, 0.03802214431762695, 0.03808870315551758, 0.03801606369018555, 0.03819513702392578, 0.03818700790405274, 0.038147071838378906, 0.03809996795654297, 0.037989376068115234, 0.03777228927612305, 0.0380579833984375, 0.03815731048583984, 0.03779481506347656, 0.03775385665893555, 0.03816243362426758, 0.03817062377929688, 0.03784806442260742, 0.03788185501098633, 0.03791155242919922, 0.03806924819946289, 0.03804467010498047, 0.03818188858032227, 0.03807743835449219, 0.03832012939453125, 0.03836415863037109, 0.03824332809448242, 0.038188030242919925, 0.038312961578369144, 0.03811328125, 0.03812467193603516, 0.03810496139526367, 0.03832831954956055, 0.038414337158203124, 0.038209537506103515, 0.03824947357177735, 0.03825971221923828, 0.038370304107666016, 0.03820851135253906, 0.03787571334838867, 0.037852161407470705, 0.03829350280761719, 0.03825254440307617, 0.03797401428222656, 0.03808870315551758, 0.03802828979492188, 0.03826483154296875, 0.0382341423034668, 0.03823715209960937, 0.12371968078613281, 0.03760332870483398, 0.03771084976196289, 0.037958656311035156, 0.03784601593017578, 0.037923839569091795, 0.03802521514892578, 0.038351871490478515, 0.038128639221191404, 0.03782656097412109, 0.03774054336547852, 0.037754878997802735, 0.03830886459350586, 0.03855462265014648, 0.038520832061767575, 0.038319103240966795, 0.03827302551269531, 0.03815321731567383, 0.03813273620605469, 0.03809075164794922, 0.03806924819946289, 0.03810815811157227, 0.03804876708984375, 0.038166526794433595, 0.038201343536376955, 0.038176769256591796, 0.038196224212646485, 0.03820544052124023, 0.03828121566772461, 0.038161407470703124, 0.038231040954589846, 0.03827507019042969, 0.03832729721069336, 0.038133758544921875, 0.038198272705078126, 0.03833958435058594, 0.03840409469604492, 0.03819007873535156, 0.03818188858032227, 0.03823728179931641, 0.038115230560302735, 0.03857920074462891, 0.03808665466308594, 0.03810406494140625, 0.03814297485351562, 0.038163455963134765, 0.03810406494140625, 0.03822390365600586, 0.03835696029663086, 0.03813580703735352, 0.03819120025634765, 0.038073249816894535, 0.03812761688232422, 0.03813580703735352, 0.038255615234375, 0.03832012939453125, 0.03836723327636719, 0.0384266242980957, 0.038354049682617186, 0.038201217651367185, 0.03825356674194336, 0.03838873672485352, 0.03826176071166992, 0.12384563446044922, 0.03769753646850586, 0.0377077751159668, 0.03770675277709961, 0.03766988754272461, 0.037743614196777346, 0.037771263122558595, 0.0378152961730957, 0.037806079864501956, 0.03771084976196289, 0.037781505584716796, 0.037716991424560545, 0.037779457092285154, 0.037735424041748046, 0.038035457611083984, 0.037872638702392575, 0.03775078582763672, 0.0377446403503418, 0.037766143798828124, 0.03778355026245117, 0.03781119918823242, 0.03782144165039063, 0.03779593658447265, 0.037872543334960936, 0.03781017684936523, 0.037804031372070314, 0.0379504623413086, 0.03786137771606445, 0.037806079864501956, 0.03806719970703125, 0.0379607048034668, 0.03789311981201172, 0.03790233612060547, 0.03789311981201172, 0.038796287536621094, 0.0385904655456543, 0.03786956787109375, 0.03787161636352539, 0.038298686981201174, 0.038311870574951175, 0.03784601593017578, 0.03782758331298828, 0.0381952018737793, 0.03825356674194336, 0.038241310119628905, 0.038227935791015626, 0.03815526580810547, 0.03832831954956055, 0.03831193542480469, 0.03818086242675781, 0.03806208038330078, 0.038307838439941407, 0.03832831954956055, 0.038284286499023434, 0.03825158309936524, 0.03834259033203125, 0.03831615829467774, 0.03820649719238281, 0.038230880737304684, 0.038338558197021484, 0.038267902374267575, 0.03824639892578125, 0.037833728790283204, 0.12362342071533203, 0.03758899307250976, 0.037658687591552734, 0.03797292709350586, 0.037956607818603515, 0.03805094528198242, 0.03803638458251953, 0.03824841690063477, 0.03817782211303711, 0.03806409454345703, 0.03774054336547852, 0.038024192810058595, 0.03819724655151367, 0.038191104888916014, 0.03809689712524414, 0.03823923110961914, 0.038079486846923825, 0.03828531265258789, 0.0381952018737793, 0.038166526794433595, 0.038397953033447264, 0.03823308944702149, 0.03816960144042969, 0.03821363067626953, 0.038250495910644534, 0.0381214714050293, 0.03815628814697265, 0.03815731048583984, 0.03859558486938477, 0.03818393707275391, 0.03780915069580078, 0.03784908676147461, 0.037795841217041014, 0.038212608337402344, 0.038368255615234374, 0.0381399040222168, 0.03785625457763672, 0.03803033447265625, 0.03820032119750977, 0.03835289764404297, 0.03822182464599609, 0.03823820877075195, 0.038312961578369144, 0.03833139038085937, 0.038302719116210936, 0.03785830307006836, 0.038201343536376955, 0.03808768081665039, 0.03809075164794922, 0.03788390350341797, 0.03812659072875976, 0.03828121566772461, 0.03820748901367187, 0.03787059020996094, 0.038319103240966795, 0.03850140762329102, 0.038340576171875, 0.03829350280761719, 0.038346752166748044, 0.03834265518188477, 0.03848908615112305, 0.03836524963378906, 0.03841939163208008, 0.12376268768310547, 0.037663745880126956, 0.03785830307006836, 0.038547454833984376, 0.039019519805908204, 0.038300670623779294, 0.03813888168334961, 0.038373374938964845, 0.03817574310302734, 0.038130687713623046, 0.03817881774902344, 0.03815935897827148, 0.0381921272277832, 0.03814300918579101, 0.038222816467285155, 0.03827711868286133, 0.03818188858032227, 0.03818905639648437, 0.03775078582763672, 0.037703678131103514, 0.03775385665893555, 0.03778355026245117, 0.03784703826904297, 0.03798425674438476, 0.03811532974243164, 0.03773952102661133, 0.03780915069580078, 0.037850112915039064, 0.03790233612060547, 0.03778867340087891, 0.03781836700439453, 0.03812249755859375, 0.03786956787109375, 0.03793407821655274, 0.03789619064331055, 0.03779993438720703, 0.03783782577514649, 0.03816348648071289, 0.038274017333984375, 0.03829452896118164, 0.03867443084716797, 0.037996543884277346, 0.03786857604980469, 0.037862369537353516, 0.03784396743774414, 0.037910526275634765, 0.03832831954956055, 0.03844403076171875, 0.038340606689453126, 0.038432769775390625, 0.03823513412475586, 0.03783679962158203, 0.03794124984741211, 0.03788595199584961, 0.037897216796875, 0.037972991943359374, 0.03790950393676758, 0.03794636917114258, 0.03800985717773438, 0.037969921112060545, 0.038084606170654296, 0.037989376068115234, 0.037981182098388674, 0.12366745758056641, 0.037678081512451174, 0.0380313606262207, 0.03768832015991211, 0.037664768218994144, 0.037705726623535156, 0.037651454925537106, 0.037765121459960936, 0.03823308944702149, 0.0381214714050293, 0.03824127960205078, 0.03817475128173828, 0.03812656021118164, 0.03820342254638672, 0.03817059326171875, 0.03828531265258789, 0.03815628814697265, 0.03828121566772461, 0.038225921630859375, 0.03824127960205078, 0.03823923110961914, 0.03807955169677734, 0.03805587387084961, 0.03824947357177735, 0.03822086334228516, 0.03822995376586914, 0.038300670623779294, 0.038225921630859375, 0.038604801177978515, 0.038209537506103515, 0.038098976135253905, 0.03794633483886719, 0.037822463989257815, 0.03784806442260742, 0.03786956787109375, 0.03780710220336914, 0.03792588806152344, 0.03784703826904297, 0.0383375358581543, 0.03831808090209961, 0.03824435043334961, 0.038247425079345705, 0.03815628814697265, 0.0392806396484375, 0.03870105743408203, 0.03885465621948242, 0.038392833709716793, 0.03845027160644531, 0.03832617568969727, 0.03838566589355469, 0.03838771057128906, 0.03839084625244141, 0.03826483154296875, 0.038224830627441406, 0.03837952041625976, 0.03798527908325195, 0.03789318466186523, 0.03797401428222656, 0.03934201431274414, 0.03860275268554687, 0.03815116882324219, 0.038354942321777344, 0.03834982299804687, 0.12387635040283203, 0.037648384094238284, 0.03801599884033203, 0.038056961059570314, 0.038084606170654296, 0.03818598556518555, 0.038201343536376955, 0.03822387313842773, 0.03820646286010742, 0.03803750228881836, 0.03776409530639648, 0.038176769256591796, 0.038141952514648435, 0.03812249755859375, 0.038112255096435545, 0.038356990814208985, 0.0381214714050293, 0.037655616760253904, 0.03771692657470703, 0.037814273834228515, 0.038089729309082034, 0.03812659072875976, 0.038125568389892575, 0.03818812942504883, 0.03820022583007812, 0.038166526794433595, 0.03816550445556641, 0.03812966537475586, 0.038217727661132815, 0.03809382247924804, 0.0381728630065918, 0.03812025451660156, 0.03832115173339844, 0.03828543853759766, 0.03833433532714844, 0.03867443084716797, 0.038456382751464846, 0.038072254180908205, 0.03804876708984375, 0.038317054748535154, 0.038351871490478515, 0.03793407821655274, 0.038225921630859375, 0.037838848114013675, 0.03786137771606445, 0.03783679962158203, 0.03786956787109375, 0.03790643310546875, 0.0379607048034668, 0.03785932922363281, 0.03829145431518555, 0.03836723327636719, 0.03836108779907227, 0.03829145431518555, 0.038300670623779294, 0.03812659072875976, 0.0379422721862793, 0.03793100738525391, 0.037906494140625, 0.03795859146118164, 0.03795455932617187, 0.03785420989990235, 0.03794636917114258]",tokens/s,25.40687556582424,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-72B,Qwen/Qwen1.5-72B,cuda,0,42,,,True,,,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",qwen2,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,internlm/internlm-20b,internlm/internlm-20b,cuda,0,42,,,True,,,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",internlm,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-110B,Qwen/Qwen1.5-110B,cuda,0,42,,,True,,,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",qwen2,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Salesforce/codegen-16B-nl,Salesforce/codegen-16B-nl,cuda,0,42,,,True,,,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",codegen,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,m,m,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/m/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669472d3-0e58b1f67aa709c95a8d7426;c59bea95-5ca1-4c17-81b1-955f74ee5309) Repository Not Found for url: https://huggingface.co/m/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: m is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,openai-community/gpt2-large,openai-community/gpt2-large,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1218.674688,2661.810176,0.0,2015.363072,1692.280832,s,10,0.24226054382324216,0.02422605438232422,0.0007668612376014522,0.023795424461364748,0.02500129566192627,0.025577368068695067,0.02603822599411011,"[0.026153440475463866, 0.02380307197570801, 0.02470319938659668, 0.02365532875061035, 0.023776927947998048, 0.023787776947021486, 0.02371471977233887, 0.023629823684692384, 0.024162975311279297, 0.024873279571533204]",tokens/s,10567.135529373794,kWh,2.8625296943444927e-07,1.5685372493336497e-07,1.0371161108329199e-06,1.4802228052007342e-06,tokens/kWh,172946936.83987907,MB,1218.674688,2663.907328,0.0,2017.460224,1740.086272,s,10,14.79459228515625,1.479459228515625,0.028526033708113827,1.4860075073242187,1.5069755615234375,1.5124290649414063,1.5167918676757812,"[1.482174072265625, 1.4794532470703126, 1.4992864990234376, 1.5020098876953125, 1.4384600830078125, 1.4480838623046874, 1.4898409423828125, 1.431637451171875, 1.505763671875, 1.517882568359375]",tokens/s,42.58312685183581,kWh,1.7695988525353617e-05,9.697412545461509e-06,3.562160766056736e-05,6.301500873138247e-05,tokens/kWh,999761.8229103729,,s,629,14.985263092041022,0.023823947682100183,0.0029445657243882214,0.023669759750366212,0.024091238021850588,0.024281702423095703,0.04714971206665039,"[0.02428620719909668, 0.023126016616821288, 0.023560192108154295, 0.023589887619018556, 0.02364825630187988, 0.023564287185668945, 0.023571456909179687, 0.02355513572692871, 0.02366048049926758, 0.023533567428588868, 0.023710720062255858, 0.02368511962890625, 0.023749696731567384, 0.023671743392944335, 0.023723007202148438, 0.02376300811767578, 0.0235662727355957, 0.023579647064208984, 0.023619583129882812, 0.023650304794311523, 0.02369740867614746, 0.02368000030517578, 0.024089599609375, 0.023996416091918944, 0.02431590461730957, 0.02368819236755371, 0.02284339141845703, 0.024205312728881836, 0.023999488830566407, 0.023608320236206053, 0.023512096405029298, 0.023624671936035156, 0.02369228744506836, 0.02323865509033203, 0.022634496688842775, 0.022583295822143554, 0.022543359756469726, 0.022750207901000977, 0.023391231536865235, 0.023592960357666014, 0.02346905517578125, 0.023654399871826173, 0.02351513671875, 0.024465408325195313, 0.02370150375366211, 0.024129535675048826, 0.023513088226318358, 0.023606271743774415, 0.022945791244506835, 0.023600128173828124, 0.023610368728637695, 0.023597055435180665, 0.023686143875122072, 0.023464960098266603, 0.02284441566467285, 0.022684671401977538, 0.023615488052368162, 0.02332979202270508, 0.022632448196411133, 0.022579200744628908, 0.022609920501708985, 0.022691839218139647, 0.04726476669311523, 0.02269491195678711, 0.022594560623168947, 0.022540351867675782, 0.02254841613769531, 0.02252288055419922, 0.022560768127441407, 0.022773759841918945, 0.02281062316894531, 0.02281881523132324, 0.023004159927368165, 0.022749183654785156, 0.022657024383544923, 0.02266316795349121, 0.022773759841918945, 0.02272870445251465, 0.02288435173034668, 0.022806528091430665, 0.022947839736938477, 0.022975488662719725, 0.022862848281860353, 0.022770687103271483, 0.022863872528076173, 0.023992319107055664, 0.02392985534667969, 0.023945215225219727, 0.024009727478027345, 0.023933952331542968, 0.023946239471435548, 0.02406399917602539, 0.024373247146606446, 0.024026111602783205, 0.0239052791595459, 0.023938047409057618, 0.024073215484619142, 0.024010751724243166, 0.024007680892944337, 0.02389606475830078, 0.023938047409057618, 0.023793664932250977, 0.02395238494873047, 0.023833599090576172, 0.023883775711059572, 0.023975936889648438, 0.023984128952026368, 0.023641088485717773, 0.023617536544799804, 0.023614463806152345, 0.023541759490966797, 0.023563264846801758, 0.02347315216064453, 0.02351923179626465, 0.023472127914428712, 0.023516159057617187, 0.02416640090942383, 0.024003616333007814, 0.02398512077331543, 0.023880704879760743, 0.023941120147705077, 0.02405990409851074, 0.024001535415649415, 0.02405580711364746, 0.024009727478027345, 0.049239040374755856, 0.02392064094543457, 0.022981632232666017, 0.02411008071899414, 0.02453708839416504, 0.024029184341430664, 0.02409779167175293, 0.02452070426940918, 0.024245248794555665, 0.024182783126831055, 0.02407219123840332, 0.02405887985229492, 0.023964672088623046, 0.024006656646728516, 0.02405171203613281, 0.0241889591217041, 0.024208351135253905, 0.023986175537109376, 0.02391961669921875, 0.023989248275756835, 0.023997440338134765, 0.023995391845703123, 0.023883775711059572, 0.023928831100463867, 0.023764991760253908, 0.02353459167480469, 0.023773183822631837, 0.023759872436523437, 0.023710720062255858, 0.02365542411804199, 0.0236810245513916, 0.02351923179626465, 0.023811071395874024, 0.023573503494262696, 0.023723007202148438, 0.023335935592651368, 0.023599103927612306, 0.02384998321533203, 0.023634944915771484, 0.02408038330078125, 0.023578624725341796, 0.023569408416748046, 0.023555072784423828, 0.023620672225952148, 0.023699392318725587, 0.023616512298583983, 0.0232857608795166, 0.023440383911132814, 0.02345062446594238, 0.023232511520385742, 0.023198720932006835, 0.023119871139526366, 0.02315673637390137, 0.02351923179626465, 0.02366873550415039, 0.023600128173828124, 0.023827455520629884, 0.023864320755004883, 0.02411212730407715, 0.023957504272460937, 0.023976959228515626, 0.023883775711059572, 0.023962623596191408, 0.047072254180908206, 0.022804479598999023, 0.022812671661376953, 0.022718463897705078, 0.022837247848510742, 0.022801408767700194, 0.023014432907104494, 0.02356425666809082, 0.02433433532714844, 0.024284160614013672, 0.02408550453186035, 0.024020992279052734, 0.024071168899536134, 0.023846912384033202, 0.024238079071044923, 0.024186880111694335, 0.024066047668457033, 0.024406015396118166, 0.023815168380737304, 0.023965696334838867, 0.023829504013061522, 0.023900159835815428, 0.024016895294189454, 0.02390630340576172, 0.0240599365234375, 0.023935968399047852, 0.024006656646728516, 0.02407423973083496, 0.024123392105102538, 0.02393600082397461, 0.02391347122192383, 0.02389299201965332, 0.023871488571166992, 0.02391756820678711, 0.0239052791595459, 0.023814144134521483, 0.02388275146484375, 0.023953407287597657, 0.023920703887939453, 0.023883712768554687, 0.023794687271118165, 0.0239052791595459, 0.02390937614440918, 0.02396883201599121, 0.02417862319946289, 0.023985151290893555, 0.02389094352722168, 0.023972864151000976, 0.023823360443115234, 0.023984128952026368, 0.023910400390625, 0.024034303665161134, 0.023645183563232423, 0.023545856475830077, 0.023669759750366212, 0.023773183822631837, 0.02392166328430176, 0.023966720581054687, 0.023983104705810547, 0.024379392623901368, 0.02393087959289551, 0.023982080459594726, 0.02384486389160156, 0.047249408721923826, 0.02275225639343262, 0.022947839736938477, 0.022789119720458984, 0.022791168212890626, 0.022795263290405272, 0.022581247329711913, 0.02262835121154785, 0.02267852783203125, 0.022626304626464845, 0.02258028793334961, 0.02260678482055664, 0.02264575958251953, 0.022642688751220705, 0.02308095932006836, 0.022971391677856445, 0.02282598304748535, 0.022541311264038084, 0.02265497589111328, 0.022761472702026365, 0.022887424468994142, 0.022861824035644532, 0.022786048889160155, 0.02287820816040039, 0.02285260772705078, 0.022853631973266602, 0.0228853759765625, 0.02284441566467285, 0.023011327743530274, 0.022889471054077147, 0.022816768646240236, 0.02286079978942871, 0.022789119720458984, 0.02282803153991699, 0.022803455352783202, 0.022845439910888672, 0.022871040344238282, 0.022820863723754883, 0.022770687103271483, 0.023230464935302734, 0.022722560882568358, 0.022861824035644532, 0.023029760360717775, 0.02293350410461426, 0.022920192718505858, 0.022814720153808594, 0.022837247848510742, 0.02282803153991699, 0.02290995216369629, 0.022832128524780275, 0.022966272354125978, 0.02285772705078125, 0.022751232147216797, 0.022759424209594727, 0.022769664764404295, 0.023022592544555662, 0.022914047241210937, 0.022897663116455077, 0.02272051239013672, 0.022816768646240236, 0.022876192092895507, 0.022813663482666016, 0.02290995216369629, 0.04728934478759766, 0.02287513542175293, 0.02289971160888672, 0.022951936721801756, 0.022730752944946288, 0.022768640518188478, 0.02280243110656738, 0.022790143966674805, 0.022832128524780275, 0.02285875129699707, 0.022898687362670898, 0.022887424468994142, 0.0228351993560791, 0.02287615966796875, 0.02290483283996582, 0.02291814422607422, 0.022806528091430665, 0.022743040084838868, 0.023193599700927735, 0.022748159408569335, 0.022795263290405272, 0.022786048889160155, 0.02241433525085449, 0.022548511505126954, 0.022550495147705078, 0.022550527572631835, 0.0229171199798584, 0.02285055923461914, 0.022737920761108397, 0.022939647674560547, 0.02304921531677246, 0.02289356803894043, 0.02290995216369629, 0.022874111175537108, 0.022772735595703125, 0.02287718391418457, 0.02267955207824707, 0.022864959716796876, 0.022771648406982422, 0.02275328063964844, 0.022758399963378906, 0.02291916847229004, 0.022840320587158205, 0.02304921531677246, 0.023395328521728515, 0.023155712127685548, 0.02281881523132324, 0.022793216705322264, 0.022583295822143554, 0.022967296600341795, 0.022743072509765625, 0.022556640625, 0.022524927139282228, 0.022777856826782225, 0.02283622360229492, 0.02364825630187988, 0.02385305595397949, 0.024081407546997072, 0.02400160026550293, 0.02400761604309082, 0.024060928344726562, 0.02394419288635254, 0.024417280197143554, 0.04715315246582031, 0.022803455352783202, 0.022922239303588866, 0.02287820816040039, 0.022845439910888672, 0.02280243110656738, 0.02289151954650879, 0.022757375717163086, 0.02294169616699219, 0.02281881523132324, 0.023001087188720702, 0.022932479858398438, 0.022840320587158205, 0.022821887969970703, 0.02290483283996582, 0.023409664154052736, 0.023778303146362305, 0.023573503494262696, 0.023511072158813477, 0.023860191345214842, 0.02391347122192383, 0.024002559661865236, 0.023932928085327147, 0.023971839904785155, 0.024000511169433594, 0.023965696334838867, 0.02345881652832031, 0.023567359924316408, 0.023825408935546875, 0.024011775970458983, 0.02387455940246582, 0.023831552505493164, 0.023956480026245116, 0.02405990409851074, 0.02391142463684082, 0.023945215225219727, 0.023948287963867186, 0.023996416091918944, 0.024022016525268555, 0.02395955276489258, 0.02432307243347168, 0.024060928344726562, 0.023792640686035156, 0.023632896423339843, 0.02391551971435547, 0.02388275146484375, 0.024005632400512695, 0.023933952331542968, 0.023931903839111326, 0.02393087959289551, 0.023931903839111326, 0.0239052791595459, 0.023872512817382813, 0.024029184341430664, 0.023967744827270508, 0.023925760269165038, 0.023924736022949217, 0.023864320755004883, 0.023860223770141603, 0.023877664566040038, 0.023890911102294923, 0.023603200912475586, 0.023904287338256835, 0.04714086532592773, 0.02276348876953125, 0.02264575958251953, 0.02263859176635742, 0.022771711349487304, 0.022618112564086915, 0.022548479080200197, 0.02269491195678711, 0.02284236717224121, 0.022782976150512696, 0.022923263549804687, 0.022899744033813476, 0.022909919738769532, 0.02330316734313965, 0.023266304016113282, 0.023214080810546874, 0.023001087188720702, 0.02285875129699707, 0.022963199615478515, 0.023052288055419923, 0.0225218563079834, 0.022631423950195313, 0.02251263999938965, 0.022459392547607423, 0.023237632751464843, 0.022738943099975584, 0.022529024124145508, 0.022443071365356445, 0.022619071960449218, 0.022542335510253905, 0.02270412826538086, 0.02268569564819336, 0.022718624114990236, 0.022625120162963867, 0.022765567779541016, 0.022776832580566408, 0.022616064071655274, 0.022525951385498046, 0.0224716796875, 0.02253926467895508, 0.02249113655090332, 0.022435840606689454, 0.022747135162353514, 0.022683647155761717, 0.022503423690795898, 0.022545408248901368, 0.02254745674133301, 0.022597631454467772, 0.02249113655090332, 0.022245376586914063, 0.022418432235717774, 0.02249113655090332, 0.02249625587463379, 0.02262937545776367, 0.02312499237060547, 0.02283622360229492, 0.02271232032775879, 0.022800384521484376, 0.022788095474243163, 0.022788095474243163, 0.022803455352783202, 0.02286591911315918, 0.022780927658081054, 0.047631359100341795, 0.023003135681152344, 0.023012351989746094, 0.022815744400024415, 0.022801408767700194, 0.022931455612182617, 0.02284339141845703, 0.02333798408508301, 0.023942144393920898, 0.02388275146484375, 0.023851007461547852, 0.023993343353271485, 0.02390732765197754, 0.023995391845703123, 0.023965696334838867, 0.024039424896240235, 0.023875583648681642, 0.024041471481323243, 0.023988224029541014, 0.023855104446411132, 0.02391551971435547, 0.023957504272460937, 0.023916543960571288, 0.023960575103759766, 0.023867391586303712, 0.023964672088623046, 0.02388991928100586, 0.023950336456298828, 0.024010751724243166, 0.023928831100463867, 0.024135679244995118, 0.023931903839111326, 0.023948287963867186, 0.024037376403808593, 0.0245483512878418, 0.024207359313964845, 0.024009727478027345, 0.023872512817382813, 0.023964672088623046, 0.023972864151000976, 0.023954431533813478, 0.02426470375061035, 0.024008703231811524, 0.024007680892944337, 0.024002559661865236, 0.023856128692626953, 0.023941120147705077, 0.02488217544555664, 0.024633344650268556, 0.024071199417114258, 0.023988191604614257, 0.024174591064453126, 0.025488384246826173, 0.024361984252929687, 0.023841791152954102, 0.023840768814086914, 0.023776256561279296, 0.023584768295288085, 0.02374143981933594, 0.023908351898193358, 0.02391142463684082, 0.024024063110351563, 0.023986207962036134, 0.04968243026733398, 0.024002527236938475, 0.024037376403808593, 0.024017919540405275, 0.02390220832824707, 0.023868576049804687, 0.023894880294799806, 0.023777280807495117, 0.023635967254638672, 0.02395136070251465, 0.02396883201599121, 0.02401683235168457, 0.023972864151000976, 0.02405580711364746, 0.024185855865478514, 0.024343551635742186, 0.024139776229858398, 0.024041471481323243, 0.024465408325195313, 0.024027135848999022, 0.024030208587646484, 0.023967744827270508, 0.023974912643432617, 0.023975936889648438, 0.023991296768188477, 0.02427289581298828, 0.02405887985229492, 0.024197120666503907, 0.023983104705810547, 0.024053760528564453, 0.023946239471435548, 0.023837696075439452, 0.024028160095214843, 0.024017919540405275, 0.024010751724243166, 0.024023040771484375, 0.024215551376342775, 0.02390118408203125, 0.023989248275756835, 0.023954431533813478, 0.024006656646728516, 0.023832576751708984, 0.023997440338134765, 0.024114175796508788, 0.02427801513671875, 0.02414080047607422, 0.024174591064453126, 0.02392268753051758, 0.024156160354614258, 0.02513817596435547, 0.024972288131713868, 0.024617984771728517, 0.02414080047607422, 0.024203264236450195, 0.024044544219970702, 0.024010751724243166, 0.023996416091918944, 0.0241213436126709, 0.024030208587646484, 0.02394419288635254, 0.024028160095214843, 0.024507392883300783, 0.023973888397216796]",tokens/s,41.97457169331081,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,huggyllama/llama-65b,huggyllama/llama-65b,cuda,0,42,,,True,,,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",llama,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,l,l,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/l/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66947448-6846d2463d5b138d5d38a1c4;e8ae6f08-ac6a-41e6-bdbb-f4b78086be8e) Repository Not Found for url: https://huggingface.co/l/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: l is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,B,B,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/B/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669324fb-7dbe01720b640ff75edb6905;f1054c35-eedf-4ab4-8721-7342c9e50de7) Repository Not Found for url: https://huggingface.co/B/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: B is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,huggyllama/llama-30b,huggyllama/llama-30b,cuda,0,42,,,True,,,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",llama,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,meta-llama/Llama-2-13b-hf,meta-llama/Llama-2-13b-hf,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ self.load_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 250, in load_model_with_no_weights self.pretrained_model = self.automodel_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3710, in from_pretrained model = cls(config, *model_args, **model_kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 1101, in __init__ self.model = LlamaModel(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 884, in __init__ [LlamaDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 884, in [LlamaDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 679, in __init__ self.mlp = LlamaMLP(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 192, in __init__ self.gate_proj = nn.Linear(self.hidden_size, self.intermediate_size, bias=config.mlp_bias) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/linear.py"", line 98, in __init__ self.weight = Parameter(torch.empty((out_features, in_features), **factory_kwargs)) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_device.py"", line 78, in __torch_function__ return func(*args, **kwargs) torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 136.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-7B,,cuda,0,42,,,,,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,217063f5c507ed7cc255df7e1f64c4333a0b4dfe,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,,MB,1303.093248,17102.798848,0.0,16456.351744,16359.853056,s,10,1.0194314193725587,0.10194314193725587,0.0002529759528872485,0.10188860702514649,0.10205652389526366,0.10236344413757324,0.1026089803314209,"[0.1026703643798828, 0.10180748748779297, 0.10198831939697266, 0.10172902679443359, 0.10191126251220703, 0.10188739013671876, 0.1017694091796875, 0.10191958618164063, 0.10188982391357422, 0.10185874938964844]",tokens/s,2511.203746864731,kWh,1.2064162498158392e-06,6.610585252520669e-07,5.900701999469618e-06,7.768176774537524e-06,tokens/kWh,32954965.80859424,MB,1303.420928,17102.798848,0.0,16456.351744,16416.242176,s,10,23.7518544921875,2.3751854492187503,0.0033105326535980578,2.3739964599609373,2.3780450195312497,2.38121611328125,2.38375298828125,"[2.38438720703125, 2.373756591796875, 2.37312353515625, 2.374914306640625, 2.37734033203125, 2.373713134765625, 2.3730146484375, 2.374236328125, 2.374585693359375, 2.37278271484375]",tokens/s,26.524244673493623,kWh,2.7978138962840816e-05,1.5333065776871895e-05,0.00011402269949472894,0.00015733390423444167,tokens/kWh,400422.27583778976,,s,629,24.629738498687743,0.03915697694544951,0.012174665303252123,0.03768115234375,0.03781857223510742,0.03790479431152344,0.14013506042480467,"[0.038335487365722655, 0.03846758270263672, 0.03829350280761719, 0.038324222564697266, 0.038228992462158204, 0.03825766372680664, 0.038346752166748044, 0.0383559684753418, 0.0382105598449707, 0.03811123275756836, 0.03820032119750977, 0.038307838439941407, 0.03793817520141601, 0.03768729782104492, 0.03760748672485351, 0.037632961273193356, 0.0377149429321289, 0.037819393157958986, 0.037556224822998044, 0.03764223861694336, 0.037634048461914066, 0.03768220901489258, 0.037555168151855466, 0.03766886520385742, 0.03762278366088867, 0.03767910385131836, 0.037694465637207034, 0.037678081512451174, 0.0376258544921875, 0.03778252792358398, 0.037634048461914066, 0.03767193603515625, 0.037579776763916016, 0.037705726623535156, 0.03761459350585938, 0.037787647247314454, 0.03773952102661133, 0.03771907043457031, 0.037696479797363285, 0.03768627166748047, 0.037663745880126956, 0.03773747253417969, 0.037759998321533206, 0.037766143798828124, 0.03769651031494141, 0.0377784309387207, 0.03773440170288086, 0.037765121459960936, 0.03769036865234375, 0.037776382446289065, 0.037722110748291016, 0.037771263122558595, 0.03773235321044922, 0.0378081283569336, 0.037767230987548826, 0.037794750213623045, 0.03775385665893555, 0.03780819320678711, 0.037739456176757814, 0.037852161407470705, 0.037789695739746096, 0.03779993438720703, 0.1402255401611328, 0.03753676986694336, 0.037548030853271484, 0.03746201705932617, 0.03756032180786133, 0.037498878479003905, 0.03753779220581055, 0.03756748962402344, 0.037614654541015625, 0.037557182312011717, 0.03754291152954101, 0.03756032180786133, 0.037612545013427735, 0.03757875061035156, 0.03768524932861328, 0.03760844802856445, 0.03763507080078125, 0.03763097763061524, 0.037610496520996094, 0.037563392639160156, 0.0376545295715332, 0.03755110549926758, 0.03762995147705078, 0.03764121627807617, 0.03769036865234375, 0.037663745880126956, 0.03767091369628906, 0.0376360969543457, 0.0377077751159668, 0.03763507080078125, 0.037631999969482424, 0.03768729782104492, 0.03768320083618164, 0.03760435104370117, 0.03767609786987305, 0.03761452865600586, 0.03770982360839844, 0.03763206481933594, 0.03764012908935547, 0.03773747253417969, 0.03769343948364258, 0.03768217468261719, 0.03765862274169922, 0.03772108840942383, 0.03775385665893555, 0.037661697387695314, 0.03785318374633789, 0.037705726623535156, 0.037787647247314454, 0.03769343948364258, 0.03775078582763672, 0.03791155242919922, 0.03785318374633789, 0.03769753646850586, 0.03773235321044922, 0.0377784309387207, 0.03779072189331055, 0.03773952102661133, 0.037789695739746096, 0.03771289443969727, 0.03783679962158203, 0.037773311614990236, 0.037784576416015625, 0.1403187255859375, 0.037648384094238284, 0.037585918426513674, 0.03748044967651367, 0.03755929565429687, 0.03746406555175781, 0.03754905700683594, 0.03770470428466797, 0.037564449310302735, 0.03755414581298828, 0.03756851196289063, 0.037561344146728515, 0.03768217468261719, 0.037553150177001955, 0.03762688064575195, 0.0375551986694336, 0.03757567977905273, 0.03752959823608398, 0.03759513473510742, 0.037528575897216795, 0.03759513473510742, 0.037532737731933594, 0.03764729690551758, 0.0376545295715332, 0.03761459350585938, 0.03759206390380859, 0.03766579055786133, 0.037577728271484374, 0.03764223861694336, 0.037634048461914066, 0.03769036865234375, 0.03760639953613281, 0.03764025497436523, 0.03760838317871094, 0.03769036865234375, 0.037639232635498045, 0.037698497772216795, 0.037645313262939455, 0.03769343948364258, 0.03773747253417969, 0.03770880126953125, 0.03768012619018555, 0.03773952102661133, 0.037610496520996094, 0.037720062255859374, 0.03766681671142578, 0.037817344665527344, 0.037743614196777346, 0.03775897598266602, 0.03767398452758789, 0.037771263122558595, 0.03769452667236328, 0.03772927856445313, 0.03769747161865234, 0.038121536254882814, 0.03773843383789063, 0.037748737335205076, 0.03772927856445313, 0.037822463989257815, 0.03765760040283203, 0.037789695739746096, 0.03767705535888672, 0.03782860946655273, 0.14010780334472656, 0.037492702484130856, 0.03754086303710937, 0.037456897735595705, 0.03749683380126953, 0.03748863983154297, 0.03758803176879883, 0.0377138557434082, 0.03757567977905273, 0.03749785614013672, 0.03758694458007812, 0.0375203857421875, 0.03761151885986328, 0.03754598236083984, 0.0375551986694336, 0.037670944213867186, 0.03758281707763672, 0.037466110229492186, 0.03760537719726562, 0.03752959823608398, 0.037765121459960936, 0.038100990295410156, 0.03772825622558594, 0.03760639953613281, 0.03764019012451172, 0.03767603302001953, 0.03773952102661133, 0.037694465637207034, 0.0377149429321289, 0.03757056045532227, 0.03771187210083008, 0.03764019012451172, 0.03767091369628906, 0.037582847595214845, 0.03769651031494141, 0.037610496520996094, 0.03769139099121094, 0.037749759674072264, 0.03801599884033203, 0.03783679962158203, 0.03787673568725586, 0.037751808166503906, 0.03780915069580078, 0.03771289443969727, 0.03777740859985351, 0.037639198303222654, 0.03775484848022461, 0.03770473480224609, 0.03778355026245117, 0.03766268920898438, 0.03778252792358398, 0.03770470428466797, 0.03780198287963867, 0.037852161407470705, 0.037751808166503906, 0.037735424041748046, 0.03782860946655273, 0.03772723388671875, 0.0378081283569336, 0.03774259185791016, 0.0378152961730957, 0.037731361389160153, 0.03784086227416992, 0.14026956176757813, 0.03751020812988281, 0.037644222259521486, 0.03782451248168945, 0.03781017684936523, 0.0376514892578125, 0.040468448638916014, 0.03775795364379883, 0.03759513473510742, 0.037544960021972655, 0.037579872131347655, 0.03750595092773437, 0.037634048461914066, 0.037574657440185545, 0.037594112396240234, 0.03759308624267578, 0.03759513473510742, 0.03757875061035156, 0.037596160888671876, 0.03754598236083984, 0.03765350341796875, 0.03757670211791992, 0.03763507080078125, 0.03760947036743164, 0.03764940643310547, 0.03763715362548828, 0.03764118576049805, 0.03759209442138672, 0.037691360473632814, 0.037670944213867186, 0.03768931198120117, 0.037645313262939455, 0.03767398452758789, 0.0376258544921875, 0.037689342498779296, 0.03764121627807617, 0.03769651031494141, 0.037667839050292966, 0.0376627197265625, 0.03771289443969727, 0.037733375549316404, 0.03769548797607422, 0.037703678131103514, 0.037663745880126956, 0.037736446380615234, 0.03767705535888672, 0.03774566268920899, 0.037730304718017575, 0.03775388717651367, 0.03769443130493164, 0.037771297454833985, 0.03775484848022461, 0.03778559875488281, 0.03772108840942383, 0.03774771118164062, 0.0377784309387207, 0.03778358459472656, 0.03788694381713867, 0.037806079864501956, 0.037706783294677734, 0.03785929489135742, 0.03772108840942383, 0.03784089660644531, 0.14026956176757813, 0.037569534301757815, 0.03760025787353516, 0.03746918487548828, 0.037526527404785154, 0.03751321411132812, 0.03753267288208008, 0.037533695220947266, 0.037550079345703126, 0.03749990463256836, 0.03758489608764649, 0.03752243041992188, 0.037553150177001955, 0.037531646728515625, 0.037563392639160156, 0.037722110748291016, 0.03790643310546875, 0.03756137466430664, 0.03763504028320312, 0.03752243041992188, 0.0376360969543457, 0.03751935958862305, 0.03760537719726562, 0.037634048461914066, 0.03763302230834961, 0.03759308624267578, 0.03767295837402344, 0.037579776763916016, 0.037765121459960936, 0.037582847595214845, 0.03768217468261719, 0.037645313262939455, 0.03769036865234375, 0.03757875061035156, 0.037645313262939455, 0.03762790298461914, 0.03769753646850586, 0.037631999969482424, 0.037838848114013675, 0.03771088027954102, 0.03775382232666016, 0.0376627197265625, 0.03774156951904297, 0.037684223175048825, 0.037773311614990236, 0.03765657424926758, 0.03774566268920899, 0.03770265579223633, 0.037754878997802735, 0.037703678131103514, 0.03773132705688476, 0.03770880126953125, 0.03779379272460937, 0.03784294509887695, 0.03805184173583984, 0.03776716613769531, 0.03781324768066406, 0.03769241714477539, 0.03783679962158203, 0.03772313690185547, 0.03783168029785156, 0.037749759674072264, 0.03782656097412109, 0.14007501220703125, 0.037479423522949216, 0.037495807647705076, 0.037482494354248046, 0.0375623664855957, 0.03752345657348633, 0.037528575897216795, 0.037553150177001955, 0.03758694458007812, 0.037528575897216795, 0.03752140808105469, 0.03754086303710937, 0.03759718322753906, 0.03755929565429687, 0.037617664337158206, 0.03762790298461914, 0.03763916778564453, 0.03753676986694336, 0.03760748672485351, 0.03757868957519531, 0.03762790298461914, 0.037617664337158206, 0.03759206390380859, 0.03768217468261719, 0.03763814544677734, 0.037651454925537106, 0.03765862274169922, 0.03760435104370117, 0.03773747253417969, 0.037623809814453124, 0.03769139099121094, 0.03765964889526367, 0.0376627197265625, 0.03766681671142578, 0.037645313262939455, 0.037647361755371096, 0.03772415924072266, 0.037645313262939455, 0.03768115234375, 0.03770163345336914, 0.037716991424560545, 0.03767910385131836, 0.03774566268920899, 0.03769343948364258, 0.03773132705688476, 0.037718017578125, 0.03771596908569336, 0.03774979019165039, 0.0377476806640625, 0.03773132705688476, 0.03773545455932617, 0.03771184158325196, 0.037784576416015625, 0.0377077751159668, 0.037748737335205076, 0.037717025756835935, 0.03779580688476562, 0.03773235321044922, 0.037787647247314454, 0.037719039916992186, 0.03779993438720703, 0.03776204681396484, 0.03779072189331055, 0.1401692199707031, 0.03748147201538086, 0.03748454284667969, 0.03749478530883789, 0.03753267288208008, 0.03746303939819336, 0.03753574371337891, 0.037533695220947266, 0.03750502395629883, 0.03750400161743164, 0.03753676986694336, 0.03753267288208008, 0.037541889190673826, 0.03750400161743164, 0.037563392639160156, 0.03753779220581055, 0.037759998321533206, 0.037498878479003905, 0.03756748962402344, 0.03752447891235351, 0.037585918426513674, 0.03750707244873047, 0.0377077751159668, 0.037746688842773435, 0.0378081283569336, 0.03756851196289063, 0.037664768218994144, 0.037563392639160156, 0.03767500686645508, 0.03760332870483398, 0.037667873382568356, 0.03766883087158203, 0.037664768218994144, 0.03758694458007812, 0.03769651031494141, 0.03760947036743164, 0.03769343948364258, 0.037631999969482424, 0.03772927856445313, 0.037820449829101564, 0.038062049865722654, 0.03822387313842773, 0.037868545532226565, 0.03786649703979492, 0.03790848159790039, 0.037722110748291016, 0.03774156951904297, 0.03774054336547852, 0.037768192291259765, 0.03770880126953125, 0.037735424041748046, 0.037664768218994144, 0.03779891204833984, 0.037678081512451174, 0.03775692749023438, 0.03786137771606445, 0.0378869743347168, 0.03781119918823242, 0.03782656097412109, 0.03774771118164062, 0.037803009033203126, 0.037748737335205076, 0.037781505584716796, 0.14017945861816405, 0.0375654411315918, 0.03756748962402344, 0.03750912094116211, 0.03756032180786133, 0.037618686676025394, 0.03790233612060547, 0.037596160888671876, 0.03760332870483398, 0.03765964889526367, 0.03783782577514649, 0.03760947036743164, 0.037615615844726565, 0.03753676986694336, 0.03757670211791992, 0.0376360969543457, 0.03759718322753906, 0.03777536010742188, 0.03771599960327148, 0.03756745529174805, 0.03767910385131836, 0.03758796691894531, 0.03759929656982422, 0.03766470336914062, 0.0376258544921875, 0.03762790298461914, 0.03765043258666992, 0.037602302551269534, 0.037725215911865236, 0.037620704650878904, 0.03767910385131836, 0.03770675277709961, 0.03775897598266602, 0.03765657424926758, 0.037668895721435544, 0.037663711547851565, 0.03768832015991211, 0.03767507171630859, 0.037687232971191406, 0.037694465637207034, 0.0377149429321289, 0.03763203048706055, 0.037721057891845704, 0.037678081512451174, 0.03777536010742188, 0.03770163345336914, 0.0377446403503418, 0.03771289443969727, 0.03772528076171875, 0.037711776733398435, 0.03773132705688476, 0.03770675277709961, 0.03779180908203125, 0.037733375549316404, 0.037736385345458985, 0.03779379272460937, 0.03784294509887695, 0.03772415924072266, 0.03779174423217774, 0.03771084976196289, 0.037819393157958986, 0.037730304718017575, 0.03779891204833984, 0.14014566040039061, 0.037498878479003905, 0.03746815872192383, 0.03746310424804687, 0.0375428466796875, 0.03753676986694336, 0.037634048461914066, 0.037684223175048825, 0.03755417633056641, 0.03752140808105469, 0.03756032180786133, 0.03753267288208008, 0.037544960021972655, 0.037539840698242184, 0.03755212783813477, 0.03754905700683594, 0.03763814544677734, 0.03749990463256836, 0.03762790298461914, 0.03752243041992188, 0.037618686676025394, 0.03753676986694336, 0.03765248107910156, 0.037678081512451174, 0.03766886520385742, 0.037580799102783204, 0.03766579055786133, 0.03757875061035156, 0.037719039916992186, 0.03756851196289063, 0.03767295837402344, 0.0377077751159668, 0.03765248107910156, 0.03765248107910156, 0.03765248107910156, 0.03757875061035156, 0.03770982360839844, 0.037634048461914066, 0.0377149429321289, 0.03766681671142578, 0.037719039916992186, 0.03769651031494141, 0.0377446403503418, 0.03769139099121094, 0.03775078582763672, 0.0377149429321289, 0.0377446403503418, 0.037733375549316404, 0.03775283050537109, 0.03765043258666992, 0.03779174423217774, 0.037689342498779296, 0.03779174423217774, 0.03768627166748047, 0.03778867340087891, 0.037733375549316404, 0.037817344665527344, 0.03772825622558594, 0.037781505584716796, 0.037763072967529294, 0.037800960540771485, 0.037759998321533206, 0.03781836700439453]",tokens/s,25.53823297935188,,,main,False,False,,,,,,,,,,,,,,,,,,,,,,, bfloat16-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,M,M,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/M/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6693228f-4b38716e103486e76012c390;511f5d78-fc8b-4247-b3a5-c31d15cd6d87) Repository Not Found for url: https://huggingface.co/M/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: M is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,stabilityai/stablelm-base-alpha-3b,stabilityai/stablelm-base-alpha-3b,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1278.410752,8550.612992,0.0,7904.165888,7767.63392,s,10,0.4308936653137207,0.04308936653137207,0.0010250389855795494,0.04273472023010254,0.043290825271606444,0.04472251720428467,0.04586787075042725,"[0.046154209136962894, 0.042775230407714845, 0.04271712112426758, 0.04277427291870117, 0.042700542449951175, 0.042699935913085935, 0.04264531326293945, 0.0427523193359375, 0.04270204925537109, 0.04297267150878906]",tokens/s,5941.140949788949,kWh,5.067934451053399e-07,2.7769848706483547e-07,2.653374629820533e-06,3.4378665619907087e-06,tokens/kWh,74464786.62969464,MB,1278.410752,8550.612992,0.0,7904.165888,7864.441856,s,10,12.76272705078125,1.2762727050781248,0.00434060207440535,1.2757932739257813,1.2776046020507812,1.2828318420410156,1.2870136340332032,"[1.28805908203125, 1.2764429931640624, 1.276372802734375, 1.2761885986328125, 1.2748446044921875, 1.27594580078125, 1.269808349609375, 1.27436669921875, 1.275057373046875, 1.2756407470703126]",tokens/s,49.362491064277336,kWh,1.4897090492227028e-05,8.163402993046847e-06,6.329267812697846e-05,8.635317161225233e-05,tokens/kWh,729562.0858361289,,s,629,13.12617985916139,0.02086833045971602,0.005128378286637428,0.02024038314819336,0.020354457473754882,0.020546355056762694,0.06330331024169922,"[0.020552703857421875, 0.020477951049804686, 0.020585472106933594, 0.0206561279296875, 0.020584447860717774, 0.020806655883789063, 0.020800512313842775, 0.020551679611206054, 0.02062335968017578, 0.02067865562438965, 0.020986879348754883, 0.020717567443847656, 0.020607040405273436, 0.02052604866027832, 0.020660192489624023, 0.02071244812011719, 0.020538368225097657, 0.020914176940917968, 0.020776960372924806, 0.020740095138549804, 0.020455423355102538, 0.02058137512207031, 0.02046976089477539, 0.020516895294189454, 0.020623327255249024, 0.020365312576293947, 0.020206592559814454, 0.020260000228881837, 0.020202335357666017, 0.020247552871704103, 0.02026188850402832, 0.020273151397705077, 0.020238336563110353, 0.02024038314819336, 0.020220928192138672, 0.020171775817871093, 0.02028544044494629, 0.020289535522460937, 0.020337663650512695, 0.020307968139648438, 0.020295679092407228, 0.020254720687866212, 0.02026700782775879, 0.02025062370300293, 0.02028441619873047, 0.02025267219543457, 0.02030899238586426, 0.02022400093078613, 0.02023423957824707, 0.020299776077270508, 0.02022809600830078, 0.020316160202026368, 0.020296703338623046, 0.0202936954498291, 0.02038368034362793, 0.020280319213867186, 0.020336639404296874, 0.02040729522705078, 0.020279296875, 0.02036735916137695, 0.020311040878295897, 0.02030899238586426, 0.06350745773315429, 0.020151296615600587, 0.020188159942626953, 0.020189184188842774, 0.020145151138305666, 0.020185087203979494, 0.020239360809326173, 0.020222976684570314, 0.0202608642578125, 0.020195327758789062, 0.020220928192138672, 0.02020966339111328, 0.020191328048706055, 0.020155296325683594, 0.02021785545349121, 0.020196352005004883, 0.020175872802734376, 0.020190271377563476, 0.02017683219909668, 0.020191232681274415, 0.02030191993713379, 0.020139936447143555, 0.02020966339111328, 0.020297727584838866, 0.020327423095703127, 0.020525056838989256, 0.0202608642578125, 0.020220928192138672, 0.02031820869445801, 0.020204544067382812, 0.02024038314819336, 0.020384767532348632, 0.020313087463378905, 0.020242431640625, 0.02027008056640625, 0.02020351982116699, 0.020214784622192384, 0.02027724838256836, 0.020197376251220703, 0.020342784881591795, 0.020315135955810547, 0.02029465675354004, 0.020282367706298828, 0.020315135955810547, 0.02023628807067871, 0.020280351638793947, 0.020252639770507813, 0.02031718444824219, 0.02027724838256836, 0.02027724838256836, 0.02023628807067871, 0.02030284881591797, 0.020237312316894532, 0.020342784881591795, 0.02029363250732422, 0.020353023529052734, 0.020313087463378905, 0.02027008056640625, 0.020248575210571287, 0.020294687271118165, 0.020247520446777342, 0.020274175643920898, 0.02026905632019043, 0.06334566497802735, 0.02010316848754883, 0.020114431381225584, 0.020171775817871093, 0.020146175384521483, 0.020117504119873047, 0.02025164794921875, 0.020205568313598633, 0.020163583755493163, 0.020206592559814454, 0.02019327926635742, 0.020164608001708984, 0.02024140739440918, 0.02019945526123047, 0.020263904571533202, 0.02017791938781738, 0.02020867156982422, 0.020264928817749023, 0.020176895141601564, 0.020172800064086914, 0.020238336563110353, 0.020157440185546875, 0.020199424743652345, 0.02027724838256836, 0.020371456146240235, 0.020420608520507814, 0.020377599716186523, 0.020214784622192384, 0.020321279525756835, 0.02022707176208496, 0.02026905632019043, 0.02027724838256836, 0.020327423095703127, 0.020175872802734376, 0.02023423957824707, 0.02021171188354492, 0.02022400093078613, 0.02028544044494629, 0.020239360809326173, 0.020331520080566406, 0.020237312316894532, 0.02025984001159668, 0.020231168746948244, 0.02023628807067871, 0.020280319213867186, 0.020246528625488282, 0.02029465675354004, 0.020369407653808593, 0.020254720687866212, 0.020262912750244142, 0.02027212715148926, 0.020242431640625, 0.020340736389160157, 0.02019430351257324, 0.020770816802978515, 0.020324352264404297, 0.020339712142944336, 0.020298751831054687, 0.020329471588134765, 0.020281343460083007, 0.02032640075683594, 0.020222976684570314, 0.020256767272949217, 0.06347980880737304, 0.020106239318847655, 0.020162559509277343, 0.020189184188842774, 0.020134912490844727, 0.020152320861816408, 0.020174848556518556, 0.02025984001159668, 0.020191232681274415, 0.020212736129760742, 0.020187135696411132, 0.020172800064086914, 0.020231168746948244, 0.020139007568359374, 0.020200447082519533, 0.02027008056640625, 0.02019536018371582, 0.020240352630615233, 0.020222976684570314, 0.020190208435058594, 0.020327423095703127, 0.020188159942626953, 0.020152320861816408, 0.02022707176208496, 0.020195327758789062, 0.020287488937377928, 0.020197376251220703, 0.020196352005004883, 0.020246528625488282, 0.02020966339111328, 0.02046156883239746, 0.020297727584838866, 0.02024345588684082, 0.020305919647216796, 0.020249599456787108, 0.020220928192138672, 0.020321279525756835, 0.020168703079223634, 0.020303871154785155, 0.020538368225097657, 0.020331520080566406, 0.020280351638793947, 0.020190176010131837, 0.020152320861816408, 0.020282367706298828, 0.020212736129760742, 0.020354047775268554, 0.020316160202026368, 0.020322303771972656, 0.020247552871704103, 0.020343807220458983, 0.02021174430847168, 0.020228063583374024, 0.020238336563110353, 0.0204400634765625, 0.02029363250732422, 0.02030182456970215, 0.02025062370300293, 0.020316160202026368, 0.020295679092407228, 0.02030284881591797, 0.020342784881591795, 0.020248575210571287, 0.06342348861694336, 0.02012774467468262, 0.020153343200683595, 0.02016972732543945, 0.020197376251220703, 0.020166656494140626, 0.020231168746948244, 0.020116479873657226, 0.02003152084350586, 0.020243423461914063, 0.02023321533203125, 0.020167680740356447, 0.02023219108581543, 0.020208639144897463, 0.020136959075927736, 0.020298751831054687, 0.020185087203979494, 0.020190208435058594, 0.020198400497436524, 0.020180992126464844, 0.0202106876373291, 0.020239360809326173, 0.020192256927490236, 0.020350976943969725, 0.020289535522460937, 0.020262912750244142, 0.02016160011291504, 0.020233152389526367, 0.02020249557495117, 0.02018611145019531, 0.02022502326965332, 0.02024140739440918, 0.02020966339111328, 0.02028339195251465, 0.020239360809326173, 0.02021683120727539, 0.02019327926635742, 0.02024448013305664, 0.02027724838256836, 0.020287488937377928, 0.020271104812622072, 0.02023219108581543, 0.02030182456970215, 0.02024140739440918, 0.020254720687866212, 0.02020147132873535, 0.020152320861816408, 0.02028544044494629, 0.020182016372680665, 0.020256767272949217, 0.02025574493408203, 0.02022809600830078, 0.020272159576416017, 0.020260831832885743, 0.02040012741088867, 0.020327423095703127, 0.02025164794921875, 0.02027519989013672, 0.020268064498901367, 0.020248544692993163, 0.02025574493408203, 0.020248640060424806, 0.020307903289794923, 0.06323404693603515, 0.020121599197387697, 0.020154367446899413, 0.020116479873657226, 0.020142080307006836, 0.020163583755493163, 0.020164608001708984, 0.020239360809326173, 0.020166656494140626, 0.020151296615600587, 0.020205568313598633, 0.020166656494140626, 0.020137983322143553, 0.02020351982116699, 0.020279296875, 0.020197376251220703, 0.020208639144897463, 0.020231168746948244, 0.020230144500732423, 0.020246528625488282, 0.02021683120727539, 0.020179967880249023, 0.02019327926635742, 0.020237312316894532, 0.020199424743652345, 0.020214784622192384, 0.020160512924194338, 0.02025881576538086, 0.02024550437927246, 0.020191232681274415, 0.020280319213867186, 0.02045030403137207, 0.020282367706298828, 0.020271104812622072, 0.02024038314819336, 0.020239360809326173, 0.02027519989013672, 0.020212736129760742, 0.02025574493408203, 0.02017695999145508, 0.020264896392822265, 0.020239360809326173, 0.020164608001708984, 0.020249599456787108, 0.02021171188354492, 0.020154367446899413, 0.020282367706298828, 0.02030284881591797, 0.020289535522460937, 0.02027827262878418, 0.020810752868652343, 0.020759552001953126, 0.020335615158081053, 0.020273151397705077, 0.020321279525756835, 0.020313087463378905, 0.020344831466674804, 0.02022400093078613, 0.020330495834350586, 0.02026905632019043, 0.02027622413635254, 0.02019327926635742, 0.02025062370300293, 0.06325759887695312, 0.02004991912841797, 0.019951648712158203, 0.019928096771240234, 0.019958784103393554, 0.019933120727539062, 0.01998028755187988, 0.02004582405090332, 0.020006912231445313, 0.019963903427124022, 0.019974143981933593, 0.019997695922851562, 0.01999564743041992, 0.019994623184204103, 0.019978336334228516, 0.02003753662109375, 0.02000588798522949, 0.019985408782958985, 0.019989503860473632, 0.019982336044311523, 0.02003046417236328, 0.02004172706604004, 0.019992671966552734, 0.02001091194152832, 0.020011007308959963, 0.01997315216064453, 0.020007904052734376, 0.019986431121826173, 0.02001203155517578, 0.0200130558013916, 0.020007936477661133, 0.020098047256469728, 0.02044825553894043, 0.020730880737304686, 0.020406272888183592, 0.020271104812622072, 0.02021990394592285, 0.02022809600830078, 0.020329471588134765, 0.020222976684570314, 0.020289535522460937, 0.02027724838256836, 0.02023321533203125, 0.020223007202148438, 0.02033660888671875, 0.02025984001159668, 0.02028441619873047, 0.020335615158081053, 0.0202445125579834, 0.020210655212402343, 0.020257791519165038, 0.02026393508911133, 0.020239360809326173, 0.020384767532348632, 0.020274175643920898, 0.02031001663208008, 0.02030284881591797, 0.020279296875, 0.020262912750244142, 0.020303871154785155, 0.02031001663208008, 0.020304895401000975, 0.02025164794921875, 0.06333030319213867, 0.02018409538269043, 0.020183008193969728, 0.020129791259765627, 0.020142080307006836, 0.020154367446899413, 0.020185087203979494, 0.020208639144897463, 0.020165632247924805, 0.020154367446899413, 0.020130815505981444, 0.020196352005004883, 0.02020249557495117, 0.02012876892089844, 0.02022400093078613, 0.02025574493408203, 0.02020351982116699, 0.020248575210571287, 0.020205568313598633, 0.020168703079223634, 0.02026905632019043, 0.020256767272949217, 0.020248575210571287, 0.020271104812622072, 0.020241472244262697, 0.02016352081298828, 0.020191232681274415, 0.020237312316894532, 0.020154367446899413, 0.020215808868408205, 0.02026700782775879, 0.02023526382446289, 0.02028646469116211, 0.020213760375976563, 0.020222976684570314, 0.02025574493408203, 0.02023219108581543, 0.02027008056640625, 0.020215808868408205, 0.020356096267700196, 0.0202106876373291, 0.02024038314819336, 0.020222976684570314, 0.020180992126464844, 0.020280319213867186, 0.020182016372680665, 0.02024448013305664, 0.020322303771972656, 0.020220928192138672, 0.02027622413635254, 0.020237312316894532, 0.02027212715148926, 0.020273151397705077, 0.020310047149658204, 0.020232160568237303, 0.020368383407592772, 0.02022707176208496, 0.020254720687866212, 0.020142080307006836, 0.0202097282409668, 0.020232128143310546, 0.020131872177124022, 0.020253664016723634, 0.06332108688354492, 0.020102144241333008, 0.020176895141601564, 0.020140031814575195, 0.020131839752197265, 0.020166656494140626, 0.02019126319885254, 0.020153343200683595, 0.020174816131591798, 0.02019327926635742, 0.020132863998413086, 0.020058208465576172, 0.020181919097900392, 0.020180992126464844, 0.02019430351257324, 0.02033459281921387, 0.020174848556518556, 0.020157440185546875, 0.02026905632019043, 0.020171775817871093, 0.020146175384521483, 0.020215808868408205, 0.02022400093078613, 0.020198400497436524, 0.020242431640625, 0.02018611145019531, 0.020215808868408205, 0.020230144500732423, 0.020281343460083007, 0.020239360809326173, 0.020199424743652345, 0.020353023529052734, 0.020247552871704103, 0.020274175643920898, 0.02025984001159668, 0.02027212715148926, 0.02021785545349121, 0.02020249557495117, 0.02027008056640625, 0.020271104812622072, 0.020288511276245116, 0.020154367446899413, 0.020264959335327147, 0.02025267219543457, 0.020231168746948244, 0.02027008056640625, 0.02025984001159668, 0.02030080032348633, 0.020287488937377928, 0.02028441619873047, 0.020316160202026368, 0.02030284881591797, 0.020260896682739258, 0.020256736755371093, 0.02035916709899902, 0.020330495834350586, 0.02024140739440918, 0.020280319213867186, 0.02028544044494629, 0.020289535522460937, 0.020337663650512695, 0.02027008056640625, 0.020296703338623046, 0.06388326263427735, 0.020200447082519533, 0.020154367446899413, 0.020135936737060548, 0.020230144500732423, 0.020182016372680665, 0.020174848556518556, 0.020205568313598633, 0.02019327926635742, 0.020190208435058594, 0.02021990394592285, 0.020158464431762696, 0.020157440185546875, 0.020221952438354493, 0.020189184188842774, 0.020221952438354493, 0.02025062370300293, 0.020175872802734376, 0.020183040618896485, 0.020197376251220703, 0.02020249557495117, 0.0202608642578125, 0.02017791938781738, 0.02022400093078613, 0.02028646469116211, 0.020154367446899413, 0.020231168746948244, 0.020230144500732423, 0.020190208435058594, 0.020190208435058594, 0.02025267219543457, 0.02023526382446289, 0.020230144500732423, 0.020178943634033202, 0.020230144500732423, 0.020221952438354493, 0.020189184188842774, 0.020264959335327147, 0.020254720687866212, 0.020330495834350586, 0.020265983581542968, 0.020303871154785155, 0.02022400093078613, 0.02025881576538086, 0.020254720687866212, 0.020198400497436524, 0.02027724838256836, 0.020281343460083007, 0.020265983581542968, 0.020242431640625, 0.02023219108581543, 0.020183263778686525, 0.020179744720458984, 0.02030182456970215, 0.020306943893432617, 0.020312063217163084, 0.02022604751586914, 0.020365312576293947, 0.020217920303344728, 0.02051679992675781, 0.02046473693847656, 0.0201757755279541, 0.02048204803466797]",tokens/s,47.91950184660858,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,google/recurrentgemma-9b,google/recurrentgemma-9b,cuda,0,42,,,True,,,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",recurrent_gemma,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,mistralai/Mixtral-8x22B-v0.1,mistralai/Mixtral-8x22B-v0.1,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/base.py"", line 43, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 54, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1021, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 403 Client Error: Forbidden for url: https://huggingface.co/mistralai/Mixtral-8x22B-v0.1/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 398, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 321, in hf_raise_for_status raise GatedRepoError(message, response) from e huggingface_hub.utils._errors.GatedRepoError: 403 Client Error. (Request ID: Root=1-66475076-1d0aacba7a2d015b42431c25;c8cb8679-00e4-4c4b-ab75-75574e36d944) Cannot access gated repo for url https://huggingface.co/mistralai/Mixtral-8x22B-v0.1/resolve/main/config.json. Access to model mistralai/Mixtral-8x22B-v0.1 is restricted and you are not in the authorized list. Visit https://huggingface.co/mistralai/Mixtral-8x22B-v0.1 to ask for access. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 87, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/base.py"", line 56, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 37, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 62, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 928, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 631, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 686, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 416, in cached_file raise EnvironmentError( OSError: You are trying to access a gated repo. Make sure to have access to it at https://huggingface.co/mistralai/Mixtral-8x22B-v0.1. 403 Client Error. (Request ID: Root=1-66475076-1d0aacba7a2d015b42431c25;c8cb8679-00e4-4c4b-ab75-75574e36d944) Cannot access gated repo for url https://huggingface.co/mistralai/Mixtral-8x22B-v0.1/resolve/main/config.json. Access to model mistralai/Mixtral-8x22B-v0.1 is restricted and you are not in the authorized list. Visit https://huggingface.co/mistralai/Mixtral-8x22B-v0.1 to ask for access. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,01-ai/Yi-6B,01-ai/Yi-6B,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1263.5136,13032.226816,0.0,12385.779712,12315.692032,s,10,0.9027714920043947,0.09027714920043946,0.0003073541008434986,0.09017732620239258,0.09042406234741211,0.09078577690124512,0.09107514854431152,"[0.09114749145507813, 0.09010883331298829, 0.0901373748779297, 0.09004342651367188, 0.09025225830078125, 0.09015289306640625, 0.0902017593383789, 0.09004361724853516, 0.09034368133544922, 0.09034015655517579]",tokens/s,2835.712051912621,kWh,1.0671894337321905e-06,5.847706018136483e-07,5.1154520403063185e-06,6.767412075852157e-06,tokens/kWh,37828345.182861395,MB,1263.5136,13032.226816,0.0,12385.779712,12315.694592,s,10,22.4199111328125,2.24199111328125,0.010449292192832046,2.238900024414063,2.2485163818359375,2.259043591308594,2.2674653588867186,"[2.26957080078125, 2.2383564453125, 2.229277587890625, 2.24528125, 2.243859619140625, 2.238340576171875, 2.23397265625, 2.235631591796875, 2.239443603515625, 2.246177001953125]",tokens/s,28.10002217528722,kWh,2.6682772180876576e-05,1.4623037018358738e-05,9.997385425329341e-05,0.00014127966345252877,tokens/kWh,445924.05205699377,,s,629,23.193609245300316,0.036873782583943235,0.010795172671264926,0.035423168182373045,0.03611709442138672,0.03682080764770508,0.12593057922363282,"[0.037945343017578126, 0.0377784309387207, 0.03743539047241211, 0.03738009643554688, 0.03809280014038086, 0.03781836700439453, 0.03786751937866211, 0.03754905700683594, 0.03729715347290039, 0.0366929931640625, 0.03640627288818359, 0.037323776245117186, 0.03779481506347656, 0.03707596969604492, 0.03611545562744141, 0.035329025268554685, 0.03527577590942383, 0.035282943725585936, 0.03533824157714844, 0.03532492828369141, 0.03524198532104492, 0.03532492828369141, 0.03537100982666016, 0.03526144027709961, 0.03552972793579102, 0.03748044967651367, 0.03624857711791992, 0.03549900817871094, 0.035372032165527346, 0.03778867340087891, 0.03685891342163086, 0.0354549446105957, 0.03546419143676758, 0.03544575881958008, 0.03627315139770508, 0.035655681610107424, 0.03532492828369141, 0.035448833465576174, 0.03551027297973633, 0.03538227081298828, 0.035323902130126955, 0.03536896133422852, 0.035332096099853515, 0.03540895843505859, 0.03534534454345703, 0.03536383819580078, 0.0353361930847168, 0.03537919998168945, 0.03531468963623047, 0.0353966064453125, 0.03545292663574219, 0.035361793518066405, 0.035312641143798826, 0.035681278228759765, 0.03563827133178711, 0.035334144592285156, 0.03529318237304688, 0.035348480224609374, 0.035356670379638674, 0.0353177604675293, 0.035381248474121094, 0.03537919998168945, 0.12592025756835937, 0.03539558410644531, 0.035225601196289064, 0.03537715148925781, 0.03528396987915039, 0.03536588668823242, 0.03528499221801758, 0.03764019012451172, 0.03668479919433594, 0.03582262420654297, 0.03553174209594727, 0.03537612915039062, 0.035369983673095705, 0.035745792388916016, 0.03543040084838867, 0.035432449340820314, 0.03534438323974609, 0.0354334716796875, 0.03545804977416992, 0.03545702362060547, 0.03537408065795898, 0.03540889739990234, 0.03530854415893555, 0.03629875183105469, 0.03618201446533203, 0.03542937469482422, 0.035434497833251956, 0.0354150390625, 0.0354252815246582, 0.03552358245849609, 0.03540991973876953, 0.035566593170166014, 0.03543756866455078, 0.03550822448730469, 0.03546419143676758, 0.03550003051757813, 0.035972095489501955, 0.035860481262207033, 0.035422206878662106, 0.035465217590332034, 0.035326976776123044, 0.03562496185302735, 0.03550310516357422, 0.03535769653320313, 0.03529830551147461, 0.03530956649780274, 0.03533824157714844, 0.03546623992919922, 0.03544575881958008, 0.035299327850341795, 0.03564748764038086, 0.03558297729492187, 0.03535871887207031, 0.0352624626159668, 0.035416065216064455, 0.03545395278930664, 0.035389438629150394, 0.03536383819580078, 0.03545091247558594, 0.035331039428710936, 0.03533004760742187, 0.03538431930541992, 0.035566593170166014, 0.1260800018310547, 0.0353259506225586, 0.03534745788574219, 0.03530342483520508, 0.03539148712158203, 0.035416065216064455, 0.035304447174072266, 0.03538431930541992, 0.03525120162963867, 0.03527167892456055, 0.03530649566650391, 0.0353259506225586, 0.03527679824829102, 0.03532799911499023, 0.035778560638427735, 0.03533926391601563, 0.03528704071044922, 0.03535974502563476, 0.03533926391601563, 0.035266559600830076, 0.03526144027709961, 0.03540172958374024, 0.03541100692749023, 0.035426239013671874, 0.03529216003417969, 0.035996673583984375, 0.03560038375854492, 0.0353361930847168, 0.035305503845214845, 0.03532182312011719, 0.03531366348266601, 0.03544780731201172, 0.0353331184387207, 0.035438591003417966, 0.03538534545898438, 0.035419136047363284, 0.03533926391601563, 0.03533926391601563, 0.035432449340820314, 0.035449886322021486, 0.03536278533935547, 0.03528192138671875, 0.03604889678955078, 0.03544063949584961, 0.035326976776123044, 0.03535769653320313, 0.035279872894287106, 0.035345409393310545, 0.03527782440185547, 0.035332096099853515, 0.03532291030883789, 0.03526652908325195, 0.0353966064453125, 0.03525632095336914, 0.03528396987915039, 0.03541708755493164, 0.03544473648071289, 0.03533107376098633, 0.03527679824829102, 0.035178497314453126, 0.03560038375854492, 0.03541401672363281, 0.03527782440185547, 0.12580044555664063, 0.035315711975097655, 0.035264511108398434, 0.03524300765991211, 0.03559423828125, 0.036241409301757815, 0.03526860809326172, 0.0353361930847168, 0.03527372741699219, 0.03528499221801758, 0.036552703857421875, 0.03542937469482422, 0.035520511627197264, 0.035364864349365234, 0.035315711975097655, 0.035381248474121094, 0.03530035018920898, 0.03541299057006836, 0.035361793518066405, 0.03577446365356445, 0.0354252815246582, 0.03545292663574219, 0.035361793518066405, 0.03542015838623047, 0.03553792190551758, 0.035506175994873046, 0.035434497833251956, 0.035489791870117186, 0.035404800415039066, 0.0354334716796875, 0.03547955322265625, 0.03554304122924805, 0.035566593170166014, 0.03562700653076172, 0.03550207901000976, 0.03544985580444336, 0.03533824157714844, 0.035492862701416016, 0.03572531127929687, 0.03552153778076172, 0.03544678497314453, 0.03540172958374024, 0.03542937469482422, 0.035399681091308595, 0.035318782806396484, 0.03541196823120117, 0.035389438629150394, 0.036357120513916014, 0.0361973762512207, 0.0362239990234375, 0.03583692932128906, 0.035383296966552735, 0.0357283821105957, 0.035579902648925785, 0.03624755096435547, 0.035476478576660156, 0.03531980895996094, 0.03536793518066406, 0.03601203155517578, 0.035675136566162106, 0.03661004638671875, 0.039000064849853515, 0.036757503509521484, 0.12629913330078124, 0.03549491119384766, 0.03553996658325195, 0.035394561767578124, 0.035418113708496096, 0.0354150390625, 0.03549798583984375, 0.03580108642578125, 0.03545804977416992, 0.03558911895751953, 0.035582015991210934, 0.03563411331176758, 0.03537408065795898, 0.035432449340820314, 0.03616767883300781, 0.03556454467773437, 0.035460094451904296, 0.03530342483520508, 0.035386367797851564, 0.03537612915039062, 0.03545804977416992, 0.03549184036254883, 0.03611443328857422, 0.03635001754760742, 0.035375038146972654, 0.03537408065795898, 0.03543251037597656, 0.03547539138793945, 0.035237918853759764, 0.03560547256469727, 0.03554508972167969, 0.035661823272705076, 0.03590963363647461, 0.03553484725952148, 0.03567001724243164, 0.03567308807373047, 0.03584102249145508, 0.03598745727539063, 0.035694625854492186, 0.035504161834716795, 0.03547846221923828, 0.03607961654663086, 0.03541708755493164, 0.03551846313476562, 0.035620864868164064, 0.0366827507019043, 0.03597312164306641, 0.03546623992919922, 0.03529734420776367, 0.035240894317626954, 0.03686707305908203, 0.03692544174194336, 0.03540889739990234, 0.035168254852294925, 0.03589529418945313, 0.03605913543701172, 0.03530956649780274, 0.035280895233154294, 0.03527679824829102, 0.03530342483520508, 0.03526553726196289, 0.03524095916748047, 0.03528195190429687, 0.12692684936523438, 0.03539254379272461, 0.035257278442382814, 0.035272705078125, 0.03520000076293945, 0.035315711975097655, 0.0352624626159668, 0.03575296020507813, 0.035253246307373046, 0.035381248474121094, 0.035310592651367184, 0.03532492828369141, 0.03540070343017578, 0.03551027297973633, 0.035639297485351565, 0.03541196823120117, 0.03560857772827149, 0.035253246307373046, 0.03529523086547852, 0.03530956649780274, 0.03541299057006836, 0.03542323303222656, 0.03534438323974609, 0.03551129531860352, 0.03550310516357422, 0.035768318176269534, 0.036291584014892575, 0.037174270629882815, 0.03606016159057617, 0.03535871887207031, 0.03536896133422852, 0.0356864013671875, 0.035620864868164064, 0.03561062240600586, 0.03554816055297851, 0.03548672103881836, 0.035389438629150394, 0.03551334381103516, 0.03549184036254883, 0.03545907211303711, 0.035451904296875, 0.03540787124633789, 0.035642368316650394, 0.03532799911499023, 0.03536588668823242, 0.035366912841796876, 0.03559936141967773, 0.03535366439819336, 0.03547846221923828, 0.03538022232055664, 0.03537510299682617, 0.03533004760742187, 0.03529420852661133, 0.03585740661621094, 0.03543552017211914, 0.03598745727539063, 0.035542015075683595, 0.035522560119628906, 0.0357212142944336, 0.03541708755493164, 0.03552358245849609, 0.035448833465576174, 0.03559628677368164, 0.12618342590332032, 0.03551129531860352, 0.03540889739990234, 0.035279872894287106, 0.03537510299682617, 0.03617587280273438, 0.03536588668823242, 0.035386367797851564, 0.03539353561401367, 0.035419136047363284, 0.03535871887207031, 0.03536588668823242, 0.03564441680908203, 0.03549798583984375, 0.03534131240844727, 0.0354252815246582, 0.03541401672363281, 0.0353546257019043, 0.03543756866455078, 0.035522560119628906, 0.035378177642822264, 0.035323902130126955, 0.03551337432861328, 0.03542012786865235, 0.035432449340820314, 0.03540377426147461, 0.03548160171508789, 0.03540582275390625, 0.03544575881958008, 0.03552358245849609, 0.03549491119384766, 0.03557785415649414, 0.03548672103881836, 0.03560857772827149, 0.03566796875, 0.03565055847167969, 0.035297279357910154, 0.03548057556152344, 0.03541302490234375, 0.035617759704589844, 0.035404800415039066, 0.03542732620239258, 0.03528806304931641, 0.03540377426147461, 0.03553792190551758, 0.03556966400146484, 0.03544063949584961, 0.03604275131225586, 0.03534342575073242, 0.03536787033081055, 0.0353331184387207, 0.035369983673095705, 0.03544268798828125, 0.035372032165527346, 0.035315711975097655, 0.03538534545898438, 0.03537100982666016, 0.03528396987915039, 0.035345409393310545, 0.03546419143676758, 0.03522969436645508, 0.035550209045410154, 0.03544063949584961, 0.1259345932006836, 0.03536281585693359, 0.03528908920288086, 0.0354304313659668, 0.03524095916748047, 0.03528291320800781, 0.035326976776123044, 0.03545395278930664, 0.03535564804077149, 0.035318782806396484, 0.035737598419189456, 0.035348480224609374, 0.03528908920288086, 0.03526144027709961, 0.03533926391601563, 0.03546419143676758, 0.03540070343017578, 0.03549388885498047, 0.03539148712158203, 0.035329025268554685, 0.03549491119384766, 0.035373054504394534, 0.035326976776123044, 0.03549184036254883, 0.0360816650390625, 0.035389438629150394, 0.035438591003417966, 0.0354252815246582, 0.03540991973876953, 0.0354068489074707, 0.035381248474121094, 0.035547134399414065, 0.03559628677368164, 0.03558297729492187, 0.03540486526489258, 0.03541804885864258, 0.0355153923034668, 0.03550822448730469, 0.0355431022644043, 0.03549484634399414, 0.03545600128173828, 0.03540787124633789, 0.03544678497314453, 0.03542015838623047, 0.03531468963623047, 0.03536896133422852, 0.035566593170166014, 0.03544268798828125, 0.03542732620239258, 0.035369983673095705, 0.03546828842163086, 0.03550515365600586, 0.036057086944580076, 0.03543961715698242, 0.035465217590332034, 0.035386367797851564, 0.035342334747314456, 0.03538227081298828, 0.03676364898681641, 0.03612364959716797, 0.03581235122680664, 0.03546623992919922, 0.03539046478271484, 0.12637081909179687, 0.035499038696289065, 0.03523273468017578, 0.03536793518066406, 0.035363903045654295, 0.035423168182373045, 0.03548876953125, 0.03588198471069336, 0.035438591003417966, 0.035590145111083986, 0.03564031982421875, 0.03560860824584961, 0.03535459136962891, 0.03585126495361328, 0.03611238479614258, 0.03641139221191406, 0.035920894622802735, 0.0354703369140625, 0.0353331184387207, 0.035334144592285156, 0.03537510299682617, 0.035415103912353516, 0.03537510299682617, 0.03546003341674805, 0.03534643173217773, 0.035342334747314456, 0.03538534545898438, 0.035410945892333984, 0.03537100982666016, 0.036212734222412106, 0.036157440185546875, 0.03552870559692383, 0.03544268798828125, 0.03551027297973633, 0.03548057556152344, 0.03584000015258789, 0.03553484725952148, 0.03541299057006836, 0.03549798583984375, 0.03583283233642578, 0.0353259506225586, 0.03536896133422852, 0.035348480224609374, 0.035345409393310545, 0.035348480224609374, 0.035318782806396484, 0.03529734420776367, 0.035428287506103516, 0.03539763259887695, 0.03522457504272461, 0.03534950256347656, 0.03524607849121094, 0.03534438323974609, 0.03537919998168945, 0.03546214294433594, 0.035343360900878903, 0.03539046478271484, 0.03630284881591797, 0.03613491058349609, 0.03605811309814453, 0.03585126495361328, 0.035606529235839846, 0.03532492828369141, 0.12794060516357422, 0.03554099273681641, 0.03534950256347656, 0.03540787124633789, 0.035379230499267576, 0.03814601516723633, 0.03671039962768555, 0.03556454467773437, 0.036195327758789066, 0.03561676788330078, 0.03538022232055664, 0.03530342483520508, 0.035419136047363284, 0.03528704071044922, 0.035369983673095705, 0.03540582275390625, 0.035337215423583986, 0.03521023941040039, 0.03540377426147461, 0.036013057708740234, 0.03557785415649414, 0.03548470306396485, 0.03525833511352539, 0.03541299057006836, 0.03540172958374024, 0.035353599548339845, 0.03539763259887695, 0.0353361930847168, 0.035381248474121094, 0.03530854415893555, 0.03537100982666016, 0.035434497833251956, 0.03544473648071289, 0.03580416107177734, 0.035460094451904296, 0.035422206878662106, 0.035369983673095705, 0.035388416290283206, 0.03562393569946289, 0.035536895751953124, 0.035448833465576174, 0.03539046478271484, 0.03549388885498047, 0.03532185745239258, 0.03535769653320313, 0.03559731292724609, 0.03766886520385742, 0.03710771179199219, 0.03554304122924805, 0.03542425537109375, 0.03549593734741211, 0.035381248474121094, 0.03532492828369141, 0.0358922233581543, 0.03613695907592773, 0.035740673065185545, 0.03533824157714844, 0.035250175476074216, 0.03567923355102539, 0.036219905853271485, 0.03537510299682617, 0.03575807952880859, 0.03569356918334961]",tokens/s,27.119539410514722,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,/,/,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 373, in cached_file raise EnvironmentError( OSError: / does not appear to have a file named config.json. Checkout 'https://huggingface.co///tree/None' for available files. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,TencentARC/Mistral_Pro_8B_v0.1,,cuda,0,42,,,,,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,d35829e539df8480b726c647eeabf91e41eae047,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,,MB,1271.263232,19665.518592,0.0,19019.071488,18747.268096,s,10,1.061557647705078,0.10615576477050781,0.000378162725663041,0.10603825378417969,0.10641835021972655,0.10679555892944335,0.1070973258972168,"[0.10717276763916016, 0.10601692962646485, 0.10626345825195313, 0.1059658203125, 0.10606905364990235, 0.10567481231689453, 0.10598377227783203, 0.106017822265625, 0.10633452606201171, 0.10605868530273438]",tokens/s,2411.550616713393,kWh,1.2540946909558703e-06,6.871848049539899e-07,7.603906667915486e-06,9.545186163825347e-06,tokens/kWh,26819801.68916946,MB,1271.263232,19667.615744,0.0,19021.16864,18774.964224,s,10,29.298705078124996,2.9298705078125,0.002554372097692277,2.929034912109375,2.930756665039062,2.934041906738281,2.9366701000976563,"[2.9373271484375, 2.928263671875, 2.92938623046875, 2.9291044921875, 2.928641845703125, 2.92896533203125, 2.930026611328125, 2.9289599609375, 2.929908447265625, 2.928121337890625]",tokens/s,21.502656800705186,kWh,3.453055689645092e-05,1.8924356512781838e-05,0.00014346296418548472,0.0001969178775947175,tokens/kWh,319930.3220688888,,s,629,30.205779903411873,0.04802190763658484,0.012651519007599461,0.0464824333190918,0.04661575622558593,0.04671159210205078,0.15297273376464843,"[0.047120319366455075, 0.047126529693603515, 0.047168510437011715, 0.04713062286376953, 0.047126529693603515, 0.047142913818359375, 0.04723814392089844, 0.04703641510009766, 0.04691763305664062, 0.04690534210205078, 0.046714881896972656, 0.04643532943725586, 0.046432254791259765, 0.046367774963378905, 0.04642095947265625, 0.04642611312866211, 0.046527488708496094, 0.0464721908569336, 0.046430206298828124, 0.04636159896850586, 0.04644454574584961, 0.046377983093261715, 0.04643430328369141, 0.046432254791259765, 0.04641180801391601, 0.04648956680297851, 0.046465057373046875, 0.046426078796386716, 0.0464640007019043, 0.04650086212158203, 0.04653055953979492, 0.046459903717041014, 0.04641996765136719, 0.04649062347412109, 0.046440448760986325, 0.04651520156860352, 0.046450687408447267, 0.046511104583740234, 0.0466319351196289, 0.046535678863525394, 0.046470142364501955, 0.04667801666259765, 0.04660224151611328, 0.0465715217590332, 0.04654182434082031, 0.04655923080444336, 0.04659814453125, 0.04662169647216797, 0.04654796981811524, 0.04660736083984375, 0.04650700759887695, 0.0465797119140625, 0.04653055953979492, 0.046604286193847655, 0.04653772735595703, 0.04658790588378906, 0.046529537200927736, 0.0465797119140625, 0.046548992156982424, 0.046755840301513675, 0.04658790588378906, 0.04661043167114258, 0.1530030059814453, 0.046432254791259765, 0.04652236938476562, 0.046448638916015625, 0.046289920806884766, 0.04637286376953125, 0.04630732727050781, 0.046430206298828124, 0.04634316635131836, 0.04641484832763672, 0.04635238265991211, 0.046412799835205076, 0.0463370246887207, 0.04643539047241211, 0.04630931091308594, 0.04643430328369141, 0.046357505798339846, 0.046497791290283204, 0.046360576629638675, 0.04642816162109375, 0.046371841430664064, 0.04642816162109375, 0.04634931182861328, 0.046473217010498044, 0.046358528137207033, 0.046491649627685545, 0.04638412857055664, 0.046440448760986325, 0.04643123245239258, 0.04645478439331055, 0.04643840026855469, 0.04649369430541992, 0.04640768051147461, 0.04653875350952148, 0.04643635177612305, 0.04656435012817383, 0.04662579345703125, 0.04656742477416992, 0.046496768951416016, 0.04655513763427734, 0.046445568084716796, 0.046562305450439455, 0.046458881378173826, 0.046532608032226565, 0.04653875350952148, 0.046504001617431644, 0.04650899124145508, 0.04660736083984375, 0.046639102935791016, 0.04659097671508789, 0.046508033752441405, 0.046543872833251954, 0.046562305450439455, 0.04649369430541992, 0.046524417877197265, 0.046599166870117184, 0.046548992156982424, 0.04655206298828125, 0.04653875350952148, 0.04655718231201172, 0.04661148834228516, 0.046559200286865235, 0.04652032089233398, 0.15322213745117189, 0.04631961441040039, 0.04639129638671875, 0.04642406463623047, 0.04642508697509766, 0.04638003158569336, 0.04640768051147461, 0.04656639862060547, 0.046789630889892575, 0.04642307281494141, 0.04638307189941406, 0.04636262512207031, 0.046394367218017575, 0.046354496002197265, 0.04638304138183594, 0.04642611312866211, 0.04642816162109375, 0.046442497253417966, 0.046358528137207033, 0.04639744186401367, 0.04642508697509766, 0.0464005126953125, 0.0464793586730957, 0.04658687973022461, 0.046456832885742184, 0.04646915054321289, 0.04644144058227539, 0.046430206298828124, 0.046458881378173826, 0.04644966506958008, 0.04654796981811524, 0.046477375030517576, 0.046449600219726564, 0.04652544021606445, 0.0464640007019043, 0.04646912002563477, 0.046465023040771485, 0.0464640007019043, 0.046630912780761716, 0.04650393676757812, 0.046491649627685545, 0.046486526489257815, 0.04656028747558594, 0.0465489616394043, 0.04655513763427734, 0.04651827239990235, 0.04656947326660156, 0.04660224151611328, 0.0466063346862793, 0.046514175415039063, 0.04652646255493164, 0.04649574279785156, 0.04659097671508789, 0.046489601135253904, 0.046601215362548826, 0.04660947036743164, 0.046621631622314454, 0.046561279296875, 0.046622718811035156, 0.04659609603881836, 0.04658073425292969, 0.046502910614013675, 0.04657052612304687, 0.1529528045654297, 0.04636675262451172, 0.04630527877807617, 0.04635337448120117, 0.04630835342407227, 0.04639846420288086, 0.04636982345581055, 0.046577632904052736, 0.046399486541748046, 0.046443519592285154, 0.04633091354370117, 0.04640047836303711, 0.04633292770385742, 0.04642201614379883, 0.04640563201904297, 0.046701568603515625, 0.04634316635131836, 0.0464640007019043, 0.046325759887695314, 0.04649062347412109, 0.04635443115234375, 0.04643942260742188, 0.04637286376953125, 0.04644454574584961, 0.04639846420288086, 0.046565376281738284, 0.04691865539550781, 0.0466063346862793, 0.04639334487915039, 0.04645273590087891, 0.04638924789428711, 0.046491649627685545, 0.04643840026855469, 0.04645478439331055, 0.04640153503417969, 0.046499839782714845, 0.04640975952148438, 0.046609375, 0.046465023040771485, 0.046514175415039063, 0.04644147109985351, 0.04648857498168945, 0.04650188827514649, 0.04655411148071289, 0.04653055953979492, 0.046543872833251954, 0.04659814453125, 0.04668723297119141, 0.0466063346862793, 0.046604320526123046, 0.04664112091064453, 0.0465530891418457, 0.04652544021606445, 0.04649577713012695, 0.04655510330200195, 0.04654489517211914, 0.04653772735595703, 0.046561344146728516, 0.04660012817382812, 0.046578689575195314, 0.04655001449584961, 0.04654489517211914, 0.04655001449584961, 0.15306343078613283, 0.04638003158569336, 0.046388225555419924, 0.0463636474609375, 0.04636979293823242, 0.04636774444580078, 0.04639030456542969, 0.046409694671630856, 0.046382080078125, 0.04640768051147461, 0.0463267822265625, 0.046413822174072264, 0.046312446594238284, 0.04638003158569336, 0.04648041534423828, 0.04643119812011719, 0.046478336334228515, 0.04650188827514649, 0.04642303848266602, 0.04637593460083008, 0.046511104583740234, 0.046635009765625, 0.04640665435791016, 0.04645580673217774, 0.04648038482666016, 0.04643635177612305, 0.04651007843017578, 0.04645273590087891, 0.04642816162109375, 0.046465023040771485, 0.04645478439331055, 0.046486526489257815, 0.046461952209472655, 0.04639129638671875, 0.046448638916015625, 0.0464005126953125, 0.046450687408447267, 0.04646092987060547, 0.046432254791259765, 0.046499839782714845, 0.046461952209472655, 0.046442497253417966, 0.04653363037109375, 0.04646092987060547, 0.04653158569335938, 0.04648755264282227, 0.046532608032226565, 0.04655513763427734, 0.04657356643676758, 0.046453758239746096, 0.04656435012817383, 0.046448638916015625, 0.04669235229492188, 0.046767105102539064, 0.04660019302368164, 0.046685184478759766, 0.046700607299804686, 0.046523326873779296, 0.04659404754638672, 0.04652032089233398, 0.04656947326660156, 0.04656332778930664, 0.04657356643676758, 0.15262413024902344, 0.046357505798339846, 0.0463001594543457, 0.046394367218017575, 0.04628070449829102, 0.04634016036987305, 0.046316478729248045, 0.04642611312866211, 0.04638515090942383, 0.04638105773925781, 0.046339073181152345, 0.046427135467529294, 0.0463185920715332, 0.0464005126953125, 0.046615550994873044, 0.04682956695556641, 0.046461952209472655, 0.04657664108276367, 0.046383102416992186, 0.04640256118774414, 0.04637388610839844, 0.046430206298828124, 0.0463370246887207, 0.04648755264282227, 0.04645478439331055, 0.04648550415039063, 0.04639539337158203, 0.04662783813476563, 0.04668108749389648, 0.04644147109985351, 0.04641689682006836, 0.046473217010498044, 0.046440448760986325, 0.04657561492919922, 0.046445568084716796, 0.04649062347412109, 0.04650502395629883, 0.04648134231567383, 0.04642508697509766, 0.04655923080444336, 0.046456832885742184, 0.04655411148071289, 0.046454814910888674, 0.046521312713623045, 0.04653977584838867, 0.04654182434082031, 0.04650495910644531, 0.04656025695800781, 0.046529537200927736, 0.04655615997314453, 0.04648755264282227, 0.04652339172363281, 0.04654694366455078, 0.04648857498168945, 0.046494720458984375, 0.046635009765625, 0.04659711837768555, 0.046578689575195314, 0.04653055953979492, 0.046561279296875, 0.046739456176757815, 0.04659404754638672, 0.04656742477416992, 0.15298048400878905, 0.046327808380126956, 0.04640358352661133, 0.04637081527709961, 0.04638719940185547, 0.04634828948974609, 0.046383102416992186, 0.04639846420288086, 0.046418975830078125, 0.046588897705078125, 0.046437374114990236, 0.04636467361450195, 0.04635647964477539, 0.0464793586730957, 0.04644966506958008, 0.04642508697509766, 0.04635955047607422, 0.046448638916015625, 0.04636671829223633, 0.04640774536132813, 0.04642502212524414, 0.04641388702392578, 0.046451648712158206, 0.046496768951416016, 0.046430206298828124, 0.04646297454833984, 0.04645478439331055, 0.04655718231201172, 0.046450687408447267, 0.04643123245239258, 0.046458881378173826, 0.04645580673217774, 0.04647423934936523, 0.046494720458984375, 0.04656742477416992, 0.04651219177246094, 0.046430145263671875, 0.04645273590087891, 0.04645785522460937, 0.046496768951416016, 0.046513153076171876, 0.04644147109985351, 0.046499839782714845, 0.04652646255493164, 0.04667391967773438, 0.04695049667358398, 0.047195041656494144, 0.04677632141113281, 0.046615550994873044, 0.0464721908569336, 0.04655411148071289, 0.04652339172363281, 0.0466165771484375, 0.046508033752441405, 0.046604286193847655, 0.04659404754638672, 0.04657459259033203, 0.046545921325683595, 0.046601215362548826, 0.04650700759887695, 0.04657564926147461, 0.04650908660888672, 0.046569408416748045, 0.15299072265625, 0.046357505798339846, 0.046333953857421874, 0.04639334487915039, 0.0464087028503418, 0.0464005126953125, 0.04631347274780274, 0.04646403121948242, 0.04631753540039062, 0.046437374114990236, 0.046311424255371096, 0.04638924789428711, 0.04634931182861328, 0.0464640007019043, 0.04636159896850586, 0.04645785522460937, 0.04635955047607422, 0.04645171356201172, 0.04636671829223633, 0.04652032089233398, 0.0463370246887207, 0.04645273590087891, 0.04641177749633789, 0.046443519592285154, 0.046410751342773435, 0.046465023040771485, 0.04637286376953125, 0.046461952209472655, 0.04641689682006836, 0.046456832885742184, 0.04640358352661133, 0.0464640007019043, 0.0464415054321289, 0.04670665740966797, 0.04679167938232422, 0.04655001449584961, 0.04644147109985351, 0.04649574279785156, 0.04646809768676758, 0.046545921325683595, 0.04658687973022461, 0.046652416229248046, 0.04650495910644531, 0.046581760406494144, 0.04652032089233398, 0.046584831237792966, 0.04655820846557617, 0.04664627075195313, 0.04658790588378906, 0.046562305450439455, 0.04654489517211914, 0.04655206298828125, 0.04655104064941406, 0.046475265502929686, 0.04654694366455078, 0.046653438568115234, 0.04659199905395508, 0.04660019302368164, 0.04652236938476562, 0.04657049560546875, 0.04653977584838867, 0.04657254409790039, 0.04655513763427734, 0.15333888244628907, 0.046357505798339846, 0.04640665435791016, 0.04636368179321289, 0.04637795257568359, 0.04640665435791016, 0.04637696075439453, 0.04643430328369141, 0.046358528137207033, 0.0464824333190918, 0.0466431999206543, 0.046442497253417966, 0.046595073699951174, 0.04654694366455078, 0.04644454574584961, 0.046450687408447267, 0.04642816162109375, 0.04664524841308594, 0.046399486541748046, 0.046413822174072264, 0.046486526489257815, 0.04645478439331055, 0.046456832885742184, 0.04643635177612305, 0.04651724624633789, 0.04659097671508789, 0.04652339172363281, 0.04642508697509766, 0.046442497253417966, 0.04642611312866211, 0.04646604919433594, 0.04643942260742188, 0.04646912002563477, 0.0464793586730957, 0.046475265502929686, 0.046473217010498044, 0.04646297454833984, 0.04643840026855469, 0.04646912002563477, 0.04650495910644531, 0.046558238983154296, 0.046458847045898435, 0.04650495910644531, 0.046513153076171876, 0.046529537200927736, 0.04651007843017578, 0.04673023986816406, 0.04679884719848633, 0.04659814453125, 0.04653363037109375, 0.04654489517211914, 0.04648755264282227, 0.046548992156982424, 0.046475265502929686, 0.04661043167114258, 0.046637054443359374, 0.04655718231201172, 0.04652339172363281, 0.04658585739135742, 0.046527488708496094, 0.046617599487304685, 0.046519294738769534, 0.046558238983154296, 0.15303471374511718, 0.046358528137207033, 0.04637491226196289, 0.04643635177612305, 0.046312446594238284, 0.04635647964477539, 0.04632985687255859, 0.04639641571044922, 0.04630835342407227, 0.04636979293823242, 0.04633190536499023, 0.046432254791259765, 0.0463370246887207, 0.04643328094482422, 0.046339073181152345, 0.046394367218017575, 0.04638617706298828, 0.046478336334228515, 0.04636671829223633, 0.04645171356201172, 0.04632371139526367, 0.04646297454833984, 0.04636979293823242, 0.04650086212158203, 0.04644659042358398, 0.04645171356201172, 0.04655001449584961, 0.0464640007019043, 0.04640464019775391, 0.046531551361083986, 0.04647628784179687, 0.04664012908935547, 0.04642611312866211, 0.046486526489257815, 0.04647628784179687, 0.046486526489257815, 0.046427135467529294, 0.04649267196655273, 0.04644460678100586, 0.0465335693359375, 0.0464793586730957, 0.04654694366455078, 0.046465023040771485, 0.046545921325683595, 0.04652236938476562, 0.046543872833251954, 0.04663603210449219, 0.04664422225952149, 0.04656025695800781, 0.04656639862060547, 0.04651007843017578, 0.04655820846557617, 0.04648755264282227, 0.04653977584838867, 0.04652032089233398, 0.04658995056152344, 0.04660224151611328, 0.046524417877197265, 0.04651520156860352, 0.046548992156982424, 0.04656639862060547, 0.04652236938476562, 0.046599166870117184]",tokens/s,20.823829148306544,,,main,False,False,,,,,,,,,,,,,,,,,,,,,,, bfloat16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-160m,EleutherAI/pythia-160m,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1266.438144,1147.666432,0.0,501.219328,454.728704,s,15,0.17122963142395015,0.011415308761596677,0.0003781902788132034,0.011530847549438476,0.011591026878356933,0.011842886066436766,0.012296513004302978,"[0.012409919738769532, 0.010972096443176269, 0.011599871635437011, 0.011107359886169434, 0.011530847549438476, 0.01149078369140625, 0.011552672386169433, 0.011513504028320312, 0.011557920455932617, 0.011577759742736816, 0.011558527946472169, 0.011557279586791992, 0.010929375648498536, 0.010935040473937988, 0.01093667221069336]",tokens/s,22426.025028883483,kWh,1.3215693055086994e-07,7.241525706236313e-08,3.0173588642951474e-07,5.063080740427478e-07,tokens/kWh,505621010.4569373,MB,1266.438144,1147.666432,0.0,501.219328,469.3632,s,15,10.17911260986328,0.6786075073242187,0.014133086741383735,0.6854160766601562,0.68841748046875,0.6907819396972655,0.6936869055175781,"[0.6944131469726562, 0.682224365234375, 0.6849143676757813, 0.6858370361328125, 0.686234375, 0.6848311767578125, 0.68544970703125, 0.6854160766601562, 0.6872051391601562, 0.6860732421875, 0.6892257080078125, 0.6709505615234375, 0.6469058837890626, 0.6533924560546875, 0.6560393676757813]",tokens/s,92.83716923264224,kWh,7.859942394311293e-06,4.306892750311403e-06,1.3307830556770776e-05,2.5474665701393468e-05,tokens/kWh,2473045.210424641,,s,944,10.325950458526622,0.010938506841659546,0.0014521603428306398,0.010857520103454591,0.01097768964767456,0.011136107110977171,0.022157588100433303,"[0.011616255760192871, 0.01164185619354248, 0.011511808395385742, 0.011109439849853516, 0.01109398365020752, 0.011061216354370117, 0.010971136093139648, 0.011229184150695801, 0.011900927543640137, 0.011702272415161133, 0.011172863960266113, 0.01164902400970459, 0.01135206413269043, 0.011190272331237794, 0.011089920043945312, 0.011148287773132324, 0.011106304168701172, 0.010964991569519043, 0.010826848030090331, 0.010662816047668456, 0.01072537612915039, 0.010760191917419434, 0.010978303909301757, 0.010917920112609864, 0.01068233585357666, 0.01084928035736084, 0.01073971176147461, 0.01070905590057373, 0.010686400413513184, 0.010681344032287597, 0.011163647651672364, 0.012259327888488769, 0.011624447822570801, 0.01101414394378662, 0.011080703735351562, 0.01113702392578125, 0.011421695709228515, 0.011254783630371093, 0.011130911827087402, 0.0110796480178833, 0.011208767890930177, 0.010994624137878418, 0.010938367843627929, 0.01103667163848877, 0.010782719612121582, 0.010729472160339355, 0.010982399940490722, 0.011087871551513672, 0.011098112106323242, 0.01123635196685791, 0.01102847957611084, 0.011513855934143067, 0.010695679664611817, 0.010400768280029296, 0.010387455940246582, 0.01033625602722168, 0.010276864051818848, 0.010326016426086425, 0.01022873592376709, 0.010184703826904297, 0.010222592353820802, 0.010358783721923828, 0.021742591857910155, 0.01023692798614502, 0.010316800117492676, 0.010248191833496094, 0.010281984329223632, 0.01031372833251953, 0.011144191741943359, 0.010928159713745117, 0.010970080375671386, 0.01091481590270996, 0.010902527809143067, 0.010927103996276855, 0.010844160079956054, 0.010876928329467773, 0.010936320304870606, 0.010912768363952637, 0.010761216163635253, 0.01072332763671875, 0.010707967758178711, 0.01076633644104004, 0.01074176025390625, 0.010925056457519532, 0.010900480270385742, 0.010873855590820313, 0.010891263961791992, 0.010855423927307128, 0.01085647964477539, 0.010862560272216796, 0.010868736267089844, 0.010868736267089844, 0.010859519958496093, 0.010859519958496093, 0.010834943771362305, 0.010915840148925781, 0.010904576301574707, 0.010966015815734862, 0.010891263961791992, 0.010801152229309082, 0.010855423927307128, 0.010851327896118163, 0.0107991361618042, 0.010724384307861328, 0.010707903861999511, 0.01063526439666748, 0.010701824188232421, 0.010851327896118163, 0.010897407531738281, 0.010899456024169921, 0.01116262435913086, 0.01093945598602295, 0.010873791694641113, 0.01094041633605957, 0.011019264221191406, 0.010903552055358886, 0.010896384239196777, 0.010842111587524414, 0.010872832298278809, 0.010925056457519532, 0.010875904083251953, 0.010880000114440918, 0.010899456024169921, 0.011200511932373047, 0.010901568412780761, 0.02298361587524414, 0.010866687774658204, 0.010891263961791992, 0.01094041633605957, 0.010964991569519043, 0.010882047653198243, 0.010904640197753905, 0.01083084774017334, 0.01087276840209961, 0.010942463874816894, 0.010905599594116211, 0.010884096145629883, 0.010926079750061036, 0.010834943771362305, 0.010885120391845703, 0.010979328155517578, 0.010996735572814942, 0.010920960426330567, 0.010868736267089844, 0.01080832004547119, 0.01072537612915039, 0.010714112281799316, 0.010743807792663575, 0.010721280097961425, 0.010895359992980956, 0.01091481590270996, 0.010699775695800781, 0.01071513557434082, 0.010693632125854492, 0.010778623580932617, 0.010658816337585449, 0.010775551795959473, 0.010911744117736816, 0.010854399681091309, 0.010904576301574707, 0.010913791656494141, 0.010930208206176758, 0.010842080116271973, 0.010927103996276855, 0.010845184326171875, 0.01082470417022705, 0.010819583892822266, 0.010888192176818847, 0.010861568450927735, 0.010842111587524414, 0.010897407531738281, 0.010848256111145019, 0.010854399681091309, 0.010875904083251953, 0.010883071899414062, 0.010880000114440918, 0.010890239715576172, 0.010920960426330567, 0.010876928329467773, 0.010913791656494141, 0.010899456024169921, 0.010866687774658204, 0.010892288208007812, 0.010972160339355469, 0.010842111587524414, 0.010880000114440918, 0.010941439628601075, 0.0109486083984375, 0.02247065544128418, 0.010869759559631348, 0.010893312454223633, 0.010933247566223145, 0.010860544204711914, 0.010853376388549805, 0.010855423927307128, 0.010895359992980956, 0.01083289623260498, 0.0108472318649292, 0.010885120391845703, 0.01084928035736084, 0.010873855590820313, 0.01093734359741211, 0.010899456024169921, 0.011033599853515624, 0.010991616249084473, 0.010903552055358886, 0.010845184326171875, 0.010911744117736816, 0.01092403221130371, 0.010856448173522949, 0.010896384239196777, 0.01093942356109619, 0.010861536026000977, 0.010844160079956054, 0.01095372772216797, 0.01073459243774414, 0.010693632125854492, 0.010843135833740235, 0.010779680252075196, 0.010652640342712402, 0.01083289623260498, 0.010905599594116211, 0.01082470417022705, 0.010871808052062988, 0.010926079750061036, 0.01094041633605957, 0.010903552055358886, 0.010897407531738281, 0.010877951622009278, 0.010882047653198243, 0.010878975868225099, 0.010904576301574707, 0.010875904083251953, 0.010881024360656738, 0.010893312454223633, 0.010865663528442383, 0.010902527809143067, 0.010848256111145019, 0.010820608139038086, 0.010874879837036134, 0.010925056457519532, 0.010870783805847169, 0.010883071899414062, 0.010904576301574707, 0.010865663528442383, 0.010906623840332032, 0.010908703804016113, 0.010843135833740235, 0.010833888053894044, 0.011000831604003907, 0.011011072158813476, 0.02289151954650879, 0.010907648086547851, 0.010848256111145019, 0.01086569595336914, 0.0109967041015625, 0.010884096145629883, 0.010878975868225099, 0.01093222427368164, 0.010810367584228516, 0.010927103996276855, 0.010874879837036134, 0.010908672332763672, 0.01080832004547119, 0.010938367843627929, 0.010926079750061036, 0.010898431777954102, 0.010820608139038086, 0.010892288208007812, 0.01081446361541748, 0.010828800201416015, 0.010926079750061036, 0.011271167755126953, 0.010899456024169921, 0.010886143684387208, 0.010895359992980956, 0.010851327896118163, 0.010828800201416015, 0.010926079750061036, 0.010868736267089844, 0.010800127983093261, 0.010897407531738281, 0.010818559646606446, 0.010825728416442871, 0.010862591743469239, 0.010913791656494141, 0.010811391830444337, 0.010888192176818847, 0.01083289623260498, 0.010815487861633302, 0.010814528465270996, 0.010874815940856934, 0.010861568450927735, 0.010864640235900879, 0.010975232124328613, 0.010868736267089844, 0.010860544204711914, 0.01094758415222168, 0.01084108829498291, 0.010899456024169921, 0.010938367843627929, 0.010848256111145019, 0.010906623840332032, 0.010936384201049804, 0.010933183670043945, 0.010860544204711914, 0.010874879837036134, 0.010972160339355469, 0.010862591743469239, 0.010887167930603027, 0.010901503562927246, 0.010819583892822266, 0.010836992263793945, 0.01093939208984375, 0.02265190315246582, 0.010775551795959473, 0.010693632125854492, 0.010894335746765137, 0.010890239715576172, 0.010856448173522949, 0.010877951622009278, 0.0108472318649292, 0.01095372772216797, 0.010858495712280274, 0.010867744445800782, 0.010960864067077638, 0.010845184326171875, 0.010873855590820313, 0.010936320304870606, 0.010829888343811035, 0.010851263999938965, 0.010858495712280274, 0.010897407531738281, 0.010817536354064941, 0.010671104431152344, 0.01074998378753662, 0.010662879943847656, 0.010718208312988281, 0.010905599594116211, 0.011018239974975585, 0.010826751708984375, 0.010889216423034668, 0.010896384239196777, 0.010810367584228516, 0.01082265567779541, 0.010881024360656738, 0.010858495712280274, 0.01082470417022705, 0.010938367843627929, 0.010868736267089844, 0.01083187198638916, 0.010901503562927246, 0.01083801555633545, 0.010892288208007812, 0.010864640235900879, 0.01084108829498291, 0.01082470417022705, 0.010813440322875977, 0.010843135833740235, 0.01082470417022705, 0.01084006404876709, 0.010878975868225099, 0.010868800163269044, 0.010874815940856934, 0.010889216423034668, 0.010784768104553222, 0.010851327896118163, 0.011002880096435547, 0.010806271553039551, 0.010902527809143067, 0.01093734359741211, 0.010868736267089844, 0.01084928035736084, 0.011341823577880859, 0.011085824012756347, 0.010858495712280274, 0.010826751708984375, 0.022920192718505858, 0.010834943771362305, 0.010874879837036134, 0.010850303649902344, 0.010836992263793945, 0.010910719871520995, 0.010803199768066407, 0.01095792007446289, 0.010877887725830078, 0.01088304042816162, 0.010845184326171875, 0.01084928035736084, 0.010918911933898925, 0.010896384239196777, 0.010882047653198243, 0.010910719871520995, 0.010860544204711914, 0.010805279731750488, 0.010933216094970704, 0.010860544204711914, 0.01081651210784912, 0.010895359992980956, 0.010902527809143067, 0.010779680252075196, 0.010826720237731933, 0.010897407531738281, 0.010842111587524414, 0.010829824447631836, 0.010899456024169921, 0.010861568450927735, 0.010880000114440918, 0.01093222427368164, 0.010867712020874023, 0.010863615989685058, 0.010908672332763672, 0.010818559646606446, 0.010829824447631836, 0.010856448173522949, 0.010887167930603027, 0.010825728416442871, 0.010811391830444337, 0.01083084774017334, 0.010833919525146484, 0.010856448173522949, 0.010907648086547851, 0.010789888381958008, 0.010819583892822266, 0.01092300796508789, 0.010813440322875977, 0.010864640235900879, 0.010865663528442383, 0.010855423927307128, 0.010854432106018067, 0.010885151863098144, 0.010957759857177734, 0.010825728416442871, 0.010855423927307128, 0.010901503562927246, 0.011415552139282227, 0.010836992263793945, 0.010897407531738281, 0.01085747241973877, 0.01083187198638916, 0.02295091247558594, 0.010811391830444337, 0.01084006404876709, 0.010886143684387208, 0.010920960426330567, 0.010868736267089844, 0.010887167930603027, 0.010863615989685058, 0.010972160339355469, 0.010986495971679687, 0.01082470417022705, 0.01092915153503418, 0.010882047653198243, 0.010897407531738281, 0.010835968017578124, 0.010898431777954102, 0.01096396827697754, 0.01084006404876709, 0.010910719871520995, 0.010959872245788574, 0.010954751968383789, 0.0109619197845459, 0.010935296058654785, 0.010875904083251953, 0.01084620761871338, 0.010899456024169921, 0.010845184326171875, 0.010829824447631836, 0.01084006404876709, 0.010877951622009278, 0.01084006404876709, 0.010836992263793945, 0.010895359992980956, 0.010880000114440918, 0.010897407531738281, 0.011098112106323242, 0.01084006404876709, 0.010877951622009278, 0.01095680046081543, 0.010843135833740235, 0.010807295799255372, 0.010876928329467773, 0.010898431777954102, 0.010796031951904296, 0.010827775955200195, 0.010926079750061036, 0.010853376388549805, 0.010819583892822266, 0.010942463874816894, 0.010889216423034668, 0.010815487861633302, 0.010742783546447754, 0.010704895973205567, 0.010655743598937988, 0.010775551795959473, 0.010927103996276855, 0.010836992263793945, 0.01084928035736084, 0.010915840148925781, 0.010871808052062988, 0.010803199768066407, 0.010882047653198243, 0.010821632385253906, 0.02310758399963379, 0.010876928329467773, 0.010934271812438966, 0.010927103996276855, 0.010836992263793945, 0.010842111587524414, 0.010909695625305176, 0.010915840148925781, 0.010818559646606446, 0.0108472318649292, 0.010875904083251953, 0.010806271553039551, 0.010827775955200195, 0.011031583786010742, 0.010915807723999024, 0.010980352401733399, 0.011009023666381837, 0.011086848258972168, 0.010862591743469239, 0.010927103996276855, 0.010893312454223633, 0.010816543579101563, 0.010771424293518066, 0.010842111587524414, 0.010896384239196777, 0.010834943771362305, 0.010874879837036134, 0.010828800201416015, 0.01084108829498291, 0.010829824447631836, 0.011107328414916993, 0.011025407791137695, 0.010964991569519043, 0.011092991828918456, 0.010954751968383789, 0.010892288208007812, 0.010946559906005859, 0.010896384239196777, 0.01093939208984375, 0.010996735572814942, 0.010983424186706543, 0.010825728416442871, 0.010920960426330567, 0.010911744117736816, 0.010910719871520995, 0.010908672332763672, 0.01094041633605957, 0.010877951622009278, 0.010874879837036134, 0.010853376388549805, 0.010896384239196777, 0.010890239715576172, 0.010877951622009278, 0.010871808052062988, 0.010898431777954102, 0.01071718406677246, 0.010712063789367676, 0.01083903980255127, 0.010880000114440918, 0.010906623840332032, 0.010872832298278809, 0.010926079750061036, 0.010897407531738281, 0.022923263549804687, 0.010904640197753905, 0.010870719909667969, 0.010835968017578124, 0.010902560234069825, 0.010816479682922363, 0.01082367992401123, 0.010938367843627929, 0.010918911933898925, 0.0108472318649292, 0.01083289623260498, 0.010833919525146484, 0.010863615989685058, 0.010906623840332032, 0.010690560340881347, 0.010722304344177246, 0.010870783805847169, 0.010976256370544434, 0.010910719871520995, 0.010903552055358886, 0.010895359992980956, 0.01082470417022705, 0.01093734359741211, 0.010893312454223633, 0.010865663528442383, 0.010850303649902344, 0.010941439628601075, 0.010874879837036134, 0.010855423927307128, 0.010867712020874023, 0.010970144271850587, 0.010878944396972656, 0.010878975868225099, 0.01103667163848877, 0.010901503562927246, 0.010921983718872071, 0.010870783805847169, 0.010828800201416015, 0.01083084774017334, 0.010853376388549805, 0.010842111587524414, 0.010845184326171875, 0.010851327896118163, 0.010908672332763672, 0.011115519523620606, 0.011001855850219726, 0.010871808052062988, 0.010845184326171875, 0.010888192176818847, 0.01083801555633545, 0.01083903980255127, 0.010892288208007812, 0.010878975868225099, 0.010860544204711914, 0.010898431777954102, 0.010913791656494141, 0.01091481590270996, 0.010828800201416015, 0.010896384239196777, 0.010910719871520995, 0.010889216423034668, 0.01095580768585205, 0.010950624465942383, 0.022895616531372072, 0.010892288208007812, 0.010900480270385742, 0.010951680183410644, 0.010972160339355469, 0.010920960426330567, 0.011139072418212891, 0.010971136093139648, 0.010980352401733399, 0.010952704429626465, 0.010968064308166504, 0.010961983680725098, 0.011170751571655273, 0.01091481590270996, 0.010964991569519043, 0.010892288208007812, 0.010854399681091309, 0.010883071899414062, 0.010894335746765137, 0.010881024360656738, 0.010853376388549805, 0.010845184326171875, 0.010907648086547851, 0.01083187198638916, 0.010827775955200195, 0.010825728416442871, 0.011065343856811523, 0.010891263961791992, 0.010868736267089844, 0.010817536354064941, 0.010893312454223633, 0.010820608139038086, 0.011160575866699218, 0.010933247566223145, 0.010866687774658204, 0.010819583892822266, 0.010901503562927246, 0.01083903980255127, 0.010818559646606446, 0.010843135833740235, 0.010895423889160157, 0.010826687812805175, 0.010897407531738281, 0.010916864395141602, 0.010942463874816894, 0.010918911933898925, 0.010919936180114746, 0.010701824188232421, 0.010712063789367676, 0.010888192176818847, 0.012494848251342774, 0.011328512191772461, 0.010982399940490722, 0.010910719871520995, 0.010853376388549805, 0.010888192176818847, 0.010916864395141602, 0.010833984375, 0.010958784103393555, 0.010911744117736816, 0.010870783805847169, 0.01081446361541748, 0.010836992263793945, 0.02288844871520996, 0.01083187198638916, 0.010908672332763672, 0.01085756778717041, 0.010835871696472169, 0.010876928329467773, 0.010934271812438966, 0.01093222427368164, 0.011050016403198241, 0.010933216094970704, 0.010894335746765137, 0.010913791656494141, 0.010862591743469239, 0.01096396827697754, 0.010896384239196777, 0.010904640197753905, 0.010870719909667969, 0.010842111587524414, 0.010825728416442871, 0.010821632385253906, 0.010826751708984375, 0.011278335571289062, 0.01083289623260498, 0.010785792350769043, 0.010842111587524414, 0.010934271812438966, 0.01083801555633545, 0.010862591743469239, 0.01082265567779541, 0.011299839973449707, 0.010986495971679687, 0.010853376388549805, 0.010903552055358886, 0.010864640235900879, 0.010894335746765137, 0.010809344291687012, 0.010862591743469239, 0.010871808052062988, 0.010803199768066407, 0.010290176391601562, 0.010188799858093261, 0.010225664138793946, 0.01022060775756836, 0.01022150421142578, 0.010214400291442872, 0.010206208229064942, 0.010242048263549805, 0.010217472076416016, 0.010238975524902343, 0.010209280014038086, 0.010216511726379395, 0.010247103691101074, 0.010214400291442872, 0.010233856201171876, 0.01025331211090088, 0.010272768020629883, 0.01023795223236084, 0.010241024017333985, 0.01022054386138916, 0.010227711677551269, 0.010247167587280273, 0.010216447830200195, 0.010242048263549805, 0.02166783905029297, 0.010225664138793946, 0.010391551971435547, 0.010296319961547852, 0.010258432388305663, 0.010249216079711914, 0.01022976016998291, 0.010234880447387695, 0.01024614429473877, 0.010238975524902343, 0.010274815559387206, 0.010216447830200195, 0.010255359649658203, 0.01023078441619873, 0.010233856201171876, 0.010283007621765136, 0.010239999771118164, 0.010226688385009765, 0.010233856201171876, 0.010227711677551269, 0.01022054386138916, 0.010315775871276855, 0.010225664138793946, 0.010207232475280761, 0.010259455680847167, 0.01021132755279541, 0.010216447830200195, 0.010213376045227051, 0.010232831954956055, 0.010227711677551269, 0.010196991920471191, 0.01021235179901123, 0.010207232475280761, 0.010247167587280273, 0.01024614429473877, 0.01022976016998291, 0.010185728073120117, 0.010254336357116698, 0.010251263618469238, 0.010216447830200195, 0.010266624450683593, 0.010270719528198241, 0.010259455680847167, 0.010201087951660156, 0.010242048263549805, 0.011087871551513672, 0.010546175956726075, 0.01031065559387207, 0.01022976016998291, 0.010262528419494628, 0.010286080360412597, 0.01022054386138916, 0.010224639892578125, 0.010261504173278809, 0.010273792266845704, 0.010222592353820802, 0.01022976016998291, 0.010235903739929199, 0.010242048263549805, 0.010271743774414062, 0.010214400291442872, 0.01025331211090088, 0.010216480255126953, 0.021653472900390627, 0.010176511764526367, 0.010324992179870606, 0.010235903739929199, 0.010364928245544434, 0.010250240325927735, 0.010293248176574708, 0.010260479927062988, 0.010225664138793946, 0.010513407707214355, 0.01035264015197754, 0.010471424102783204, 0.010255359649658203, 0.01021132755279541, 0.010261568069458007, 0.01021945571899414, 0.010295295715332031, 0.010203136444091796, 0.010224639892578125, 0.010224639892578125, 0.010283007621765136, 0.010243071556091308, 0.010224639892578125, 0.010250240325927735, 0.01023078441619873, 0.010248191833496094, 0.01051955223083496, 0.010399744033813477, 0.011139072418212891, 0.010443807601928711, 0.010301407814025878, 0.010349568367004394, 0.01030246353149414, 0.010396672248840331, 0.010314751625061035, 0.010490880012512208, 0.010562560081481934, 0.010489855766296387, 0.01057692813873291, 0.010511327743530273, 0.01052569580078125, 0.010614784240722656, 0.010570752143859862, 0.010514431953430176, 0.010545151710510254, 0.010505215644836426, 0.010582015991210938, 0.010566656112670898, 0.010548224449157715, 0.010584063529968261, 0.010543104171752929, 0.01031167984008789, 0.010287103652954101, 0.010203136444091796, 0.010262528419494628, 0.010208255767822265, 0.010259455680847167, 0.010215423583984374, 0.010225664138793946, 0.010232831954956055, 0.010209280014038086, 0.01032192039489746, 0.010202112197875977, 0.02167705535888672, 0.010592255592346191, 0.010585087776184082, 0.010556415557861328, 0.010518527984619141, 0.010580991744995117, 0.01052467155456543, 0.010568703651428223, 0.010547200202941894, 0.01084006404876709, 0.010678272247314453, 0.010574912071228027, 0.010524607658386231, 0.010516480445861816, 0.010661888122558593, 0.010602496147155761, 0.01054207992553711, 0.010539008140563964, 0.010498047828674317, 0.010620927810668946, 0.010570752143859862, 0.010483712196350097, 0.01053286361694336, 0.010530816078186036, 0.010520575523376464, 0.010520575523376464, 0.010507264137268067, 0.01051136016845703, 0.01064038372039795, 0.010530816078186036, 0.010522624015808106, 0.01051750373840332, 0.01050931167602539, 0.010505215644836426, 0.010528767585754394, 0.010349568367004394, 0.01024614429473877, 0.010217472076416016, 0.010223615646362304, 0.010176511764526367, 0.010224639892578125, 0.010262528419494628, 0.01021951961517334, 0.01019596767425537, 0.010202112197875977, 0.010288127899169922, 0.01017958354949951, 0.010234880447387695, 0.010207232475280761, 0.010225664138793946, 0.010223615646362304, 0.010238975524902343, 0.010198016166687012, 0.01021951961517334, 0.010201087951660156, 0.01022976016998291, 0.01022054386138916, 0.010209280014038086, 0.010283007621765136, 0.010211392402648926, 0.010212287902832032, 0.01021132755279541, 0.010233856201171876]",tokens/s,91.42015582889958,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-6.7b,,cuda,0,42,,,,,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,217063f5c507ed7cc255df7e1f64c4333a0b4dfe,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,,MB,1287.274496,15125.184512,0.0,14478.737408,14241.165312,s,10,0.8262053833007812,0.08262053833007812,0.00121076900145203,0.08222753524780274,0.08265813827514648,0.08445479774475097,0.08589212532043457,"[0.08625145721435547, 0.08217072296142579, 0.08225759887695312, 0.08218064117431641, 0.08218204498291015, 0.08225004577636719, 0.08220502471923828, 0.08225421142578125, 0.08225888061523437, 0.08219475555419922]",tokens/s,3098.503170933744,kWh,9.743499693431924e-07,5.338999817169797e-07,4.857622283180546e-06,6.365872234240718e-06,tokens/kWh,40214442.04032695,MB,1287.274496,15125.184512,0.0,14478.737408,14438.898176,s,10,23.88951025390625,2.3889510253906248,0.007925318159070799,2.3869569091796876,2.3999682617187497,2.4018271484375,2.4033142578124997,"[2.383995361328125, 2.386448486328125, 2.39711279296875, 2.380974365234375, 2.37921484375, 2.388448974609375, 2.39955517578125, 2.40368603515625, 2.38746533203125, 2.38260888671875]",tokens/s,26.371407086379545,kWh,2.808398980857679e-05,1.53910336679044e-05,0.00011702191966481678,0.000160496943141298,tokens/kWh,392530.8405689458,,s,629,24.591448070526134,0.03909610186093184,0.009791683602354094,0.03785324859619141,0.038254598999023434,0.03832709121704102,0.12031136047363282,"[0.03765760040283203, 0.03772927856445313, 0.037774368286132814, 0.037732318878173826, 0.03763814544677734, 0.037731361389160153, 0.037762016296386716, 0.037705726623535156, 0.03785932922363281, 0.03764326477050781, 0.03775078582763672, 0.037783615112304686, 0.03797804641723633, 0.03842047882080078, 0.038012928009033206, 0.03769036865234375, 0.03768115234375, 0.03785113525390625, 0.03768012619018555, 0.03778355026245117, 0.03778662490844727, 0.037776382446289065, 0.037994495391845705, 0.03778355026245117, 0.037746688842773435, 0.03783782577514649, 0.03771187210083008, 0.037855232238769534, 0.037855232238769534, 0.03778867340087891, 0.03791872024536133, 0.037868545532226565, 0.03785625457763672, 0.03783270263671875, 0.03786342239379883, 0.03787059020996094, 0.03777027130126953, 0.037800926208496094, 0.03788390350341797, 0.03778665542602539, 0.037865440368652345, 0.03775078582763672, 0.03777740859985351, 0.03783475112915039, 0.03782963180541992, 0.0378869743347168, 0.03796582412719727, 0.03783987045288086, 0.03789619064331055, 0.037901313781738284, 0.03782656097412109, 0.03796889495849609, 0.03785728073120117, 0.037889057159423825, 0.03794326400756836, 0.03787980651855469, 0.03788083267211914, 0.0379156494140625, 0.037868545532226565, 0.03789004898071289, 0.03781635284423828, 0.03791049575805664, 0.12029440307617187, 0.037648384094238284, 0.037703678131103514, 0.03761151885986328, 0.0376258544921875, 0.03767603302001953, 0.03767500686645508, 0.03775494384765625, 0.037726142883300784, 0.03766067123413086, 0.037766143798828124, 0.037689342498779296, 0.037759998321533206, 0.03773235321044922, 0.03772415924072266, 0.03778252792358398, 0.037688385009765624, 0.037685184478759765, 0.03783987045288086, 0.037787647247314454, 0.03774156951904297, 0.03775692749023438, 0.037677120208740235, 0.0378223991394043, 0.03784806442260742, 0.037722110748291016, 0.03782656097412109, 0.03768627166748047, 0.03778252792358398, 0.037763072967529294, 0.03776204681396484, 0.03781631851196289, 0.03782451248168945, 0.03792793655395508, 0.03823820877075195, 0.03789209747314453, 0.03785932922363281, 0.03777433776855469, 0.03785830307006836, 0.03788185501098633, 0.03780505752563477, 0.03777433776855469, 0.03776716613769531, 0.03774569702148438, 0.03779888153076172, 0.037779457092285154, 0.03781222534179687, 0.03788288116455078, 0.037768192291259765, 0.03769139099121094, 0.03781324768066406, 0.03812659072875976, 0.03837849426269531, 0.03829248046875, 0.03834368133544922, 0.03848191833496094, 0.03822284698486328, 0.03826073455810547, 0.03828121566772461, 0.038212608337402344, 0.03830886459350586, 0.03856793594360351, 0.03826278305053711, 0.120416259765625, 0.037631999969482424, 0.037736446380615234, 0.037596160888671876, 0.03758489608764649, 0.03760332870483398, 0.03761356735229492, 0.0377262077331543, 0.03767295837402344, 0.03763711929321289, 0.03764841461181641, 0.03760329437255859, 0.03767603302001953, 0.03769343948364258, 0.03768115234375, 0.03775795364379883, 0.03783065414428711, 0.03834265518188477, 0.038199295043945314, 0.03809280014038086, 0.03811328125, 0.0380682258605957, 0.038507553100585935, 0.038306785583496095, 0.03810406494140625, 0.038073345184326174, 0.038079486846923825, 0.03810406494140625, 0.038141952514648435, 0.038056961059570314, 0.038114303588867186, 0.03821158218383789, 0.0381030387878418, 0.038145023345947264, 0.03807743835449219, 0.03820851135253906, 0.0379791374206543, 0.03781222534179687, 0.037992481231689454, 0.03818288040161133, 0.038073345184326174, 0.03809894561767578, 0.038400001525878906, 0.03832115173339844, 0.03815321731567383, 0.03824844741821289, 0.03816550445556641, 0.03827814483642578, 0.038623233795166016, 0.038193153381347655, 0.03820236968994141, 0.03818086242675781, 0.038089729309082034, 0.038147071838378906, 0.03823820877075195, 0.03833139038085937, 0.03817779159545898, 0.0382371826171875, 0.0382105598449707, 0.038254623413085935, 0.03821360015869141, 0.03820032119750977, 0.038209537506103515, 0.12031795501708985, 0.03759718322753906, 0.037648384094238284, 0.03760537719726562, 0.037571582794189456, 0.03763711929321289, 0.03763302230834961, 0.03775795364379883, 0.03769139099121094, 0.03775283050537109, 0.03767091369628906, 0.03766579055786133, 0.0377149429321289, 0.03762483215332031, 0.03773747253417969, 0.03781017684936523, 0.037644287109375, 0.03790643310546875, 0.038356990814208985, 0.03805286407470703, 0.03795667266845703, 0.03774764633178711, 0.037748737335205076, 0.03784908676147461, 0.037711902618408205, 0.03766678237915039, 0.03775078582763672, 0.037736446380615234, 0.03770470428466797, 0.03770982360839844, 0.03777433776855469, 0.03778662490844727, 0.03783782577514649, 0.03773235321044922, 0.03782451248168945, 0.03779072189331055, 0.038171646118164065, 0.03774156951904297, 0.037787647247314454, 0.03781222534179687, 0.037695518493652345, 0.037827552795410155, 0.037797889709472655, 0.03779177474975586, 0.03779068756103516, 0.037719039916992186, 0.03781119918823242, 0.03790028762817383, 0.037822463989257815, 0.0377968635559082, 0.037862464904785155, 0.03777734375, 0.03782553482055664, 0.03779174423217774, 0.037823486328125, 0.037910526275634765, 0.03788083267211914, 0.03782457733154297, 0.03783571243286133, 0.03788185501098633, 0.03787366485595703, 0.03783168029785156, 0.03791155242919922, 0.12024217224121093, 0.03752140808105469, 0.03763926315307617, 0.03759811019897461, 0.037658687591552734, 0.037622718811035155, 0.03765862274169922, 0.03773440170288086, 0.03770470428466797, 0.037664768218994144, 0.03769651031494141, 0.037610496520996094, 0.03769651031494141, 0.037591041564941405, 0.03767295837402344, 0.03769651031494141, 0.03767398452758789, 0.03760950469970703, 0.03772003173828125, 0.03760639953613281, 0.03772927856445313, 0.037684223175048825, 0.03778252792358398, 0.03794636917114258, 0.037797889709472655, 0.03768627166748047, 0.03782758331298828, 0.037720062255859374, 0.03776204681396484, 0.03775897598266602, 0.03778047943115234, 0.03778355026245117, 0.037787647247314454, 0.03770374298095703, 0.03779065704345703, 0.037713920593261716, 0.03784908676147461, 0.03774259185791016, 0.03783782577514649, 0.03783168029785156, 0.037833728790283204, 0.03767705535888672, 0.03780198287963867, 0.03773651123046875, 0.037854145050048825, 0.037754878997802735, 0.037850112915039064, 0.03785017776489258, 0.03784185409545898, 0.03775897598266602, 0.03786751937866211, 0.03782860946655273, 0.03783993530273438, 0.03777017593383789, 0.0379791374206543, 0.03790643310546875, 0.03791360092163086, 0.03786342239379883, 0.03786137771606445, 0.037795841217041014, 0.037920768737792966, 0.03784089660644531, 0.03794432067871094, 0.12033740997314453, 0.037531646728515625, 0.03764223861694336, 0.03759001541137695, 0.03765555191040039, 0.03767295837402344, 0.03766681671142578, 0.037776382446289065, 0.03765248107910156, 0.03767295837402344, 0.03768012619018555, 0.038027263641357424, 0.037698558807373043, 0.037718017578125, 0.037700607299804685, 0.037716991424560545, 0.037722110748291016, 0.03765964889526367, 0.037700607299804685, 0.03765043258666992, 0.03768217468261719, 0.03769548797607422, 0.03770982360839844, 0.037917697906494144, 0.037789695739746096, 0.03769753646850586, 0.037935104370117184, 0.03805081558227539, 0.03806719970703125, 0.03806105422973633, 0.03812351989746094, 0.03814604949951172, 0.03818905639648437, 0.03809075164794922, 0.038160446166992185, 0.03812243270874023, 0.038168575286865236, 0.03829248046875, 0.038269950866699216, 0.037975040435791016, 0.03779481506347656, 0.037986305236816405, 0.03818188858032227, 0.038193153381347655, 0.03830374526977539, 0.03806003189086914, 0.03784908676147461, 0.03794028854370117, 0.03785004806518555, 0.037784576416015625, 0.03778972625732422, 0.037763038635253904, 0.0378787841796875, 0.03785420989990235, 0.038035457611083984, 0.03862428665161133, 0.0383037109375, 0.038209537506103515, 0.03810713577270508, 0.03789516830444336, 0.03792486572265625, 0.03785932922363281, 0.03792697525024414, 0.12031890869140625, 0.03758899307250976, 0.0375838737487793, 0.03757363128662109, 0.0377149429321289, 0.03777542495727539, 0.03768108749389648, 0.03780198287963867, 0.03770982360839844, 0.037684223175048825, 0.03771289443969727, 0.037700607299804685, 0.03769241714477539, 0.037781505584716796, 0.037991424560546876, 0.03811635208129883, 0.03810611343383789, 0.03807849502563477, 0.038070240020751954, 0.03826073455810547, 0.03833651351928711, 0.03822182464599609, 0.0381317138671875, 0.03828121566772461, 0.038042625427246096, 0.03846246337890625, 0.03809075164794922, 0.03808665466308594, 0.038073345184326174, 0.0381399040222168, 0.03804876708984375, 0.03825459289550781, 0.03811532974243164, 0.038133792877197266, 0.03803753662109375, 0.03813267135620117, 0.03806924819946289, 0.03812460708618164, 0.03808153533935547, 0.03820640182495117, 0.03804774475097656, 0.03806617736816406, 0.038117374420166016, 0.038182910919189454, 0.0381399040222168, 0.03828736114501953, 0.03825766372680664, 0.038300704956054685, 0.038214622497558595, 0.03823110580444336, 0.038174720764160154, 0.03845727920532226, 0.0383559684753418, 0.038279232025146485, 0.03825350570678711, 0.038368255615234374, 0.038214656829833986, 0.03828224182128906, 0.03824435043334961, 0.03822694396972656, 0.03827507019042969, 0.03840716934204102, 0.03835084915161133, 0.12041932678222657, 0.03776102447509765, 0.03801599884033203, 0.03795251083374023, 0.03797401428222656, 0.038029312133789066, 0.038040576934814455, 0.03810508728027344, 0.037921791076660154, 0.037981182098388674, 0.03805286407470703, 0.03806412887573242, 0.03846451187133789, 0.03798425674438476, 0.03807846450805664, 0.03814809417724609, 0.03802624130249024, 0.03804876708984375, 0.03813683319091797, 0.03818393707275391, 0.03808870315551758, 0.03827609634399414, 0.03832524871826172, 0.038307838439941407, 0.03811532974243164, 0.038056961059570314, 0.03817062377929688, 0.03809894561767578, 0.038059009552001956, 0.03813478469848633, 0.0381102066040039, 0.038245376586914064, 0.038204414367675785, 0.03817574310302734, 0.03817062377929688, 0.03806105422973633, 0.03815423965454102, 0.038161407470703124, 0.038553600311279294, 0.038302719116210936, 0.038145023345947264, 0.038191104888916014, 0.03815935897827148, 0.038130687713623046, 0.03828121566772461, 0.03811942291259766, 0.03814604949951172, 0.03822079849243164, 0.03817062377929688, 0.03815423965454102, 0.03828121566772461, 0.03815526580810547, 0.03818700790405274, 0.038196224212646485, 0.038215679168701173, 0.03828736114501953, 0.03822182464599609, 0.03827302551269531, 0.038263809204101565, 0.03830374526977539, 0.03823513412475586, 0.038247425079345705, 0.03824332809448242, 0.12043571472167969, 0.037596160888671876, 0.03760025787353516, 0.03772927856445313, 0.03765350341796875, 0.03773747253417969, 0.03774774551391601, 0.03776713562011719, 0.03772313690185547, 0.037617664337158206, 0.03774156951904297, 0.037582847595214845, 0.037792766571044925, 0.03766886520385742, 0.03776409530639648, 0.037776382446289065, 0.03773747253417969, 0.03772415924072266, 0.037779457092285154, 0.03768115234375, 0.03796582412719727, 0.038109184265136715, 0.03815731048583984, 0.038150142669677735, 0.038163455963134765, 0.03808358383178711, 0.038166526794433595, 0.038130687713623046, 0.0382371826171875, 0.03808051300048828, 0.03820140838623047, 0.03816032028198242, 0.03814604949951172, 0.038095870971679685, 0.0382033920288086, 0.03806719970703125, 0.038234111785888675, 0.03810815811157227, 0.03802828979492188, 0.03782451248168945, 0.03784396743774414, 0.03775897598266602, 0.03786547088623047, 0.037766143798828124, 0.03785318374633789, 0.037746688842773435, 0.03784294509887695, 0.03785324859619141, 0.037895103454589844, 0.037823486328125, 0.03787571334838867, 0.03780505752563477, 0.03786649703979492, 0.03780198287963867, 0.03785420989990235, 0.037901313781738284, 0.037928958892822266, 0.03783679962158203, 0.037917697906494144, 0.03789926528930664, 0.03789823913574219, 0.03783168029785156, 0.03795254516601562, 0.12032611083984375, 0.03749273681640625, 0.03768729782104492, 0.03760332870483398, 0.0377077751159668, 0.037571582794189456, 0.037648384094238284, 0.03781119918823242, 0.03775795364379883, 0.03766886520385742, 0.03766681671142578, 0.037628929138183595, 0.037713920593261716, 0.03765043258666992, 0.037738494873046875, 0.03777740859985351, 0.037719039916992186, 0.037751808166503906, 0.037738494873046875, 0.03761971282958984, 0.03770675277709961, 0.03771084976196289, 0.037736446380615234, 0.03785625457763672, 0.03772931289672852, 0.03771798324584961, 0.03778559875488281, 0.03774566268920899, 0.0377446403503418, 0.037733375549316404, 0.03778047943115234, 0.03779072189331055, 0.03781324768066406, 0.03775692749023438, 0.03778252792358398, 0.03771596908569336, 0.03785932922363281, 0.03779379272460937, 0.037789695739746096, 0.037907455444335936, 0.03832831954956055, 0.038424575805664066, 0.03815321731567383, 0.03777024078369141, 0.03784089660644531, 0.03783987045288086, 0.037956607818603515, 0.03789311981201172, 0.037820415496826174, 0.037804031372070314, 0.03785625457763672, 0.03780198287963867, 0.03786342239379883, 0.037817344665527344, 0.037852161407470705, 0.03788288116455078, 0.037923839569091795, 0.03825356674194336, 0.038114303588867186, 0.0379607048034668, 0.03799552154541016, 0.03803443145751953, 0.03790950393676758]",tokens/s,25.577997611042786,,,main,False,False,,,,,,,,,,,,,,,,,,,,,,, bfloat16-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-13b,facebook/opt-13b,cuda,0,42,,,True,,,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",opt,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-6.7b,,cuda,0,42,,,,,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,d35829e539df8480b726c647eeabf91e41eae047,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,,MB,1188.245504,14382.792704,0.0,13736.3456,13487.531008,s,10,0.7251225891113281,0.0725122589111328,0.0011519276516436144,0.0721160774230957,0.07271428680419922,0.07433507919311523,0.07563171310424804,"[0.07595587158203125, 0.07203961944580078, 0.07235411071777344, 0.07218275451660157, 0.07221212768554687, 0.0720672607421875, 0.07216489410400391, 0.07205974578857421, 0.07203350067138672, 0.07205270385742188]",tokens/s,3530.437526622086,kWh,8.531013457514018e-07,4.6746000959154755e-07,4.600003680000037e-06,5.9205650353429865e-06,tokens/kWh,43239116.27890252,MB,1188.573184,14382.792704,0.0,13736.3456,13661.259776,s,10,21.688784423828125,2.1688784423828125,0.0017743762365916294,2.1693094482421875,2.1706022705078127,2.1710042602539064,2.1713258520507814,"[2.170212890625, 2.167744873046875, 2.17008251953125, 2.17140625, 2.1680048828125, 2.170512939453125, 2.17001708984375, 2.165689453125, 2.16651171875, 2.168601806640625]",tokens/s,29.04727105442839,kWh,2.5559453352106564e-05,1.4007375430240984e-05,0.00011063022739299909,0.00015019705617534665,tokens/kWh,419448.96660591685,,s,629,22.303378452301015,0.035458471307314825,0.008606498420304457,0.03440435028076172,0.03462369384765625,0.03471974258422852,0.10675292083740234,"[0.03420979309082031, 0.03431935882568359, 0.03419136047363281, 0.034323455810546875, 0.03461017608642578, 0.0345425910949707, 0.03450163269042969, 0.03425689697265625, 0.0346879997253418, 0.03490304183959961, 0.03458047866821289, 0.03445452880859375, 0.034490367889404294, 0.034473983764648435, 0.034541568756103515, 0.034598911285400394, 0.03448115158081055, 0.034500606536865236, 0.0341923828125, 0.03441664123535156, 0.03428966522216797, 0.034293758392333985, 0.034418689727783204, 0.03435728073120117, 0.03423638534545898, 0.03431219100952149, 0.034301952362060545, 0.03432550430297852, 0.03427840042114258, 0.03441049575805664, 0.03431628799438476, 0.03431219100952149, 0.034320384979248046, 0.03451903915405274, 0.034587646484375, 0.03446476745605469, 0.03448320007324219, 0.03434188842773438, 0.034871295928955076, 0.03489689636230469, 0.03439616012573242, 0.03453440093994141, 0.034460670471191404, 0.03456719970703125, 0.03429475021362305, 0.034618366241455076, 0.034436161041259766, 0.03445344161987305, 0.034328575134277346, 0.03440435028076172, 0.03442073440551758, 0.0343900146484375, 0.03446476745605469, 0.034408447265625, 0.03440435028076172, 0.03451289749145508, 0.03434598541259765, 0.03449753570556641, 0.03434905624389648, 0.034557952880859374, 0.03438489532470703, 0.03449139022827148, 0.10669977569580077, 0.034229248046875, 0.03427635192871094, 0.034233345031738284, 0.034315265655517575, 0.03440435028076172, 0.03422822570800781, 0.03426611328125, 0.03429785537719727, 0.03426406478881836, 0.03435212707519531, 0.0344268798828125, 0.034348033905029295, 0.03422719955444336, 0.034260990142822266, 0.03437363052368164, 0.034372608184814454, 0.034721790313720705, 0.03434188842773438, 0.034388992309570314, 0.03448627090454102, 0.034362400054931644, 0.03438179016113281, 0.03437363052368164, 0.034374656677246096, 0.03439411163330078, 0.03429683303833008, 0.0342845458984375, 0.03431219100952149, 0.03420467376708984, 0.03433171081542969, 0.03441145706176758, 0.034288639068603514, 0.034219009399414066, 0.03433267211914062, 0.03425689697265625, 0.03434393692016602, 0.03440639877319336, 0.03459174346923828, 0.03462758255004883, 0.03438796615600586, 0.034372608184814454, 0.0343818244934082, 0.03482624053955078, 0.034543617248535156, 0.034530303955078126, 0.03443404769897461, 0.03444224166870117, 0.03436646270751953, 0.03455590438842773, 0.034484222412109376, 0.034429950714111326, 0.034462718963623046, 0.03441766357421875, 0.03449753570556641, 0.03469107055664063, 0.03464704132080078, 0.034586654663085935, 0.03443299102783203, 0.03443609619140625, 0.034423809051513675, 0.03468288040161133, 0.03446579360961914, 0.10744217681884766, 0.034310142517089845, 0.03446783828735352, 0.03440947341918945, 0.03449760055541992, 0.03435103988647461, 0.03426508712768555, 0.03425894546508789, 0.03429177474975586, 0.03420665740966797, 0.03419955062866211, 0.034149375915527344, 0.034298881530761716, 0.03424358367919922, 0.03424361419677734, 0.03426095962524414, 0.03437977600097656, 0.03428147125244141, 0.03446579360961914, 0.03429580688476563, 0.03462656021118164, 0.03503923034667969, 0.03473100662231445, 0.03474739074707031, 0.03449651336669922, 0.03450265502929688, 0.03454265594482422, 0.034509761810302735, 0.03430099105834961, 0.034675647735595706, 0.03442483139038086, 0.034385921478271485, 0.03446681594848633, 0.034513919830322266, 0.0345456657409668, 0.034560001373291016, 0.03430092620849609, 0.03426816177368164, 0.03445452880859375, 0.03434598541259765, 0.0343900146484375, 0.034351104736328124, 0.034377727508544925, 0.034369537353515625, 0.03440947341918945, 0.03434188842773438, 0.03449139022827148, 0.034710529327392575, 0.034597888946533206, 0.03457228851318359, 0.03456512069702149, 0.03438387298583984, 0.034411518096923825, 0.03448115158081055, 0.03442278289794922, 0.03449343872070312, 0.034479103088378905, 0.03433779144287109, 0.03449760055541992, 0.0343438720703125, 0.03448524856567383, 0.034505729675292966, 0.03469004821777344, 0.10690354919433594, 0.034181121826171876, 0.03421593475341797, 0.0341739501953125, 0.0342476806640625, 0.03423027038574219, 0.03422208023071289, 0.03424358367919922, 0.03428870391845703, 0.034237377166748045, 0.03433779144287109, 0.0344453125, 0.03448534393310547, 0.03440732955932617, 0.03457024002075195, 0.03455590438842773, 0.03450470352172851, 0.03446783828735352, 0.03448831939697266, 0.03445558547973633, 0.0344749755859375, 0.034552833557128904, 0.03450368118286133, 0.034522113800048826, 0.03448320007324219, 0.034511871337890625, 0.03450470352172851, 0.034482177734375, 0.03456204986572266, 0.03451289749145508, 0.03451903915405274, 0.03458150482177735, 0.03436032104492188, 0.03428659057617187, 0.034551807403564457, 0.03445248031616211, 0.03450982284545898, 0.034411518096923825, 0.03543961715698242, 0.0348671989440918, 0.034667518615722655, 0.03457331085205078, 0.03457436752319336, 0.034474014282226566, 0.0343807373046875, 0.034374656677246096, 0.03438387298583984, 0.03441971206665039, 0.034495487213134765, 0.034391040802001956, 0.03444121551513672, 0.03450470352172851, 0.03460710525512695, 0.03432755279541016, 0.03460812759399414, 0.03464396667480469, 0.03460710525512695, 0.034388992309570314, 0.034484222412109376, 0.03435520172119141, 0.03448425674438477, 0.03430089569091797, 0.03439616012573242, 0.1069834213256836, 0.034255870819091795, 0.0343818244934082, 0.03443814468383789, 0.03428659057617187, 0.03420979309082031, 0.03417292785644531, 0.03434393692016602, 0.034290687561035156, 0.03420467376708984, 0.03422515106201172, 0.034219009399414066, 0.034280479431152346, 0.03424457550048828, 0.034305057525634765, 0.03450467300415039, 0.034232383728027345, 0.03423839950561523, 0.034435073852539064, 0.03436544036865234, 0.03454880142211914, 0.03447187042236328, 0.034500606536865236, 0.03454054260253906, 0.03443404769897461, 0.034233345031738284, 0.03427020645141601, 0.034272254943847655, 0.034374656677246096, 0.0342927360534668, 0.03430604934692383, 0.03441561508178711, 0.034359294891357424, 0.03443711853027344, 0.03440947341918945, 0.034285568237304685, 0.03464704132080078, 0.034477054595947264, 0.03460300827026367, 0.034579456329345705, 0.034971649169921876, 0.03463372802734375, 0.03470336151123047, 0.0344453125, 0.03456921768188476, 0.034331649780273435, 0.03446169662475586, 0.03466649627685547, 0.034533374786376955, 0.03430604934692383, 0.03444838333129883, 0.034334720611572264, 0.034375679016113284, 0.034476032257080076, 0.034403358459472656, 0.03446780776977539, 0.03441971206665039, 0.034418689727783204, 0.034418689727783204, 0.03427942276000977, 0.03443814468383789, 0.03443302536010742, 0.03463782501220703, 0.10678374481201172, 0.03428966522216797, 0.03423231887817383, 0.03411455917358398, 0.03462348937988281, 0.03505868911743164, 0.03434086227416992, 0.03432550430297852, 0.03446169662475586, 0.03417190551757812, 0.03433881759643555, 0.03462451171875, 0.03439923095703125, 0.03423539352416992, 0.03427840042114258, 0.034277374267578126, 0.034305023193359374, 0.03430303955078125, 0.034270145416259765, 0.034200576782226565, 0.03432755279541016, 0.03424460983276367, 0.03428147125244141, 0.034342910766601564, 0.03434393692016602, 0.034310142517089845, 0.034523136138916014, 0.034477054595947264, 0.03455590438842773, 0.03432352066040039, 0.03450361633300781, 0.03452928161621094, 0.034590721130371094, 0.034462753295898436, 0.03460095977783203, 0.034369503021240234, 0.034356224060058595, 0.03446579360961914, 0.03438489532470703, 0.03445043182373047, 0.034541568756103515, 0.03445148849487305, 0.03433577728271484, 0.03437254333496094, 0.03438387298583984, 0.03434393692016602, 0.03438796615600586, 0.03441766357421875, 0.034590721130371094, 0.03496345520019531, 0.03461734390258789, 0.03465625762939453, 0.034786304473876956, 0.03452928161621094, 0.03464704132080078, 0.03445555114746094, 0.03449753570556641, 0.03448831939697266, 0.034716670989990234, 0.03457535934448242, 0.03458867263793945, 0.03456204986572266, 0.034609153747558595, 0.10681753540039063, 0.03440947341918945, 0.034353153228759765, 0.03425996780395508, 0.03448627090454102, 0.034207775115966794, 0.03421795272827149, 0.03425689697265625, 0.03425791931152344, 0.03428659057617187, 0.03438387298583984, 0.03441254425048828, 0.03426303863525391, 0.034184192657470705, 0.03433574295043945, 0.0342999038696289, 0.03426508712768555, 0.03417702484130859, 0.03438079833984375, 0.03442278289794922, 0.034476032257080076, 0.03443404769897461, 0.034282497406005856, 0.034423809051513675, 0.034321407318115234, 0.03422208023071289, 0.034405376434326174, 0.03430912017822266, 0.034358272552490236, 0.03432550430297852, 0.03446169662475586, 0.03441766357421875, 0.03434905624389648, 0.03451903915405274, 0.03460505676269531, 0.03421491241455078, 0.034522113800048826, 0.03439616012573242, 0.03458873748779297, 0.03447907257080078, 0.0343592643737793, 0.0343296012878418, 0.03453849411010742, 0.03445248031616211, 0.03442073440551758, 0.03454873657226563, 0.03458457565307617, 0.034699264526367186, 0.034601982116699216, 0.034358272552490236, 0.034597888946533206, 0.03484467315673828, 0.03479142379760742, 0.03462246322631836, 0.03464601516723633, 0.03461324691772461, 0.03464908981323242, 0.034759681701660154, 0.034631679534912106, 0.03459481430053711, 0.0347248649597168, 0.03444121551513672, 0.034609153747558595, 0.10678886413574219, 0.03424256134033203, 0.034212894439697265, 0.03421488189697266, 0.034162689208984375, 0.03419136047363281, 0.034318336486816405, 0.034299934387207034, 0.03422825622558594, 0.03414729690551758, 0.034276321411132814, 0.03429683303833008, 0.034477054595947264, 0.03433369445800781, 0.03431628799438476, 0.03452108764648437, 0.03461119842529297, 0.03444940948486328, 0.034334720611572264, 0.03433062362670898, 0.034402305603027344, 0.03428147125244141, 0.03428659057617187, 0.03450368118286133, 0.03431321716308594, 0.03421491241455078, 0.03437977600097656, 0.03425900650024414, 0.03428448104858398, 0.034334720611572264, 0.03436544036865234, 0.03433267211914062, 0.034476032257080076, 0.03435520172119141, 0.034361343383789066, 0.03432755279541016, 0.03435520172119141, 0.03432447814941406, 0.03438387298583984, 0.03441971206665039, 0.03445248031616211, 0.03428147125244141, 0.03431423950195313, 0.034356224060058595, 0.034365470886230466, 0.03432751846313477, 0.03450777435302734, 0.03462656021118164, 0.03443199920654297, 0.034293758392333985, 0.03455487823486328, 0.034374656677246096, 0.03443097686767578, 0.0343818244934082, 0.03443916702270508, 0.03469619369506836, 0.034716670989990234, 0.034425918579101565, 0.03452000045776367, 0.03436032104492188, 0.0344453125, 0.03435417556762695, 0.03442483139038086, 0.10674483489990234, 0.034162689208984375, 0.03439718246459961, 0.03475763320922851, 0.03462041473388672, 0.034200576782226565, 0.03426201629638672, 0.03427948760986328, 0.034311103820800784, 0.03417702484130859, 0.0343296012878418, 0.03426611328125, 0.03422822570800781, 0.03416883087158203, 0.03442483139038086, 0.03434188842773438, 0.03446681594848633, 0.0345241584777832, 0.03436032104492188, 0.034318336486816405, 0.034307071685791016, 0.034229248046875, 0.03439011383056641, 0.0343551025390625, 0.034253822326660154, 0.034255870819091795, 0.03427840042114258, 0.03420979309082031, 0.03428761672973633, 0.034595840454101565, 0.03462656021118164, 0.034479103088378905, 0.03430912017822266, 0.034405376434326174, 0.03462348937988281, 0.03444736099243164, 0.03433881759643555, 0.034323455810546875, 0.03452320098876953, 0.034376640319824216, 0.034374656677246096, 0.034288639068603514, 0.034328575134277346, 0.03437363052368164, 0.034362400054931644, 0.034394081115722654, 0.03445248031616211, 0.034375679016113284, 0.034531326293945314, 0.034338878631591796, 0.03454252624511719, 0.034477054595947264, 0.03448934555053711, 0.034388992309570314, 0.03443199920654297, 0.03444121551513672, 0.03439923095703125, 0.034377727508544925, 0.03453747177124023, 0.03433779144287109, 0.03443404769897461, 0.03438796615600586, 0.034533409118652346, 0.10675606536865234, 0.03417292785644531, 0.034283584594726565, 0.03416057586669922, 0.03428966522216797, 0.034522113800048826, 0.034925567626953126, 0.03478732681274414, 0.034490367889404294, 0.034407424926757815, 0.03430809783935547, 0.03421184158325195, 0.03434086227416992, 0.03420876693725586, 0.03429171371459961, 0.03427840042114258, 0.03431219100952149, 0.03421388626098633, 0.03424665451049805, 0.0343296012878418, 0.034313247680664065, 0.03422819137573242, 0.03426713562011719, 0.03434086227416992, 0.03426201629638672, 0.03422719955444336, 0.03427840042114258, 0.03436441421508789, 0.03439206314086914, 0.034301952362060545, 0.0343900146484375, 0.03433062362670898, 0.03437670516967774, 0.03425791931152344, 0.03453747177124023, 0.034784255981445314, 0.03486310577392578, 0.03429788970947266, 0.03442582321166992, 0.03439923095703125, 0.03456921768188476, 0.03443609619140625, 0.034587646484375, 0.03454771041870117, 0.03459686279296875, 0.034563072204589845, 0.03444838333129883, 0.03437977600097656, 0.03458150482177735, 0.034328575134277346, 0.034457599639892575, 0.03441459274291992, 0.03470745468139649, 0.03450982284545898, 0.03444224166870117, 0.03444326400756836, 0.03463577651977539, 0.034353153228759765, 0.0345846061706543, 0.03445756912231445, 0.03467161560058594, 0.03435417556762695, 0.03446579360961914]",tokens/s,28.20200542017465,,,main,False,False,,,,,,,,,,,,,,,,,,,,,,, bfloat16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,microsoft/rho-math-1b-v0.1,microsoft/rho-math-1b-v0.1,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1262.280704,3045.588992,0.0,2399.141888,2279.02464,s,10,0.26803974342346193,0.026803974342346194,0.0008872097979867859,0.02658950424194336,0.027117384910583495,0.028239908695220944,0.029137927722930907,"[0.029362432479858397, 0.026646303176879882, 0.025867424011230468, 0.026602527618408204, 0.026576480865478515, 0.026551040649414062, 0.02645132827758789, 0.026464672088623048, 0.026867935180664063, 0.026649599075317384]",tokens/s,9550.822453801526,kWh,3.082963249956568e-07,1.6892854569618973e-07,1.2088724948749882e-06,1.6860973655668348e-06,tokens/kWh,151829903.3187431,MB,1262.280704,3045.588992,0.0,2399.141888,2279.0272,s,10,15.70307958984375,1.570307958984375,0.01128738515553608,1.5718734741210938,1.580663903808594,1.580894451904297,1.5810788903808592,"[1.567904052734375, 1.56319921875, 1.5758428955078125, 1.567379150390625, 1.5806126708984376, 1.5433612060546875, 1.580206787109375, 1.563749267578125, 1.581125, 1.5796993408203126]",tokens/s,40.119519002340404,kWh,1.8042147945509187e-05,9.887089594834605e-06,3.866046495612481e-05,6.658970249646861e-05,tokens/kWh,946092.2280489393,,s,629,15.915335668563834,0.025302600426969545,0.0032474409372474896,0.025011199951171875,0.025325159072875976,0.025517874908447265,0.05186424896240235,"[0.02527846336364746, 0.02408243179321289, 0.024011775970458983, 0.023972864151000976, 0.023982080459594726, 0.023946239471435548, 0.024259584426879883, 0.023997440338134765, 0.023938047409057618, 0.023942144393920898, 0.024162303924560546, 0.024039424896240235, 0.024019968032836913, 0.02411520004272461, 0.024186880111694335, 0.024184831619262694, 0.024024063110351563, 0.024056831359863282, 0.023998464584350586, 0.024804351806640625, 0.024863744735717775, 0.024839168548583986, 0.025450496673583983, 0.025654272079467775, 0.025636863708496094, 0.024983552932739257, 0.024997888565063478, 0.024921087265014647, 0.025003007888793945, 0.02493235206604004, 0.025284608840942382, 0.026445823669433592, 0.02552524757385254, 0.025208831787109375, 0.025007104873657225, 0.02510438346862793, 0.025204736709594725, 0.02574131202697754, 0.02515558433532715, 0.025396223068237304, 0.025199615478515625, 0.024928255081176756, 0.025037824630737306, 0.025242624282836915, 0.024870912551879884, 0.02510233688354492, 0.025449472427368162, 0.02551603126525879, 0.024990751266479493, 0.02501321601867676, 0.025166847229003905, 0.025240575790405274, 0.02514739227294922, 0.025082880020141602, 0.02554982376098633, 0.02502348709106445, 0.02491494369506836, 0.02512179183959961, 0.025347072601318358, 0.024994815826416016, 0.02497331237792969, 0.024953855514526366, 0.0523581428527832, 0.02492518424987793, 0.024698879241943358, 0.024224767684936522, 0.024817663192749022, 0.024937471389770507, 0.024987680435180664, 0.025025503158569336, 0.024814592361450196, 0.025164800643920897, 0.024839168548583986, 0.025011199951171875, 0.02471731185913086, 0.024242176055908202, 0.024130559921264647, 0.024374271392822267, 0.024225791931152343, 0.02422684860229492, 0.02419708824157715, 0.024233983993530273, 0.02434764862060547, 0.024209407806396483, 0.024606719970703125, 0.025027584075927735, 0.02550067138671875, 0.024970239639282226, 0.02493337631225586, 0.025020416259765626, 0.025166847229003905, 0.024970239639282226, 0.02494259262084961, 0.024964096069335938, 0.02512076759338379, 0.025218048095703126, 0.025241600036621094, 0.026783775329589844, 0.02593276786804199, 0.02533171272277832, 0.02512384033203125, 0.024308736801147462, 0.024182783126831055, 0.023976959228515626, 0.024236032485961914, 0.024078336715698243, 0.02408243179321289, 0.024251392364501953, 0.02405171203613281, 0.024066047668457033, 0.024048639297485352, 0.024980480194091798, 0.025074687957763672, 0.025191423416137695, 0.0250849609375, 0.02521494483947754, 0.025069568634033205, 0.025251840591430662, 0.025076736450195314, 0.024937471389770507, 0.025033727645874023, 0.024986623764038086, 0.025056320190429686, 0.02503059196472168, 0.025047040939331053, 0.05121535873413086, 0.02489958381652832, 0.025034751892089844, 0.02515456008911133, 0.024936447143554686, 0.025058303833007813, 0.02492313575744629, 0.02508902359008789, 0.024851455688476562, 0.0248985595703125, 0.024830976486206056, 0.024938495635986328, 0.025091072082519532, 0.025020416259765626, 0.025209856033325196, 0.025207807540893554, 0.02503167915344238, 0.025000959396362304, 0.024904703140258787, 0.02492416000366211, 0.02490060806274414, 0.024854528427124024, 0.02508185577392578, 0.02503987121582031, 0.024840192794799806, 0.024913919448852538, 0.024919040679931642, 0.025133056640625, 0.02489958381652832, 0.024937471389770507, 0.024978431701660156, 0.025057279586791992, 0.024958976745605467, 0.024997888565063478, 0.02508185577392578, 0.025187328338623048, 0.025030656814575194, 0.024986623764038086, 0.025082880020141602, 0.02491494369506836, 0.025196544647216795, 0.024987648010253907, 0.025006080627441408, 0.02490982437133789, 0.02498252868652344, 0.02506547164916992, 0.025011199951171875, 0.025111551284790038, 0.025028608322143556, 0.02490572738647461, 0.025050111770629883, 0.025011199951171875, 0.024922111511230468, 0.024975360870361327, 0.02492620849609375, 0.025093120574951173, 0.024970239639282226, 0.025049087524414062, 0.025192447662353516, 0.025093120574951173, 0.025061376571655275, 0.025015296936035155, 0.025158655166625975, 0.05207040023803711, 0.02484223937988281, 0.025009151458740234, 0.024770559310913084, 0.02409984016418457, 0.02430668830871582, 0.02430975914001465, 0.024188928604125977, 0.02411110305786133, 0.02431692886352539, 0.024376352310180663, 0.024242143630981445, 0.024135679244995118, 0.024191999435424806, 0.02415718460083008, 0.024276992797851563, 0.024401920318603516, 0.025268224716186522, 0.025027584075927735, 0.024971263885498047, 0.025034751892089844, 0.02508188819885254, 0.025491424560546875, 0.026066944122314452, 0.025316383361816405, 0.02499171257019043, 0.02492313575744629, 0.02498252868652344, 0.024964096069335938, 0.024996864318847657, 0.02498252868652344, 0.025038848876953124, 0.02494054412841797, 0.024984575271606444, 0.025013248443603517, 0.02506547164916992, 0.024983552932739257, 0.025064512252807616, 0.024803264617919922, 0.02508799934387207, 0.02485043144226074, 0.02492313575744629, 0.024843263626098632, 0.024844287872314453, 0.024928255081176756, 0.024887327194213868, 0.025060319900512697, 0.025034751892089844, 0.024988672256469727, 0.024913919448852538, 0.0251013126373291, 0.025159679412841796, 0.02491801643371582, 0.024959999084472655, 0.024922111511230468, 0.02507161521911621, 0.025042943954467774, 0.025409536361694338, 0.025070592880249022, 0.02511769676208496, 0.02508902359008789, 0.024999935150146483, 0.02498150444030762, 0.052120609283447264, 0.024907743453979492, 0.02495078468322754, 0.0248985595703125, 0.024975391387939454, 0.025040864944458008, 0.02492620849609375, 0.02505936050415039, 0.024880096435546874, 0.024878080368041993, 0.02458527946472168, 0.024750015258789063, 0.025052160263061524, 0.025447423934936524, 0.025183231353759765, 0.02508185577392578, 0.025767936706542968, 0.0253439998626709, 0.025091072082519532, 0.02508697509765625, 0.0249487361907959, 0.02527539253234863, 0.025648128509521483, 0.025867263793945314, 0.025367551803588868, 0.025151487350463866, 0.02506342315673828, 0.024945663452148437, 0.024979455947875977, 0.025026559829711914, 0.02509619140625, 0.025057279586791992, 0.024816640853881834, 0.025000959396362304, 0.02513817596435547, 0.025078784942626952, 0.025143295288085937, 0.025221120834350585, 0.025166847229003905, 0.02517913627624512, 0.024930303573608398, 0.02488115119934082, 0.025034751892089844, 0.024977407455444335, 0.024930303573608398, 0.024992767333984374, 0.02506547164916992, 0.024975360870361327, 0.02507980728149414, 0.025185279846191407, 0.02502348709106445, 0.02535424041748047, 0.025074687957763672, 0.02494156837463379, 0.025169919967651368, 0.02510438346862793, 0.025042943954467774, 0.025078784942626952, 0.02514022445678711, 0.02502348709106445, 0.02495078468322754, 0.02503167915344238, 0.025042943954467774, 0.05189836883544922, 0.0249036808013916, 0.024863807678222657, 0.02485958480834961, 0.024105983734130858, 0.024183807373046876, 0.024022016525268555, 0.024174591064453126, 0.024414207458496092, 0.024164352416992187, 0.02405887985229492, 0.024061952590942383, 0.024351743698120116, 0.02408038330078125, 0.024187904357910156, 0.024135679244995118, 0.024163328170776367, 0.024035327911376952, 0.02406707191467285, 0.024069120407104492, 0.02405580711364746, 0.024071168899536134, 0.02409984016418457, 0.024062976837158204, 0.024188928604125977, 0.024030208587646484, 0.02405887985229492, 0.02386534309387207, 0.023981056213378905, 0.024089599609375, 0.024024063110351563, 0.02410393524169922, 0.024161279678344725, 0.023974912643432617, 0.024467456817626954, 0.024985599517822265, 0.025022464752197264, 0.02491801643371582, 0.02551910400390625, 0.025252864837646483, 0.0248668155670166, 0.025037824630737306, 0.024959999084472655, 0.024855552673339845, 0.025095232009887697, 0.024868799209594727, 0.024862720489501954, 0.024871936798095705, 0.02493337631225586, 0.02515558433532715, 0.025158655166625975, 0.024337408065795898, 0.0241530876159668, 0.024114175796508788, 0.024240127563476564, 0.024143871307373048, 0.025028608322143556, 0.024922111511230468, 0.024980480194091798, 0.024993791580200195, 0.02490163230895996, 0.024852479934692383, 0.024860671997070313, 0.051776512145996094, 0.026083391189575197, 0.02483705520629883, 0.02497331237792969, 0.02491801643371582, 0.025012224197387696, 0.02494259262084961, 0.024971263885498047, 0.02495692825317383, 0.025011199951171875, 0.02494156837463379, 0.02489958381652832, 0.02493337631225586, 0.02491596794128418, 0.02490982437133789, 0.02510233688354492, 0.02515660858154297, 0.02509619140625, 0.02509004783630371, 0.02487603187561035, 0.0251463680267334, 0.02513408088684082, 0.02487295913696289, 0.024979455947875977, 0.024990720748901366, 0.02508697509765625, 0.025340927124023437, 0.02496614456176758, 0.025057279586791992, 0.025027584075927735, 0.025028608322143556, 0.02511359977722168, 0.025034751892089844, 0.025161727905273438, 0.024970239639282226, 0.0243507194519043, 0.0243507194519043, 0.02431590461730957, 0.024653823852539062, 0.025184255599975586, 0.02519862365722656, 0.024940511703491212, 0.02506342315673828, 0.025074687957763672, 0.025407487869262696, 0.024989696502685548, 0.025020416259765626, 0.025586687088012695, 0.025421823501586914, 0.027659263610839844, 0.02534604835510254, 0.025019392013549805, 0.025203712463378908, 0.02502556800842285, 0.025029600143432616, 0.02508799934387207, 0.025186304092407227, 0.025024511337280272, 0.0249354248046875, 0.02533478355407715, 0.025038848876953124, 0.025000959396362304, 0.02493440055847168, 0.05200083160400391, 0.02500499153137207, 0.024988672256469727, 0.024877056121826172, 0.024807424545288087, 0.02491494369506836, 0.024809471130371095, 0.024913919448852538, 0.024787967681884765, 0.024930303573608398, 0.025018367767333984, 0.024962047576904296, 0.024837120056152344, 0.024896511077880858, 0.024830976486206056, 0.025092096328735353, 0.02507366371154785, 0.024912895202636717, 0.02487603187561035, 0.024919040679931642, 0.024896511077880858, 0.02490982437133789, 0.024832000732421877, 0.02508799934387207, 0.02503987121582031, 0.024912895202636717, 0.024870912551879884, 0.024954879760742187, 0.024739839553833007, 0.024065023422241212, 0.024066047668457033, 0.02427903938293457, 0.024172544479370117, 0.02404351997375488, 0.02411929512023926, 0.024010751724243166, 0.024163328170776367, 0.02409267234802246, 0.024155136108398437, 0.024177663803100585, 0.02413670349121094, 0.024152063369750978, 0.024061952590942383, 0.024022016525268555, 0.024181760787963868, 0.025453567504882812, 0.025448448181152345, 0.025057279586791992, 0.025185279846191407, 0.025261056900024413, 0.02515558433532715, 0.025127935409545898, 0.02507366371154785, 0.02529177665710449, 0.02549452781677246, 0.02553548812866211, 0.025379840850830077, 0.025164800643920897, 0.02532352066040039, 0.025196544647216795, 0.025177087783813477, 0.025166847229003905, 0.025124864578247072, 0.05255782318115235, 0.025229312896728515, 0.02508902359008789, 0.025656320571899413, 0.025281536102294923, 0.025242624282836915, 0.026050559997558592, 0.025387008666992186, 0.02508083152770996, 0.025227264404296876, 0.0251146240234375, 0.02509823989868164, 0.025157632827758788, 0.025208831787109375, 0.025290752410888673, 0.025226240158081056, 0.02516275215148926, 0.025351167678833008, 0.02549862480163574, 0.024993791580200195, 0.025001983642578125, 0.025025535583496093, 0.02498252868652344, 0.025069568634033205, 0.025054208755493163, 0.02505625534057617, 0.02508595275878906, 0.02511769676208496, 0.024995840072631836, 0.02508083152770996, 0.02510233688354492, 0.025029632568359376, 0.024968255996704103, 0.02517190361022949, 0.02510745620727539, 0.025036863327026367, 0.025021375656127928, 0.025236480712890624, 0.025203712463378908, 0.02513715171813965, 0.025225248336791992, 0.025107423782348634, 0.025290752410888673, 0.02506547164916992, 0.02511769676208496, 0.025264127731323242, 0.025033727645874023, 0.023986175537109376, 0.024232959747314452, 0.024037376403808593, 0.023983104705810547, 0.02505625534057617, 0.02510438346862793, 0.025068544387817384, 0.024814592361450196, 0.025033727645874023, 0.025059328079223633, 0.025029632568359376, 0.025571327209472656, 0.02512588882446289, 0.025124864578247072, 0.025083904266357423, 0.025265151977539063, 0.05209395217895508, 0.02475212860107422, 0.024958976745605467, 0.024977407455444335, 0.024992767333984374, 0.024928255081176756, 0.024954879760742187, 0.025021440505981447, 0.025052160263061524, 0.02492416000366211, 0.02503167915344238, 0.025116672515869142, 0.025134111404418947, 0.025076736450195314, 0.025121759414672852, 0.025250816345214845, 0.025054208755493163, 0.025012224197387696, 0.024810495376586913, 0.0251463680267334, 0.02472038459777832, 0.02509004783630371, 0.02550579261779785, 0.02527539253234863, 0.02488934326171875, 0.02507776069641113, 0.025366527557373047, 0.025151487350463866, 0.024987648010253907, 0.02570035171508789, 0.025272319793701172, 0.02491801643371582, 0.025188352584838865, 0.024176639556884767, 0.02426982307434082, 0.024205312728881836, 0.02405068778991699, 0.02403945541381836, 0.02415203285217285, 0.02573311996459961, 0.025563135147094726, 0.025259008407592775, 0.025394176483154295, 0.025249792098999024, 0.025178112030029298, 0.025214975357055663, 0.025247743606567383, 0.025359359741210938, 0.025126911163330077, 0.025214975357055663, 0.025193471908569336, 0.02509823989868164, 0.025426944732666015, 0.02524569511413574, 0.025322496414184572, 0.025362432479858397, 0.025199615478515625, 0.0251463680267334, 0.025264127731323242, 0.02531123161315918, 0.025229312896728515, 0.02523750305175781, 0.02529484748840332]",tokens/s,39.52162952129299,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-2.7b,EleutherAI/pythia-2.7b,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1279.287296,6797.39392,0.0,6150.946816,5989.292032,s,10,0.3887236557006836,0.03887236557006836,0.0014476602822727281,0.038411041259765626,0.03893620071411132,0.041074644851684564,0.04278540016174316,"[0.04321308898925781, 0.038460990905761716, 0.038429759979248045, 0.038342273712158204, 0.03834566497802734, 0.03831264114379883, 0.038414592742919924, 0.03835456085205078, 0.03840748977661133, 0.03844259262084961]",tokens/s,6585.655291251929,kWh,4.5476279133914886e-07,2.4918756812649705e-07,2.3326225925923065e-06,3.0365729520579523e-06,tokens/kWh,84305565.53120293,MB,1279.287296,6797.39392,0.0,6150.946816,6115.763712,s,10,16.95939392089844,1.6959393920898438,0.01658036035440633,1.700268310546875,1.7105757202148437,1.7186706115722656,1.7251465246582032,"[1.7267655029296876, 1.6696607666015626, 1.7003759765625, 1.69278466796875, 1.674107421875, 1.6788770751953126, 1.70877685546875, 1.7042218017578126, 1.7036632080078125, 1.70016064453125]",tokens/s,37.14755391250593,kWh,2.0089898790177118e-05,1.1009501097148761e-05,6.44621530654086e-05,9.556155295273448e-05,tokens/kWh,659260.9480840093,,s,629,17.276188758850115,0.027466118853497774,0.004633908549784592,0.02698851203918457,0.02728058891296387,0.027619328308105467,0.06496825531005859,"[0.027757568359375, 0.027805696487426756, 0.027840511322021484, 0.027918336868286132, 0.027936767578125, 0.027876352310180662, 0.027886592864990234, 0.028139551162719725, 0.02875491142272949, 0.027892736434936522, 0.026928127288818358, 0.0267509765625, 0.02698240089416504, 0.027613183975219727, 0.027894784927368164, 0.027662336349487306, 0.02750771141052246, 0.02796134376525879, 0.03131596755981445, 0.029711360931396483, 0.0285614070892334, 0.027473920822143554, 0.027185152053833008, 0.027123807907104492, 0.027047840118408203, 0.027011072158813477, 0.026999807357788085, 0.02692915153503418, 0.026806272506713868, 0.02655539131164551, 0.02688928031921387, 0.026999744415283203, 0.027049983978271484, 0.02696396827697754, 0.02695782470703125, 0.02696294403076172, 0.026995712280273438, 0.026854400634765626, 0.027240447998046875, 0.026995712280273438, 0.027012096405029298, 0.026842111587524413, 0.026944511413574217, 0.026937408447265623, 0.02707961654663086, 0.02693836784362793, 0.02699673652648926, 0.02690559959411621, 0.02698137664794922, 0.026934272766113283, 0.026894336700439454, 0.026945535659790038, 0.0269117431640625, 0.026960895538330077, 0.027157503128051756, 0.026959871292114256, 0.02691276741027832, 0.027456512451171877, 0.027404287338256835, 0.028463104248046874, 0.02730291175842285, 0.027026432037353516, 0.06489600372314454, 0.025996288299560546, 0.02595020866394043, 0.025981952667236328, 0.02590924835205078, 0.02591436767578125, 0.025839616775512695, 0.02598297691345215, 0.0259368953704834, 0.025993215560913087, 0.02614476776123047, 0.025964544296264647, 0.025977855682373048, 0.02595737648010254, 0.025948160171508788, 0.026062847137451172, 0.025965568542480468, 0.026005504608154296, 0.026010623931884767, 0.025994239807128908, 0.025935871124267578, 0.026876928329467774, 0.02693222427368164, 0.027124736785888674, 0.027088895797729492, 0.027019264221191407, 0.027026432037353516, 0.026951679229736326, 0.027002880096435547, 0.027088895797729492, 0.026909696578979493, 0.027230207443237304, 0.027076608657836915, 0.026960895538330077, 0.02696294403076172, 0.027002880096435547, 0.026960895538330077, 0.02695680046081543, 0.026948640823364258, 0.027046880722045898, 0.027125759124755858, 0.02707148742675781, 0.02635468864440918, 0.02588569641113281, 0.026051584243774413, 0.026060800552368164, 0.026020927429199217, 0.02623891258239746, 0.026065919876098635, 0.026041343688964845, 0.026073087692260744, 0.026056703567504884, 0.026022911071777344, 0.02688204765319824, 0.02697113609313965, 0.027099136352539063, 0.02690662384033203, 0.026892288208007813, 0.027035648345947266, 0.027023359298706053, 0.026884096145629883, 0.026928127288818358, 0.02692915153503418, 0.06595686340332031, 0.026848255157470705, 0.02690355110168457, 0.027062271118164064, 0.027304960250854493, 0.027249664306640626, 0.02709199905395508, 0.027053024291992186, 0.02736025619506836, 0.02719436836242676, 0.02696703910827637, 0.027068416595458986, 0.026876928329467774, 0.026070016860961914, 0.026005504608154296, 0.0261212158203125, 0.02615603256225586, 0.026998783111572267, 0.027281408309936524, 0.027034624099731445, 0.02705311965942383, 0.027073471069335938, 0.027032575607299804, 0.02713907241821289, 0.02696601676940918, 0.027000831604003905, 0.0265850887298584, 0.02697113609313965, 0.027052032470703126, 0.026936319351196288, 0.026845184326171875, 0.027288576126098633, 0.026901504516601563, 0.026901504516601563, 0.027543552398681642, 0.027382783889770508, 0.027064319610595702, 0.02697222328186035, 0.02688505554199219, 0.02705820846557617, 0.02699977684020996, 0.027274240493774415, 0.02700595283508301, 0.027044864654541017, 0.027072511672973632, 0.027026432037353516, 0.02698240089416504, 0.027081727981567383, 0.02712883186340332, 0.02756096076965332, 0.026704896926879884, 0.026908672332763672, 0.02693836784362793, 0.02698956871032715, 0.026874879837036132, 0.026993663787841796, 0.02699673652648926, 0.02710527992248535, 0.026969120025634764, 0.027040735244750976, 0.02698240089416504, 0.026942464828491212, 0.026953727722167968, 0.06499635314941406, 0.026013696670532226, 0.026012672424316406, 0.026011648178100585, 0.025959423065185547, 0.026016767501831056, 0.025971712112426756, 0.02608742332458496, 0.02591744041442871, 0.02608332824707031, 0.02695782470703125, 0.02701312065124512, 0.0273756160736084, 0.02715443229675293, 0.027220991134643553, 0.0271329288482666, 0.027045919418334962, 0.02695369529724121, 0.026945535659790038, 0.026999807357788085, 0.0269434871673584, 0.026984447479248046, 0.027085823059082033, 0.027030527114868166, 0.026842111587524413, 0.027010047912597656, 0.026959871292114256, 0.02695884895324707, 0.02690559959411621, 0.026885120391845704, 0.027057151794433593, 0.027222015380859374, 0.026969087600708007, 0.026955808639526367, 0.026957792282104494, 0.02694758415222168, 0.026993663787841796, 0.02694041633605957, 0.026885120391845704, 0.027072511672973632, 0.026987520217895508, 0.026918912887573244, 0.02691481590270996, 0.026952703475952147, 0.027241472244262696, 0.02697318458557129, 0.026908735275268554, 0.027116479873657225, 0.027019264221191407, 0.02715443229675293, 0.02705516815185547, 0.0272608642578125, 0.027087936401367186, 0.026987455368041993, 0.027001855850219726, 0.027142143249511717, 0.0269752311706543, 0.026951679229736326, 0.02697420883178711, 0.027038719177246092, 0.027018239974975586, 0.027067392349243165, 0.02691993522644043, 0.06595174407958984, 0.027057151794433593, 0.027023359298706053, 0.027166719436645507, 0.026986495971679687, 0.02707967948913574, 0.02693836784362793, 0.027037727355957032, 0.02694243240356445, 0.0271646728515625, 0.027600896835327147, 0.02740838432312012, 0.027089920043945313, 0.027032575607299804, 0.02694655990600586, 0.027088895797729492, 0.02698137664794922, 0.026933248519897462, 0.027059200286865235, 0.02700492858886719, 0.027124736785888674, 0.027060224533081056, 0.027140096664428712, 0.027157503128051756, 0.027017215728759765, 0.026070016860961914, 0.026064895629882814, 0.025956352233886718, 0.025959423065185547, 0.025968639373779297, 0.026014720916748047, 0.026077184677124023, 0.02592665672302246, 0.0259368953704834, 0.026001407623291017, 0.02598297691345215, 0.02595327949523926, 0.026063871383666993, 0.02593382453918457, 0.02611404800415039, 0.026009599685668947, 0.026070016860961914, 0.026034175872802736, 0.025990144729614258, 0.025981952667236328, 0.026011648178100585, 0.02607923126220703, 0.02615910339355469, 0.02596249580383301, 0.02595840072631836, 0.025903104782104492, 0.025989120483398437, 0.025960447311401368, 0.02597068786621094, 0.026014720916748047, 0.02726911926269531, 0.02705311965942383, 0.02698851203918457, 0.027003936767578125, 0.027001792907714844, 0.02695884895324707, 0.02697420883178711, 0.02713599967956543, 0.06481613159179687, 0.025976831436157227, 0.026065919876098635, 0.025956352233886718, 0.025976863861083985, 0.025912288665771485, 0.025742336273193358, 0.025738239288330078, 0.02571980857849121, 0.025683967590332032, 0.0259420166015625, 0.02596659278869629, 0.02596767997741699, 0.025823232650756835, 0.025933759689331055, 0.02612531280517578, 0.026023935317993165, 0.025999359130859375, 0.026004480361938476, 0.02614784049987793, 0.025992191314697266, 0.026029056549072265, 0.026171392440795898, 0.026100736618041992, 0.02592870330810547, 0.026853376388549805, 0.02718003273010254, 0.026991615295410155, 0.027104352951049803, 0.02730793571472168, 0.027288576126098633, 0.027250688552856447, 0.02710937690734863, 0.027083776473999024, 0.027062271118164064, 0.027378688812255858, 0.027240447998046875, 0.027054079055786134, 0.026983423233032225, 0.027189279556274416, 0.02697929573059082, 0.027081727981567383, 0.02687283134460449, 0.02730086326599121, 0.028120063781738282, 0.02719340705871582, 0.026993600845336915, 0.027061248779296877, 0.027297792434692384, 0.027084800720214845, 0.027064384460449217, 0.026728384017944334, 0.027010047912597656, 0.02698137664794922, 0.02695577621459961, 0.027065343856811523, 0.026885120391845704, 0.02692198371887207, 0.02696703910827637, 0.026927167892456055, 0.026924991607666017, 0.02699673652648926, 0.026983488082885743, 0.0659486083984375, 0.027015167236328123, 0.02710937690734863, 0.027196416854858397, 0.02693939208984375, 0.02695577621459961, 0.02694758415222168, 0.027455488204956056, 0.02853068733215332, 0.027623424530029295, 0.02710527992248535, 0.027217920303344727, 0.02715545654296875, 0.0269803524017334, 0.02695680046081543, 0.027021312713623048, 0.027044864654541017, 0.027027456283569336, 0.02714112091064453, 0.027085823059082033, 0.027046911239624022, 0.026953727722167968, 0.027108352661132814, 0.0271329288482666, 0.027089920043945313, 0.02710425567626953, 0.02698137664794922, 0.027035680770874024, 0.027056095123291015, 0.02707155227661133, 0.02708883285522461, 0.027195423126220704, 0.02705200004577637, 0.026961919784545898, 0.02696703910827637, 0.027030527114868166, 0.02693529510498047, 0.0269434871673584, 0.026990591049194337, 0.027045888900756834, 0.026895360946655275, 0.026999807357788085, 0.02698137664794922, 0.027048959732055664, 0.02694963264465332, 0.02733465576171875, 0.027691007614135742, 0.02839347267150879, 0.02731520080566406, 0.027185152053833008, 0.0269803524017334, 0.026992639541625976, 0.027035648345947266, 0.027030527114868166, 0.02698240089416504, 0.027072511672973632, 0.027027456283569336, 0.027001855850219726, 0.027053056716918947, 0.026984447479248046, 0.027047935485839843, 0.026951679229736326, 0.027015167236328123, 0.06589030456542969, 0.026983423233032225, 0.027019264221191407, 0.026968063354492186, 0.026977279663085937, 0.02730700874328613, 0.02697318458557129, 0.027048959732055664, 0.02712678337097168, 0.027051008224487305, 0.02731110382080078, 0.027174911499023437, 0.026953727722167968, 0.027000831604003905, 0.02689945602416992, 0.02705510330200195, 0.027011072158813477, 0.027107328414916993, 0.02698854446411133, 0.027014144897460936, 0.026961919784545898, 0.027059200286865235, 0.027106304168701172, 0.02715340805053711, 0.026968063354492186, 0.027051008224487305, 0.026736640930175783, 0.027594751358032226, 0.027266048431396486, 0.02698956871032715, 0.02697318458557129, 0.027159551620483398, 0.02696499252319336, 0.02697420883178711, 0.027280384063720704, 0.027048959732055664, 0.027031551361083983, 0.02696703910827637, 0.02696396827697754, 0.02707865524291992, 0.02705510330200195, 0.026945535659790038, 0.026918912887573244, 0.026953727722167968, 0.027000831604003905, 0.027018239974975586, 0.02693939208984375, 0.027100160598754884, 0.02695884895324707, 0.027007999420166014, 0.02700595283508301, 0.027022367477416993, 0.027075551986694337, 0.027001855850219726, 0.02714521598815918, 0.027258880615234377, 0.027001855850219726, 0.02711756706237793, 0.02711961555480957, 0.026942464828491212, 0.026952703475952147, 0.02694963264465332, 0.026950656890869142, 0.06597734069824218, 0.02693734359741211, 0.026918975830078126, 0.026946495056152344, 0.02690457534790039, 0.026993663787841796, 0.027066368103027344, 0.02732748794555664, 0.027100160598754884, 0.027014144897460936, 0.026986495971679687, 0.02709401512145996, 0.027072511672973632, 0.02720256042480469, 0.026878976821899415, 0.027148288726806642, 0.027073535919189453, 0.027064319610595702, 0.027024383544921874, 0.027012096405029298, 0.027732992172241212, 0.027082752227783204, 0.027009023666381835, 0.027107328414916993, 0.027035648345947266, 0.027151359558105468, 0.02692403221130371, 0.02736639976501465, 0.02715545654296875, 0.02713497543334961, 0.02694963264465332, 0.02708684730529785, 0.02692300796508789, 0.0269803524017334, 0.027059200286865235, 0.027048959732055664, 0.026961919784545898, 0.02698956871032715, 0.027046911239624022, 0.02731724739074707, 0.026680320739746095, 0.026995712280273438, 0.02692095947265625, 0.027060224533081056, 0.026994688034057617, 0.02700595283508301, 0.026944511413574217, 0.027204608917236327, 0.02691276741027832, 0.026984447479248046, 0.026936384201049806, 0.026974143981933593, 0.0269752311706543, 0.026978303909301758, 0.02693939208984375, 0.027051008224487305, 0.02686566352844238, 0.02696703910827637, 0.02697420883178711, 0.027019264221191407, 0.026950656890869142, 0.026977279663085937, 0.02696601676940918, 0.06687026977539062, 0.027146240234375, 0.026933248519897462, 0.02695475196838379, 0.02692403221130371, 0.026852352142333984, 0.0269486083984375, 0.027090944290161133, 0.026853376388549805, 0.027206655502319335, 0.027046943664550783, 0.02696291160583496, 0.026818559646606444, 0.027083776473999024, 0.02691584014892578, 0.027059200286865235, 0.027009023666381835, 0.027001855850219726, 0.02693836784362793, 0.026978303909301758, 0.02692403221130371, 0.02694144058227539, 0.027006975173950197, 0.027058176040649414, 0.02693734359741211, 0.02692198371887207, 0.026976255416870116, 0.026933248519897462, 0.026986495971679687, 0.027115520477294923, 0.026999807357788085, 0.027028480529785157, 0.02694764709472656, 0.026938304901123047, 0.02696499252319336, 0.026902528762817384, 0.0270960636138916, 0.027151359558105468, 0.02697113609313965, 0.027124799728393555, 0.02686867141723633, 0.026969087600708007, 0.02690355110168457, 0.02699673652648926, 0.026851327896118164, 0.026612735748291014, 0.026868736267089844, 0.027020288467407227, 0.026901504516601563, 0.027037696838378908, 0.026894336700439454, 0.0269752311706543, 0.02696396827697754, 0.027007999420166014, 0.026968063354492186, 0.027192319869995117, 0.026908672332763672, 0.027029535293579102, 0.02698031997680664, 0.027067392349243165, 0.027448320388793947, 0.02666700744628906, 0.0260250244140625]",tokens/s,36.408493145097246,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen2-beta-72B,Qwen/Qwen2-beta-72B,cuda,0,42,,,True,,,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",qwen2,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-0.5B,,cuda,0,42,,,,,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,217063f5c507ed7cc255df7e1f64c4333a0b4dfe,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,,MB,1276.850176,2227.699712,0.0,1581.252608,1426.269184,s,10,0.2340151023864746,0.02340151023864746,0.001731743654549342,0.02261070346832275,0.024222504997253414,0.026330564212799067,0.028017011585235597,"[0.028438623428344727, 0.023754047393798827, 0.022628543853759765, 0.022533279418945312, 0.02259286308288574, 0.022489376068115234, 0.023263328552246092, 0.02247862434387207, 0.02331657600402832, 0.022519840240478515]",tokens/s,10939.464905868232,kWh,2.694022354522074e-07,1.476201905299898e-07,7.213116375399033e-07,1.1383340635221005e-06,tokens/kWh,224890046.0800713,MB,1276.850176,2227.699712,0.0,1581.252608,1426.271744,s,10,13.46384130859375,1.346384130859375,0.03165139668584073,1.3318592529296875,1.39713642578125,1.4051128295898436,1.4114939526367187,"[1.4130892333984375, 1.3953638916015625, 1.357199951171875, 1.3206170654296876, 1.3152825927734375, 1.3277080078125, 1.3215374755859375, 1.3493245849609374, 1.3283558349609375, 1.3353626708984374]",tokens/s,46.79199535706658,kWh,1.5588696662216848e-05,8.542403064964462e-06,2.722993294565652e-05,5.136103267283782e-05,tokens/kWh,1226610.8510960182,,s,629,13.645661182403565,0.021694214916380867,0.002823342458080262,0.020990976333618162,0.022175539779663085,0.022793011474609375,0.043916859741210935,"[0.02307379150390625, 0.022808576583862306, 0.02289254379272461, 0.022944768905639647, 0.022986751556396484, 0.023622655868530275, 0.023863296508789062, 0.02272870445251465, 0.021751808166503905, 0.02318137550354004, 0.022915008544921876, 0.022987775802612305, 0.022785024642944338, 0.02305023956298828, 0.022798336029052735, 0.02291814422607422, 0.02308403205871582, 0.022915071487426757, 0.02163302421569824, 0.02274412727355957, 0.022627264022827148, 0.022748159408569335, 0.02290995216369629, 0.02229248046875, 0.02208665657043457, 0.022099967956542968, 0.02204569625854492, 0.02202726364135742, 0.022174720764160157, 0.02212761688232422, 0.022236160278320313, 0.02211020851135254, 0.0220579833984375, 0.022221824645996095, 0.023005184173583985, 0.02249830436706543, 0.022128639221191407, 0.02207027244567871, 0.02209587287902832, 0.022008832931518556, 0.02206719970703125, 0.022149120330810547, 0.02206105613708496, 0.02211020851135254, 0.022046720504760742, 0.022063104629516602, 0.022046720504760742, 0.02204876708984375, 0.022077472686767578, 0.022110176086425782, 0.022042623519897463, 0.022099967956542968, 0.02208153533935547, 0.021992448806762696, 0.02200377655029297, 0.02216032028198242, 0.022022144317626953, 0.022221824645996095, 0.022037504196166992, 0.02209587287902832, 0.022071296691894532, 0.022172672271728516, 0.04598169708251953, 0.021979135513305666, 0.021978111267089845, 0.02207334327697754, 0.02206924819946289, 0.022037504196166992, 0.02207334327697754, 0.02208665657043457, 0.02207744026184082, 0.022128639221191407, 0.021013504028320314, 0.021819391250610352, 0.02208768081665039, 0.022212608337402344, 0.022033407211303712, 0.02207334327697754, 0.02209280014038086, 0.022145023345947267, 0.022081600189208985, 0.022113216400146483, 0.02209587287902832, 0.022148128509521484, 0.022171615600585937, 0.02262838363647461, 0.023275487899780272, 0.022421503067016603, 0.022156288146972656, 0.022099967956542968, 0.022140928268432617, 0.022105119705200196, 0.02210915184020996, 0.022235136032104492, 0.02227712059020996, 0.02209382438659668, 0.022183935165405275, 0.022178815841674804, 0.022198272705078126, 0.022195199966430663, 0.02215936088562012, 0.022129663467407225, 0.02208153533935547, 0.02204569625854492, 0.022041599273681642, 0.022191104888916017, 0.02211737632751465, 0.02204876708984375, 0.022143999099731446, 0.02209587287902832, 0.022071296691894532, 0.022092832565307616, 0.022013919830322266, 0.022012928009033202, 0.023031808853149413, 0.022691839218139647, 0.02243891143798828, 0.022163455963134765, 0.02206003189086914, 0.022129663467407225, 0.022017023086547852, 0.02210406494140625, 0.02207846450805664, 0.022182912826538087, 0.022122495651245116, 0.04466175842285156, 0.022090751647949217, 0.0221214714050293, 0.022363136291503907, 0.02210201644897461, 0.022120447158813478, 0.022029312133789062, 0.022039552688598633, 0.02204569625854492, 0.02193715286254883, 0.021987327575683592, 0.021935104370117187, 0.021979135513305666, 0.022014976501464844, 0.022041599273681642, 0.022007808685302735, 0.022030336380004883, 0.02205388832092285, 0.022090751647949217, 0.022082559585571288, 0.022137887954711916, 0.02212348747253418, 0.02205183982849121, 0.022023231506347656, 0.02197190475463867, 0.022025215148925782, 0.022001663208007814, 0.021959680557250977, 0.022041599273681642, 0.021978143692016602, 0.022012895584106445, 0.022339584350585938, 0.022271999359130858, 0.021811199188232423, 0.021120000839233398, 0.02086092758178711, 0.020957183837890626, 0.0208353271484375, 0.020966400146484376, 0.020916223526000977, 0.020914176940917968, 0.02089369583129883, 0.020892671585083008, 0.02086809539794922, 0.020918272018432618, 0.020898815155029296, 0.020911136627197267, 0.020932607650756836, 0.0208885440826416, 0.020958208084106447, 0.020938751220703124, 0.020890623092651366, 0.020952064514160155, 0.020962303161621093, 0.020968448638916014, 0.020976640701293944, 0.021113855361938477, 0.020942848205566408, 0.020996095657348633, 0.020930559158325195, 0.020990976333618162, 0.020908031463623047, 0.02109644889831543, 0.043960319519042966, 0.020942848205566408, 0.02090598487854004, 0.020867071151733398, 0.02104319953918457, 0.020984832763671874, 0.020972543716430665, 0.020965375900268556, 0.020962303161621093, 0.020720640182495118, 0.020733951568603515, 0.02068992042541504, 0.020703231811523438, 0.020884479522705078, 0.020958208084106447, 0.020765695571899414, 0.02084659194946289, 0.02097977638244629, 0.020972480773925783, 0.02101759910583496, 0.020981760025024415, 0.020985855102539062, 0.020957183837890626, 0.020952064514160155, 0.02123980712890625, 0.02166988754272461, 0.02127872085571289, 0.020953088760375976, 0.020961280822753905, 0.020996095657348633, 0.020931583404541015, 0.02103398323059082, 0.020972543716430665, 0.02105753517150879, 0.020990976333618162, 0.02099718475341797, 0.02094380760192871, 0.020962303161621093, 0.02108518409729004, 0.021175296783447265, 0.020928512573242186, 0.020997119903564454, 0.021001216888427734, 0.020961280822753905, 0.021009408950805664, 0.020983808517456053, 0.02090291213989258, 0.02088140869140625, 0.020898815155029296, 0.020926464080810548, 0.020884479522705078, 0.02088243293762207, 0.02087833595275879, 0.020904960632324218, 0.02085683250427246, 0.02089779281616211, 0.020985855102539062, 0.02087424087524414, 0.020913152694702147, 0.020877344131469727, 0.020882400512695312, 0.0209039363861084, 0.020921344757080077, 0.044036094665527346, 0.02083635139465332, 0.020839424133300782, 0.020981760025024415, 0.02089472007751465, 0.02084966468811035, 0.020754432678222655, 0.020677631378173827, 0.02080460739135742, 0.02084867286682129, 0.020721632003784178, 0.020941823959350587, 0.020769792556762694, 0.020703231811523438, 0.020611072540283205, 0.020776960372924806, 0.020736000061035157, 0.020815872192382814, 0.02095414352416992, 0.02090390396118164, 0.020943872451782225, 0.020618240356445314, 0.020791296005249024, 0.020883455276489257, 0.020855808258056642, 0.020909055709838868, 0.020933631896972657, 0.0209685115814209, 0.020915136337280274, 0.020931583404541015, 0.02091007995605469, 0.020936704635620116, 0.02087116813659668, 0.02068070411682129, 0.020931583404541015, 0.020703231811523438, 0.020813823699951172, 0.020896768569946288, 0.020708351135253905, 0.02089574432373047, 0.020867071151733398, 0.0208353271484375, 0.020946943283081054, 0.020904960632324218, 0.020890623092651366, 0.020908031463623047, 0.020978687286376953, 0.021012479782104493, 0.02089574432373047, 0.020888576507568358, 0.020929536819458007, 0.02101043128967285, 0.02088755226135254, 0.020899839401245117, 0.020995071411132812, 0.020890623092651366, 0.02104115104675293, 0.02101759910583496, 0.02085478401184082, 0.02090291213989258, 0.02091007995605469, 0.020992000579833983, 0.020931583404541015, 0.043834369659423826, 0.020983808517456053, 0.020840448379516603, 0.022182912826538087, 0.0212541446685791, 0.020898815155029296, 0.02086092758178711, 0.02084147262573242, 0.020851776123046874, 0.020900863647460938, 0.02097043228149414, 0.02082713508605957, 0.02082918357849121, 0.020862016677856445, 0.020846527099609376, 0.020957183837890626, 0.02088960075378418, 0.020847616195678712, 0.020917312622070312, 0.02092844772338867, 0.020976640701293944, 0.02094598388671875, 0.02092844772338867, 0.0208035831451416, 0.02084659194946289, 0.020945920944213867, 0.020969472885131835, 0.022007808685302735, 0.021611520767211914, 0.020917247772216797, 0.020935680389404295, 0.020999168395996092, 0.020964351654052735, 0.020974592208862306, 0.021337152481079102, 0.021625791549682617, 0.020974592208862306, 0.02089472007751465, 0.020980735778808594, 0.02088140869140625, 0.02126848030090332, 0.020975616455078124, 0.020817920684814452, 0.020642816543579103, 0.02109337615966797, 0.020933631896972657, 0.020922367095947265, 0.020922367095947265, 0.020928512573242186, 0.021300224304199217, 0.021151744842529296, 0.020892671585083008, 0.020956159591674805, 0.02106777572631836, 0.021562368392944335, 0.021545984268188476, 0.021988351821899413, 0.021511167526245118, 0.021176319122314453, 0.021531648635864258, 0.021357568740844726, 0.02089779281616211, 0.020933631896972657, 0.04533871841430664, 0.020919263839721678, 0.020908992767333986, 0.020939775466918945, 0.0212807674407959, 0.02090598487854004, 0.020899839401245117, 0.020962303161621093, 0.020920320510864256, 0.021612543106079102, 0.020933631896972657, 0.020920320510864256, 0.020907007217407226, 0.021046272277832033, 0.02128281593322754, 0.020907007217407226, 0.020931583404541015, 0.021137407302856445, 0.02167807960510254, 0.020971519470214844, 0.020942848205566408, 0.02089369583129883, 0.020935680389404295, 0.021020671844482423, 0.020974592208862306, 0.020912128448486327, 0.020944896697998046, 0.020813823699951172, 0.02085683250427246, 0.020832256317138673, 0.020794368743896483, 0.02127257537841797, 0.021082111358642578, 0.02102272033691406, 0.020965408325195313, 0.02095510482788086, 0.020879392623901368, 0.020833248138427733, 0.021013504028320314, 0.02088652801513672, 0.020926464080810548, 0.020912160873413087, 0.020961248397827148, 0.02088652801513672, 0.020917247772216797, 0.020928512573242186, 0.02082508850097656, 0.02084764862060547, 0.020839391708374025, 0.02090188789367676, 0.020864000320434572, 0.02085990333557129, 0.02085478401184082, 0.020896768569946288, 0.02087936019897461, 0.020915199279785156, 0.021081087112426757, 0.020876287460327148, 0.02087936019897461, 0.020875263214111327, 0.02086092758178711, 0.020823040008544923, 0.020935680389404295, 0.04392345428466797, 0.02124083137512207, 0.021547008514404296, 0.021695487976074217, 0.021511167526245118, 0.02142207908630371, 0.021364736557006835, 0.020832256317138673, 0.020909055709838868, 0.02086604881286621, 0.021145599365234375, 0.02148249626159668, 0.02142617607116699, 0.02086604881286621, 0.020947967529296875, 0.020892671585083008, 0.02102681541442871, 0.020953088760375976, 0.021786624908447266, 0.021573631286621094, 0.021582847595214845, 0.020917280197143555, 0.020985824584960937, 0.020944896697998046, 0.021694496154785158, 0.023026655197143555, 0.02201907157897949, 0.02169343948364258, 0.02167296028137207, 0.02143129539489746, 0.02079539108276367, 0.020920320510864256, 0.021525503158569336, 0.021534719467163087, 0.021478464126586914, 0.022917055130004884, 0.021817344665527344, 0.021292032241821288, 0.02087116813659668, 0.021596160888671875, 0.021536767959594725, 0.02145792007446289, 0.02147123146057129, 0.020947967529296875, 0.02084454345703125, 0.0208855037689209, 0.02127769660949707, 0.021565439224243164, 0.021724159240722657, 0.021543935775756837, 0.021180416107177736, 0.02086502456665039, 0.02067148780822754, 0.021309440612792968, 0.02290073585510254, 0.021507135391235353, 0.022521791458129884, 0.021754880905151368, 0.02151628875732422, 0.020916223526000977, 0.021719039916992186, 0.02187161636352539, 0.021745664596557617, 0.04532633590698242, 0.020979711532592774, 0.020876287460327148, 0.021184511184692383, 0.020931583404541015, 0.02091007995605469, 0.020816896438598635, 0.020912128448486327, 0.020847616195678712, 0.020957183837890626, 0.020912128448486327, 0.020945920944213867, 0.02104422378540039, 0.02085068893432617, 0.02084249687194824, 0.02087014389038086, 0.02083945655822754, 0.021004255294799806, 0.020941823959350587, 0.02089369583129883, 0.020891647338867187, 0.021015552520751952, 0.020929536819458007, 0.02082918357849121, 0.020900863647460938, 0.020890623092651366, 0.020923391342163086, 0.021150720596313476, 0.020987903594970703, 0.02086092758178711, 0.020896768569946288, 0.02124595260620117, 0.021541919708251953, 0.021606367111206056, 0.021142559051513674, 0.02093257522583008, 0.0211015682220459, 0.020883455276489257, 0.020960256576538085, 0.020853759765625, 0.021374975204467773, 0.02110361671447754, 0.020900863647460938, 0.02085683250427246, 0.020921344757080077, 0.020855871200561524, 0.02332972717285156, 0.022330368041992187, 0.02102579116821289, 0.020882463455200194, 0.02093974494934082, 0.0210565128326416, 0.020932607650756836, 0.020935680389404295, 0.020958208084106447, 0.02122137641906738, 0.022742015838623047, 0.021292032241821288, 0.02104319953918457, 0.020831232070922853, 0.020956159591674805, 0.02090598487854004, 0.02089779281616211, 0.04389990234375, 0.020907007217407226, 0.02085683250427246, 0.02089574432373047, 0.020925439834594727, 0.02085273551940918, 0.02089779281616211, 0.02088755226135254, 0.020917247772216797, 0.02086809539794922, 0.020955135345458984, 0.02085171127319336, 0.02092748832702637, 0.02087321662902832, 0.02165555191040039, 0.021935104370117187, 0.021541887283325196, 0.021351423263549805, 0.02091110420227051, 0.02119478416442871, 0.021586912155151367, 0.020912128448486327, 0.02086911964416504, 0.020926464080810548, 0.020971519470214844, 0.020899839401245117, 0.020942848205566408, 0.021614591598510743, 0.02182252883911133, 0.02175686454772949, 0.021568511962890623, 0.021465087890625, 0.021542911529541017, 0.021608448028564452, 0.021584896087646483, 0.021384191513061524, 0.02083737564086914, 0.020901920318603516, 0.021010400772094727, 0.02088652801513672, 0.020929599761962892, 0.02150704002380371, 0.021629919052124025, 0.021352447509765626, 0.02103091239929199, 0.020978687286376953, 0.02103603172302246, 0.02087424087524414, 0.02171392059326172, 0.02165452766418457, 0.02123776054382324, 0.02091007995605469, 0.02090188789367676, 0.02090598487854004, 0.020890623092651366, 0.02087833595275879, 0.02164735984802246, 0.021362688064575194, 0.022535167694091796, 0.021362688064575194, 0.020936704635620116, 0.020944896697998046, 0.02146918487548828]",tokens/s,46.09523800950824,,,main,False,False,,,,,,,,,,,,,,,,,,,,,,, bfloat16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-1.8B,,cuda,0,42,,,,,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,217063f5c507ed7cc255df7e1f64c4333a0b4dfe,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,,MB,1277.599744,5222.432768,0.0,4575.985664,4408.404992,s,10,0.25400146865844725,0.025400146865844724,0.0011341561227013026,0.02571254348754883,0.026372268295288086,0.02708008651733398,0.027646341094970703,"[0.027787904739379882, 0.025948863983154297, 0.025637727737426758, 0.024192192077636718, 0.024207199096679687, 0.0257873592376709, 0.025839487075805665, 0.026214975357055664, 0.024190399169921876, 0.02419536018371582]",tokens/s,10078.681881333536,kWh,2.890672709785768e-07,1.5835740973770973e-07,1.3255090745330683e-06,1.7729337552493548e-06,tokens/kWh,144393437.8495686,MB,1277.599744,5222.432768,0.0,4575.985664,4408.407552,s,10,14.477070800781249,1.4477070800781249,0.0393811811920201,1.469327880859375,1.4762752685546874,1.4882212524414062,1.4977780395507814,"[1.500167236328125, 1.47362060546875, 1.4732127685546874, 1.4460653076171874, 1.3849686279296876, 1.4007744140625, 1.4701929931640625, 1.472905517578125, 1.3867005615234376, 1.4684627685546876]",tokens/s,43.51709048532127,kWh,1.7152702484570325e-05,9.398499110037754e-06,4.147702516746769e-05,6.802822676207577e-05,tokens/kWh,926086.1703824552,,s,629,14.677949451446525,0.02333537273679894,0.0030918058258084533,0.02332979202270508,0.023565312576293945,0.024161074829101563,0.0465539079284668,"[0.02447667121887207, 0.024321023941040038, 0.024489984512329102, 0.024466432571411133, 0.02455449676513672, 0.02447667121887207, 0.024414207458496092, 0.024225791931152343, 0.024343551635742186, 0.024534015655517577, 0.02448896026611328, 0.025266176223754884, 0.024423423767089843, 0.02433126449584961, 0.02464460754394531, 0.02484230422973633, 0.024510400772094727, 0.024212480545043946, 0.024246271133422852, 0.02432614326477051, 0.024137727737426756, 0.02411520004272461, 0.0240762882232666, 0.023802879333496094, 0.023525375366210938, 0.023545856475830077, 0.023785472869873047, 0.023442432403564452, 0.023411712646484374, 0.023390207290649414, 0.023533567428588868, 0.023368703842163087, 0.023343103408813477, 0.023393280029296876, 0.023318527221679687, 0.023384063720703126, 0.023366655349731445, 0.02333695983886719, 0.023515167236328124, 0.02356118392944336, 0.023359487533569336, 0.023322656631469728, 0.023494623184204103, 0.023362560272216795, 0.023364608764648437, 0.02345471954345703, 0.023444480895996093, 0.023385087966918947, 0.02328166389465332, 0.023339008331298827, 0.0233175048828125, 0.02329599952697754, 0.02326118469238281, 0.02332262420654297, 0.023351295471191406, 0.02327142333984375, 0.02330828857421875, 0.02329599952697754, 0.023344127655029297, 0.02331648063659668, 0.023401472091674806, 0.02328883171081543, 0.049775615692138675, 0.023232511520385742, 0.023362560272216795, 0.02326323127746582, 0.023357440948486328, 0.023326719284057617, 0.023739391326904297, 0.02349158477783203, 0.02331443214416504, 0.02325606346130371, 0.023326719284057617, 0.023348224639892577, 0.023294015884399413, 0.023399360656738283, 0.023000064849853515, 0.023372800827026367, 0.02332876777648926, 0.023459840774536132, 0.023368703842163087, 0.023368703842163087, 0.02327142333984375, 0.023347200393676756, 0.023371776580810546, 0.023367679595947266, 0.023267328262329103, 0.023342079162597656, 0.023376895904541017, 0.023392255783081056, 0.023330816268920897, 0.02331648063659668, 0.023311424255371093, 0.02346284866333008, 0.023378944396972655, 0.02342911911010742, 0.023407648086547852, 0.02336764717102051, 0.023372800827026367, 0.023384063720703126, 0.023370752334594725, 0.02352025604248047, 0.023463935852050782, 0.023369728088378908, 0.023355392456054686, 0.023436288833618164, 0.023401472091674806, 0.023582719802856447, 0.023446527481079102, 0.023497791290283204, 0.0233686408996582, 0.02371788787841797, 0.02332569694519043, 0.023311359405517578, 0.023310335159301757, 0.023382015228271484, 0.023443456649780273, 0.02346905517578125, 0.023377920150756838, 0.023430143356323242, 0.023335935592651368, 0.023427072525024413, 0.023382015228271484, 0.023638015747070314, 0.023356416702270507, 0.04945612716674805, 0.023371776580810546, 0.02288332748413086, 0.022931455612182617, 0.023343103408813477, 0.023410688400268553, 0.023391231536865235, 0.023394304275512694, 0.023285791397094725, 0.02334511947631836, 0.023353343963623048, 0.02345577621459961, 0.0233287353515625, 0.023373823165893554, 0.02330828857421875, 0.023167999267578124, 0.02294988822937012, 0.023584768295288085, 0.02352639961242676, 0.023370752334594725, 0.023351295471191406, 0.023391231536865235, 0.023330816268920897, 0.02350387191772461, 0.023357440948486328, 0.0232806396484375, 0.023353343963623048, 0.023665664672851562, 0.023292959213256834, 0.02336867141723633, 0.023363584518432616, 0.02346905517578125, 0.02348134422302246, 0.02332876777648926, 0.02332979202270508, 0.023380992889404296, 0.023434240341186522, 0.023362560272216795, 0.023406591415405274, 0.02354380798339844, 0.023455743789672853, 0.023426048278808592, 0.023359487533569336, 0.023416831970214845, 0.023427072525024413, 0.02351206398010254, 0.023417856216430662, 0.02353049659729004, 0.023412736892700195, 0.023426048278808592, 0.023446527481079102, 0.023378944396972655, 0.02352639961242676, 0.023415807723999024, 0.023421951293945312, 0.023446527481079102, 0.02329395294189453, 0.023426048278808592, 0.023406591415405274, 0.023362560272216795, 0.023357440948486328, 0.023335935592651368, 0.023455743789672853, 0.046565376281738284, 0.021884927749633788, 0.021916671752929686, 0.021942272186279296, 0.02190336036682129, 0.0218920955657959, 0.02230886459350586, 0.021970943450927736, 0.02189619255065918, 0.0219289608001709, 0.02200371170043945, 0.022008832931518556, 0.02182143974304199, 0.021789695739746092, 0.0218470401763916, 0.021946367263793946, 0.02185523223876953, 0.02183782386779785, 0.021854240417480467, 0.022307807922363282, 0.02395238494873047, 0.02511769676208496, 0.02366771125793457, 0.023417856216430662, 0.02329910469055176, 0.023410655975341796, 0.023536640167236327, 0.0234967041015625, 0.023414783477783203, 0.023388160705566406, 0.023240703582763672, 0.023409664154052736, 0.023385087966918947, 0.023340032577514647, 0.02327347183227539, 0.02330419158935547, 0.02332057571411133, 0.022975488662719725, 0.02330521583557129, 0.023426048278808592, 0.023577600479125976, 0.02333798408508301, 0.023370752334594725, 0.023376895904541017, 0.02339638328552246, 0.023255008697509766, 0.023418880462646483, 0.023599103927612306, 0.022979583740234375, 0.02332569694519043, 0.023723007202148438, 0.023369728088378908, 0.02333695983886719, 0.02326118469238281, 0.023065631866455077, 0.023078880310058593, 0.023218175888061524, 0.02335436820983887, 0.02333798408508301, 0.023550975799560548, 0.023391231536865235, 0.023400447845458985, 0.023398399353027344, 0.04666470336914062, 0.02182963180541992, 0.021924863815307616, 0.021833728790283204, 0.021873664855957032, 0.021828607559204103, 0.02191155242919922, 0.02188800048828125, 0.021737472534179687, 0.021612543106079102, 0.021813247680664064, 0.021980159759521483, 0.022039552688598633, 0.02167807960510254, 0.021634048461914062, 0.021769216537475586, 0.0218787841796875, 0.021970943450927736, 0.021805055618286134, 0.021801984786987305, 0.021866559982299805, 0.021868480682373046, 0.021864511489868163, 0.0219299201965332, 0.02184499168395996, 0.02180201530456543, 0.021610464096069336, 0.02186649513244629, 0.02186140823364258, 0.02192278480529785, 0.02242355155944824, 0.022611967086791994, 0.022054912567138672, 0.021972991943359374, 0.021813247680664064, 0.02186956787109375, 0.02186649513244629, 0.021843967437744142, 0.021840896606445313, 0.0219289608001709, 0.0218603515625, 0.021803007125854493, 0.021843967437744142, 0.021865472793579102, 0.02185830307006836, 0.02186137580871582, 0.021849088668823242, 0.02189004707336426, 0.02191974449157715, 0.02183782386779785, 0.021820415496826173, 0.021811199188232423, 0.02184601593017578, 0.02184499168395996, 0.021910528182983398, 0.02192793655395508, 0.021783552169799804, 0.02183782386779785, 0.02210406494140625, 0.022245376586914063, 0.024010751724243166, 0.023772159576416017, 0.02352332878112793, 0.04823859024047852, 0.022140928268432617, 0.021952512741088868, 0.021931007385253908, 0.021873664855957032, 0.021944320678710938, 0.02192793655395508, 0.02205183982849121, 0.02203651237487793, 0.02211734390258789, 0.021949472427368163, 0.021790687561035155, 0.02185625648498535, 0.02189619255065918, 0.021811199188232423, 0.021939199447631837, 0.021946367263793946, 0.021799936294555664, 0.021850112915039063, 0.02185420799255371, 0.02184499168395996, 0.02186342430114746, 0.021836799621582033, 0.021944320678710938, 0.0219238395690918, 0.0218787841796875, 0.02190745544433594, 0.02188390350341797, 0.02165862464904785, 0.021594112396240234, 0.021590015411376954, 0.021778432846069336, 0.0218787841796875, 0.02187059211730957, 0.0218920955657959, 0.021918720245361328, 0.021906431198120118, 0.021917695999145507, 0.021638143539428712, 0.02188083267211914, 0.02190540885925293, 0.021850112915039063, 0.02188902473449707, 0.021811199188232423, 0.021818368911743165, 0.021917695999145507, 0.02184806442260742, 0.022122495651245116, 0.025406463623046875, 0.024176639556884767, 0.023573503494262696, 0.022031423568725585, 0.021895103454589844, 0.02208768081665039, 0.02344960021972656, 0.023444480895996093, 0.023790592193603514, 0.023463935852050782, 0.023374847412109375, 0.023293983459472655, 0.023298015594482423, 0.023365631103515624, 0.02331340789794922, 0.04960768127441406, 0.02347110366821289, 0.02325299263000488, 0.022972415924072266, 0.02291097640991211, 0.023639039993286134, 0.02306764793395996, 0.023257087707519532, 0.023323648452758788, 0.02327039909362793, 0.02353561592102051, 0.02330009651184082, 0.023311359405517578, 0.02306662368774414, 0.023012351989746094, 0.02332569694519043, 0.02325606346130371, 0.02329497528076172, 0.02353049659729004, 0.023403520584106444, 0.02347417640686035, 0.02330419158935547, 0.02325503921508789, 0.023347200393676756, 0.02323865509033203, 0.023563264846801758, 0.02330316734313965, 0.023342079162597656, 0.02330316734313965, 0.023249919891357423, 0.023331840515136718, 0.023557119369506836, 0.02345062446594238, 0.02333286476135254, 0.023408639907836915, 0.023401472091674806, 0.023712799072265624, 0.0233338565826416, 0.023393280029296876, 0.023537664413452147, 0.023341056823730468, 0.02333286476135254, 0.023266304016113282, 0.02330828857421875, 0.023387136459350585, 0.023393280029296876, 0.02327654457092285, 0.023404544830322265, 0.023556095123291015, 0.02332569694519043, 0.02328678321838379, 0.02352332878112793, 0.023340032577514647, 0.023266304016113282, 0.02330009651184082, 0.02331340789794922, 0.023361568450927735, 0.02331644821166992, 0.023174144744873046, 0.022986751556396484, 0.02326937675476074, 0.023345151901245118, 0.023339008331298827, 0.050165760040283204, 0.023383039474487305, 0.023378944396972655, 0.023378944396972655, 0.023404544830322265, 0.023276575088500978, 0.023394271850585936, 0.02388479995727539, 0.02341996765136719, 0.023425983428955077, 0.023451648712158202, 0.023438335418701172, 0.023404544830322265, 0.023399423599243165, 0.023740415573120118, 0.023417856216430662, 0.023403520584106444, 0.023441408157348635, 0.023444480895996093, 0.02332467269897461, 0.02328883171081543, 0.02335436820983887, 0.02328883171081543, 0.023378944396972655, 0.02342300796508789, 0.023326688766479493, 0.023351295471191406, 0.02330419158935547, 0.023304224014282227, 0.02335331153869629, 0.02330009651184082, 0.023431167602539063, 0.023299072265625, 0.02327244758605957, 0.023376928329467774, 0.023258079528808595, 0.02328883171081543, 0.02330624008178711, 0.023213056564331053, 0.023398399353027344, 0.02325299263000488, 0.023318527221679687, 0.023224384307861327, 0.023303104400634767, 0.02329804801940918, 0.02329292869567871, 0.023352319717407227, 0.023337024688720703, 0.02335532760620117, 0.023352319717407227, 0.02330931282043457, 0.023388160705566406, 0.0232857608795166, 0.02333695983886719, 0.023387136459350585, 0.023377920150756838, 0.02327859115600586, 0.0237260799407959, 0.023356416702270507, 0.023431167602539063, 0.023385087966918947, 0.02332467269897461, 0.02338102340698242, 0.04647727966308594, 0.02190342330932617, 0.021847999572753907, 0.02185215950012207, 0.02184601593017578, 0.021879840850830078, 0.02191663932800293, 0.021952512741088868, 0.021849088668823242, 0.021779455184936524, 0.021884927749633788, 0.021772287368774415, 0.021669952392578126, 0.02179372787475586, 0.021574655532836915, 0.021803007125854493, 0.021801984786987305, 0.02185523223876953, 0.021818368911743165, 0.02181228828430176, 0.021816255569458008, 0.021786624908447266, 0.021769216537475586, 0.021960704803466798, 0.021794815063476563, 0.02163609504699707, 0.02163609504699707, 0.02165657615661621, 0.0218603515625, 0.021793792724609375, 0.021816320419311523, 0.021959680557250977, 0.021901311874389647, 0.021914623260498048, 0.02184294319152832, 0.021933055877685546, 0.021770240783691407, 0.021763071060180664, 0.022226943969726562, 0.021901311874389647, 0.021803007125854493, 0.021792768478393554, 0.02187571144104004, 0.021816320419311523, 0.021815296173095702, 0.02183065605163574, 0.0218603515625, 0.021608448028564452, 0.02171801567077637, 0.02186342430114746, 0.02187980842590332, 0.02183475112915039, 0.02188083267211914, 0.021850112915039063, 0.021857280731201172, 0.022622207641601562, 0.023740415573120118, 0.023199743270874023, 0.023341056823730468, 0.02297145652770996, 0.023253952026367188, 0.023359487533569336, 0.02332467269897461, 0.046524417877197265, 0.021865472793579102, 0.02186444854736328, 0.022347776412963868, 0.023578624725341796, 0.023421951293945312, 0.023407615661621094, 0.02346086311340332, 0.023472127914428712, 0.023517183303833008, 0.023386112213134767, 0.023340032577514647, 0.023165952682495116, 0.023371776580810546, 0.02330112075805664, 0.02349567985534668, 0.023341056823730468, 0.02332979202270508, 0.02333695983886719, 0.023961599349975587, 0.023372800827026367, 0.023352319717407227, 0.02353971290588379, 0.023423999786376954, 0.023411712646484374, 0.023373823165893554, 0.023416831970214845, 0.02333695983886719, 0.023384063720703126, 0.023319551467895508, 0.023377920150756838, 0.023401472091674806, 0.023443456649780273, 0.023398399353027344, 0.02332262420654297, 0.023348224639892577, 0.023180288314819338, 0.02289254379272461, 0.02342911911010742, 0.023622655868530275, 0.023385087966918947, 0.023380992889404296, 0.02290790367126465, 0.023373823165893554, 0.023331840515136718, 0.023384063720703126, 0.023384063720703126, 0.023433216094970705, 0.023366655349731445, 0.023374847412109375, 0.023327743530273438, 0.023344127655029297, 0.023349248886108398, 0.023369728088378908, 0.023315456390380858, 0.023380992889404296, 0.02332057571411133, 0.023335935592651368, 0.02330931282043457, 0.023560192108154295, 0.023612415313720703, 0.02391756820678711, 0.023368703842163087]",tokens/s,42.85339734141209,,,main,False,False,,,,,,,,,,,,,,,,,,,,,,, bfloat16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-70m,,cuda,0,42,,,,,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,217063f5c507ed7cc255df7e1f64c4333a0b4dfe,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,,MB,1277.698048,883.42528,0.0,236.978176,215.486464,s,29,0.17640089559555053,0.006082789503294847,0.00022680963954793804,0.006005599975585938,0.006212409496307373,0.006309433555603028,0.006930284843444824,"[0.007161632061004639, 0.006335391998291016, 0.006155327796936035, 0.0061970877647399905, 0.0061084160804748535, 0.006005599975585938, 0.0059482879638671875, 0.006001664161682129, 0.006270495891571045, 0.0059714560508728025, 0.006031680107116699, 0.005983071804046631, 0.006070847988128662, 0.005954751968383789, 0.006053376197814942, 0.006032544136047363, 0.005984543800354004, 0.0060167679786682126, 0.005988063812255859, 0.005991104125976562, 0.005993375778198242, 0.00590396785736084, 0.005985184192657471, 0.006040031909942627, 0.005987520217895508, 0.006197887897491455, 0.005983744144439697, 0.0059443840980529785, 0.006102687835693359]",tokens/s,42085.954127022356,kWh,7.357807153646984e-08,4.0316963893367914e-08,1.367569456809803e-07,2.5065198111081803e-07,tokens/kWh,1021336431.7548223,MB,1277.698048,883.42528,0.0,236.978176,220.639744,s,29,9.89432583618164,0.34118364952350483,0.005771339517882901,0.3394775390625,0.342862060546875,0.34968533935546875,0.3637021728515625,"[0.3675633239746094, 0.35377349853515627, 0.3392811279296875, 0.3392323913574219, 0.33871035766601565, 0.3394775390625, 0.33949673461914065, 0.3409992065429687, 0.339360595703125, 0.3426893005371094, 0.3401512145996094, 0.3389798278808594, 0.3390479431152344, 0.3402869262695313, 0.34069406127929686, 0.3401373291015625, 0.33826144409179687, 0.3396458740234375, 0.34009146118164063, 0.33914071655273437, 0.33709063720703125, 0.33751400756835936, 0.3421714172363281, 0.3382021179199219, 0.3412197265625, 0.33901348876953125, 0.3392912292480469, 0.3435531005859375, 0.3392492370605469]",tokens/s,184.65128703554655,kWh,3.9671648864970565e-06,2.173837858738458e-06,6.133211033252553e-06,1.2274213778488067e-05,tokens/kWh,5132711.645483522,,s,1826,10.058050559043883,0.005508242365303332,0.0008053850906751835,0.005362688064575195,0.005576703786849976,0.005743616104125977,0.011779352188110352,"[0.006313983917236328, 0.006158336162567139, 0.006238272190093994, 0.006118336200714111, 0.006149119853973388, 0.005810175895690918, 0.005876736164093017, 0.005772287845611572, 0.0057825279235839844, 0.005745664119720459, 0.00582144021987915, 0.005788671970367432, 0.005681151866912842, 0.005615615844726562, 0.0057118721008300784, 0.005739520072937012, 0.005747712135314941, 0.0060702719688415525, 0.005836800098419189, 0.006054912090301514, 0.005781504154205322, 0.0057415680885314945, 0.0056780800819396975, 0.0058091521263122555, 0.006086656093597412, 0.005795839786529541, 0.005851136207580566, 0.005817344188690185, 0.005777408123016357, 0.006204480171203613, 0.005838784217834472, 0.005856256008148194, 0.005846015930175781, 0.0056863040924072265, 0.00562172794342041, 0.0056442880630493165, 0.005667840003967285, 0.005728256225585937, 0.005602303981781006, 0.005728256225585937, 0.0057415680885314945, 0.0057272958755493165, 0.005800896167755127, 0.005884928226470947, 0.005942336082458496, 0.005877696037292481, 0.005898240089416504, 0.0057784318923950195, 0.005705728054046631, 0.005815296173095703, 0.0058757119178771975, 0.005727231979370117, 0.0058009600639343266, 0.005793791770935058, 0.005638144016265869, 0.00570470380783081, 0.005734399795532226, 0.005743616104125977, 0.005792768001556397, 0.0060067839622497555, 0.00570470380783081, 0.0056145920753479005, 0.01265561580657959, 0.005763072013854981, 0.005723135948181152, 0.0057190399169921875, 0.0056852478981018065, 0.005777408123016357, 0.005843967914581299, 0.005984255790710449, 0.005910528182983398, 0.005896192073822021, 0.005831679821014404, 0.005780479907989502, 0.00581324815750122, 0.00566374397277832, 0.005734399795532226, 0.005743616104125977, 0.005633024215698243, 0.005583871841430664, 0.0054988799095153805, 0.005482495784759522, 0.005686272144317627, 0.005700607776641845, 0.0055511040687561035, 0.00537497615814209, 0.005398528099060058, 0.005381120204925537, 0.0056514558792114256, 0.005586944103240967, 0.005690368175506591, 0.0057415680885314945, 0.00568012809753418, 0.005773312091827393, 0.005863423824310303, 0.005948416233062744, 0.0056780800819396975, 0.005535744190216065, 0.005591040134429932, 0.005526527881622314, 0.005665791988372803, 0.005612544059753418, 0.0055920639038085935, 0.0058122239112854005, 0.005549056053161621, 0.005542912006378174, 0.005465087890625, 0.005375999927520752, 0.005398528099060058, 0.0052930560111999515, 0.005303296089172363, 0.005338111877441406, 0.00566476821899414, 0.0053944320678710935, 0.005369855880737305, 0.0053647360801696775, 0.005395455837249756, 0.00535756778717041, 0.005338111877441406, 0.005541888236999512, 0.005661695957183838, 0.005550079822540284, 0.00553984022140503, 0.005502975940704345, 0.005534719944000244, 0.012116991996765136, 0.005483520030975342, 0.005332992076873779, 0.005361663818359375, 0.005352447986602784, 0.00535756778717041, 0.005378047943115235, 0.0053606400489807126, 0.005339136123657226, 0.005392384052276611, 0.005407743930816651, 0.005362688064575195, 0.0053647360801696775, 0.005336063861846924, 0.005354496002197265, 0.005352447986602784, 0.005345280170440674, 0.005341184139251709, 0.005378047943115235, 0.005360703945159912, 0.005350336074829102, 0.005368832111358642, 0.005343232154846191, 0.0053729920387268065, 0.005359551906585693, 0.005335135936737061, 0.005351327896118164, 0.005395455837249756, 0.005352447986602784, 0.005349376201629639, 0.0053678078651428224, 0.005354496002197265, 0.005337088108062744, 0.005384191989898681, 0.005352447986602784, 0.0053340158462524415, 0.0053647360801696775, 0.00535859203338623, 0.0053658242225646975, 0.005357503890991211, 0.005347328186035156, 0.00536575984954834, 0.005349376201629639, 0.005343232154846191, 0.005370880126953125, 0.005344255924224854, 0.005351424217224121, 0.00537497615814209, 0.005325823783874512, 0.005339136123657226, 0.00535756778717041, 0.005326848030090332, 0.0053340158462524415, 0.00537395191192627, 0.005342207908630371, 0.0053534722328186036, 0.005361663818359375, 0.005348351955413819, 0.005340159893035889, 0.005327871799468994, 0.005354496002197265, 0.005477375984191894, 0.006415359973907471, 0.012149760246276856, 0.0053606400489807126, 0.005527552127838135, 0.005488639831542969, 0.005429247856140137, 0.00537497615814209, 0.005479423999786377, 0.0053534722328186036, 0.005378047943115235, 0.005352447986602784, 0.005345280170440674, 0.005404672145843506, 0.005327871799468994, 0.005341184139251709, 0.0053678078651428224, 0.005344255924224854, 0.0053678078651428224, 0.005344255924224854, 0.005332992076873779, 0.005379072189331055, 0.00535968017578125, 0.005352384090423584, 0.0053944320678710935, 0.005368832111358642, 0.005341184139251709, 0.005372928142547607, 0.005435391902923584, 0.00538316822052002, 0.00537497615814209, 0.00535859203338623, 0.005355552196502685, 0.0053974719047546384, 0.005349376201629639, 0.005351424217224121, 0.005403647899627686, 0.005532671928405761, 0.005416959762573242, 0.00551526403427124, 0.005355519771575928, 0.005356544017791748, 0.00535859203338623, 0.005359615802764893, 0.005354496002197265, 0.005386240005493164, 0.005338111877441406, 0.005377024173736572, 0.005349376201629639, 0.005352447986602784, 0.005370880126953125, 0.005361728191375733, 0.005322688102722168, 0.005368832111358642, 0.005345280170440674, 0.005327871799468994, 0.005375999927520752, 0.005341248035430908, 0.005380032062530518, 0.005387263774871826, 0.0053821439743042, 0.005340159893035889, 0.00537395191192627, 0.0053606400489807126, 0.005347328186035156, 0.011882495880126954, 0.005362688064575195, 0.005336063861846924, 0.005328896045684814, 0.005351424217224121, 0.0053340158462524415, 0.005336063861846924, 0.00538316822052002, 0.00536678409576416, 0.0053340158462524415, 0.005375999927520752, 0.005339136123657226, 0.0052930560111999515, 0.005287936210632324, 0.005313536167144775, 0.005288959980010987, 0.005288959980010987, 0.00531763219833374, 0.005287936210632324, 0.005340159893035889, 0.00532480001449585, 0.005345280170440674, 0.005371903896331787, 0.005341184139251709, 0.005347328186035156, 0.00536678409576416, 0.0053606400489807126, 0.005338111877441406, 0.005557248115539551, 0.005486591815948487, 0.005342207908630371, 0.005379072189331055, 0.005361695766448974, 0.005337056159973145, 0.005378047943115235, 0.005686272144317627, 0.005356607913970947, 0.00538105583190918, 0.005341184139251709, 0.005380095958709717, 0.005340159893035889, 0.005348351955413819, 0.00535859203338623, 0.005349376201629639, 0.005328896045684814, 0.0053534722328186036, 0.005349376201629639, 0.0053534722328186036, 0.005375040054321289, 0.00535753583908081, 0.005344223976135254, 0.0053534722328186036, 0.005339136123657226, 0.005351424217224121, 0.005426176071166992, 0.005348351955413819, 0.005405695915222168, 0.005618688106536865, 0.005480447769165039, 0.005538815975189209, 0.005420032024383545, 0.005370880126953125, 0.00535756778717041, 0.011831295967102052, 0.00538316822052002, 0.005391456127166748, 0.005357471942901612, 0.005374015808105469, 0.005349311828613281, 0.005359615802764893, 0.0053688640594482424, 0.005364704132080078, 0.005368832111358642, 0.005331967830657959, 0.005341184139251709, 0.00536678409576416, 0.005343232154846191, 0.005377024173736572, 0.005362720012664795, 0.0053339838981628414, 0.005360703945159912, 0.005383103847503662, 0.005329919815063477, 0.005347328186035156, 0.00570470380783081, 0.005509119987487793, 0.005478400230407715, 0.005463039875030518, 0.005436416149139404, 0.005406720161437988, 0.005381120204925537, 0.0053002238273620605, 0.00531763219833374, 0.005339136123657226, 0.005305344104766845, 0.005315584182739258, 0.005403647899627686, 0.005352447986602784, 0.00537497615814209, 0.00538316822052002, 0.00535859203338623, 0.00539958381652832, 0.005403615951538086, 0.005355519771575928, 0.005341184139251709, 0.005310463905334473, 0.005305344104766845, 0.005327871799468994, 0.005322751998901368, 0.0053340158462524415, 0.005396480083465576, 0.005347328186035156, 0.005548064231872559, 0.005544928073883057, 0.005512191772460938, 0.00536575984954834, 0.005387263774871826, 0.005349376201629639, 0.005339136123657226, 0.005366847991943359, 0.005347328186035156, 0.005440447807312012, 0.005534719944000244, 0.0053821439743042, 0.005389311790466309, 0.005349376201629639, 0.01174732780456543, 0.005354559898376464, 0.005354432106018067, 0.0053637118339538575, 0.00536678409576416, 0.0053678078651428224, 0.005354496002197265, 0.005337088108062744, 0.005359615802764893, 0.00535756778717041, 0.005343232154846191, 0.005371903896331787, 0.005368896007537842, 0.005429215908050537, 0.005661664009094239, 0.005492735862731934, 0.005494783878326416, 0.0055316481590271, 0.005501952171325684, 0.005454847812652588, 0.005371903896331787, 0.005356544017791748, 0.005372928142547607, 0.005359615802764893, 0.005346303939819336, 0.005355519771575928, 0.0053320322036743166, 0.005343167781829834, 0.005378047943115235, 0.00536678409576416, 0.005342207908630371, 0.005379072189331055, 0.005337183952331543, 0.005353375911712647, 0.005361663818359375, 0.0053647360801696775, 0.00535756778717041, 0.005375999927520752, 0.00535148811340332, 0.005338047981262207, 0.00536678409576416, 0.005339136123657226, 0.005587967872619629, 0.005502975940704345, 0.005468160152435303, 0.005332992076873779, 0.005421088218688965, 0.0053554878234863285, 0.005356544017791748, 0.005400576114654541, 0.005339136123657226, 0.005384191989898681, 0.005350399971008301, 0.005336063861846924, 0.005381120204925537, 0.005348351955413819, 0.005345280170440674, 0.00535756778717041, 0.005336063861846924, 0.005339136123657226, 0.00537395191192627, 0.005336063861846924, 0.005350399971008301, 0.011754495620727539, 0.005355519771575928, 0.005342207908630371, 0.005338111877441406, 0.005495808124542236, 0.005502975940704345, 0.005635072231292724, 0.00552345609664917, 0.005486591815948487, 0.005488639831542969, 0.005548031806945801, 0.005497856140136719, 0.005484543800354004, 0.005370880126953125, 0.0053309440612792965, 0.005375999927520752, 0.0053637118339538575, 0.005340159893035889, 0.005352447986602784, 0.005340159893035889, 0.005322751998901368, 0.00535756778717041, 0.005387263774871826, 0.005338111877441406, 0.0053647360801696775, 0.005352447986602784, 0.005339136123657226, 0.00535859203338623, 0.005645311832427978, 0.005430272102355957, 0.0054271998405456545, 0.005314559936523438, 0.005286911964416504, 0.005331967830657959, 0.005656576156616211, 0.005596159934997558, 0.005695487976074219, 0.005502975940704345, 0.005515327930450439, 0.00548140811920166, 0.005501952171325684, 0.005471295833587646, 0.0053554558753967285, 0.005350399971008301, 0.00532480001449585, 0.005338111877441406, 0.005387263774871826, 0.0053637118339538575, 0.005332992076873779, 0.005352447986602784, 0.005349376201629639, 0.005362688064575195, 0.005346303939819336, 0.005346303939819336, 0.005338111877441406, 0.00536575984954834, 0.0053534722328186036, 0.005345280170440674, 0.005368832111358642, 0.00537497615814209, 0.0053534722328186036, 0.0053678078651428224, 0.005343232154846191, 0.012178496360778808, 0.0054824318885803225, 0.005512191772460938, 0.0054999680519104005, 0.005510079860687256, 0.005391359806060791, 0.005339136123657226, 0.005380095958709717, 0.005359615802764893, 0.005349376201629639, 0.00535756778717041, 0.0053534722328186036, 0.005345280170440674, 0.005379072189331055, 0.005348383903503418, 0.0053647680282592775, 0.005370816230773926, 0.0053678078651428224, 0.005350399971008301, 0.005390336036682129, 0.005352447986602784, 0.005347328186035156, 0.005380095958709717, 0.0053678078651428224, 0.005355519771575928, 0.00537497615814209, 0.005350399971008301, 0.005577727794647217, 0.005508096218109131, 0.005371903896331787, 0.005336063861846924, 0.005331967830657959, 0.0053002238273620605, 0.00531660795211792, 0.005321728229522705, 0.0053534722328186036, 0.0053340158462524415, 0.005405695915222168, 0.005338111877441406, 0.0052899842262268066, 0.0053012480735778805, 0.005284863948822022, 0.005295104026794434, 0.005359615802764893, 0.005347328186035156, 0.005342207908630371, 0.00535859203338623, 0.005286911964416504, 0.005273600101470947, 0.005314559936523438, 0.00530944013595581, 0.005312511920928955, 0.00531660795211792, 0.0055285758972167965, 0.005432320117950439, 0.005513216018676758, 0.005667840003967285, 0.005526591777801513, 0.005488639831542969, 0.00534009599685669, 0.005401599884033203, 0.005368832111358642, 0.005328896045684814, 0.011801600456237793, 0.005369855880737305, 0.005362688064575195, 0.005342207908630371, 0.005371903896331787, 0.005385216236114502, 0.005341184139251709, 0.0053647360801696775, 0.0053463678359985355, 0.0053441920280456546, 0.00536678409576416, 0.005340159893035889, 0.005340159893035889, 0.005377024173736572, 0.0053340158462524415, 0.0053534722328186036, 0.00536575984954834, 0.005349376201629639, 0.005343232154846191, 0.005362688064575195, 0.005559296131134033, 0.005533696174621582, 0.005509119987487793, 0.005470208168029785, 0.005434368133544922, 0.005352447986602784, 0.005353536128997803, 0.005365695953369141, 0.005341216087341309, 0.005350368022918701, 0.006254591941833496, 0.005903359889984131, 0.005732351779937744, 0.0055203838348388675, 0.005519360065460205, 0.0055285758972167965, 0.005469279766082764, 0.005543839931488037, 0.00551526403427124, 0.005505023956298828, 0.005550079822540284, 0.005489664077758789, 0.005401599884033203, 0.005377024173736572, 0.005341184139251709, 0.00535859203338623, 0.005354496002197265, 0.005350399971008301, 0.0055808000564575196, 0.005488639831542969, 0.005454847812652588, 0.005575679779052735, 0.005500959873199463, 0.005437407970428467, 0.0053320322036743166, 0.00533292818069458, 0.005350399971008301, 0.005325856208801269, 0.005336031913757324, 0.005359615802764893, 0.005357632160186768, 0.005358528137207031, 0.00536678409576416, 0.011817983627319336, 0.0053534722328186036, 0.005362688064575195, 0.00536678409576416, 0.0053504319190979, 0.005343200206756592, 0.0053647360801696775, 0.005343232154846191, 0.005343232154846191, 0.00536678409576416, 0.005361663818359375, 0.005352447986602784, 0.005574656009674072, 0.005502975940704345, 0.005496831893920898, 0.0053821439743042, 0.0053534722328186036, 0.005343232154846191, 0.0053658242225646975, 0.005363647937774658, 0.0053340158462524415, 0.00568012809753418, 0.005463039875030518, 0.005513216018676758, 0.005483520030975342, 0.00546611213684082, 0.005438464164733887, 0.005339136123657226, 0.005356544017791748, 0.005399551868438721, 0.005381120204925537, 0.0053432960510253905, 0.005366720199584961, 0.005386240005493164, 0.005349376201629639, 0.0053647360801696775, 0.005581823825836182, 0.005502975940704345, 0.005479423999786377, 0.005471231937408447, 0.00551526403427124, 0.005350399971008301, 0.005351424217224121, 0.0053821439743042, 0.0053647360801696775, 0.0053340158462524415, 0.005344255924224854, 0.0053309440612792965, 0.0053350400924682614, 0.00537395191192627, 0.005355519771575928, 0.005337088108062744, 0.005399551868438721, 0.005356544017791748, 0.005340159893035889, 0.005529600143432617, 0.0053534722328186036, 0.005343232154846191, 0.005388288021087646, 0.005326848030090332, 0.005323775768280029, 0.005349376201629639, 0.005325823783874512, 0.011727871894836426, 0.005333055973052978, 0.0053482880592346195, 0.0053309440612792965, 0.005343232154846191, 0.005620736122131348, 0.0054876160621643065, 0.005439487934112549, 0.0053340158462524415, 0.005323840141296387, 0.005373888015747071, 0.005350399971008301, 0.005314559936523438, 0.00535859203338623, 0.005327871799468994, 0.005350399971008301, 0.005351424217224121, 0.005405695915222168, 0.0053340158462524415, 0.005368832111358642, 0.005356544017791748, 0.0053821439743042, 0.005371903896331787, 0.005338111877441406, 0.005336063861846924, 0.005405695915222168, 0.005347328186035156, 0.005325823783874512, 0.005359615802764893, 0.005332992076873779, 0.005338111877441406, 0.005359615802764893, 0.0053340158462524415, 0.0055244798660278325, 0.005502975940704345, 0.005495808124542236, 0.005436416149139404, 0.005378047943115235, 0.005328896045684814, 0.005341184139251709, 0.0053821439743042, 0.005349376201629639, 0.0053606400489807126, 0.005389311790466309, 0.005453824043273926, 0.005435391902923584, 0.005342207908630371, 0.005339136123657226, 0.005340159893035889, 0.005349376201629639, 0.005341184139251709, 0.005350399971008301, 0.0053534722328186036, 0.005341216087341309, 0.005353439807891846, 0.005340223789215088, 0.005339072227478027, 0.005337088108062744, 0.005350399971008301, 0.005344255924224854, 0.005381120204925537, 0.005348351955413819, 0.005641215801239013, 0.011813887596130371, 0.005369855880737305, 0.005348351955413819, 0.005338111877441406, 0.005362688064575195, 0.0053309440612792965, 0.005328896045684814, 0.005352447986602784, 0.005355519771575928, 0.005339136123657226, 0.005356544017791748, 0.005345280170440674, 0.005326848030090332, 0.005378047943115235, 0.005348351955413819, 0.005352447986602784, 0.005369855880737305, 0.0053534722328186036, 0.005355519771575928, 0.005386240005493164, 0.005342207908630371, 0.00536678409576416, 0.005381120204925537, 0.005339200019836426, 0.005358528137207031, 0.005387263774871826, 0.005534719944000244, 0.005512191772460938, 0.005459968090057373, 0.0053637118339538575, 0.00537395191192627, 0.005352447986602784, 0.005361663818359375, 0.00537497615814209, 0.0053340158462524415, 0.00537395191192627, 0.005370880126953125, 0.00535859203338623, 0.005340159893035889, 0.005386240005493164, 0.005360703945159912, 0.005342144012451172, 0.005384191989898681, 0.005329919815063477, 0.0053340158462524415, 0.0053637118339538575, 0.005354496002197265, 0.005336063861846924, 0.005348351955413819, 0.005356544017791748, 0.005336063861846924, 0.005361663818359375, 0.005342207908630371, 0.005326848030090332, 0.005493760108947754, 0.005607423782348633, 0.005530623912811279, 0.005495808124542236, 0.005470208168029785, 0.005435391902923584, 0.005351424217224121, 0.0053534722328186036, 0.005355552196502685, 0.011767775535583495, 0.005453824043273926, 0.00535859203338623, 0.005389311790466309, 0.005378047943115235, 0.0053422718048095705, 0.005389247894287109, 0.0053647360801696775, 0.005388288021087646, 0.00538316822052002, 0.0053534722328186036, 0.005356544017791748, 0.005375999927520752, 0.00535859203338623, 0.005554175853729248, 0.005590015888214111, 0.005517343997955323, 0.005497824192047119, 0.0054876160621643065, 0.005471231937408447, 0.0053678078651428224, 0.0055797758102417, 0.005529600143432617, 0.005526527881622314, 0.005372928142547607, 0.005450751781463623, 0.0054906878471374515, 0.005368832111358642, 0.005380095958709717, 0.005348351955413819, 0.005352447986602784, 0.005419007778167725, 0.005341184139251709, 0.005339136123657226, 0.005387263774871826, 0.005421088218688965, 0.005370848178863526, 0.005391359806060791, 0.005359615802764893, 0.0053350400924682614, 0.00537497615814209, 0.005368832111358642, 0.0053606400489807126, 0.005362688064575195, 0.005339136123657226, 0.00536575984954834, 0.005454847812652588, 0.005406720161437988, 0.005347328186035156, 0.005380095958709717, 0.00535859203338623, 0.005384191989898681, 0.00536678409576416, 0.005347328186035156, 0.005387263774871826, 0.005337088108062744, 0.005345280170440674, 0.005356544017791748, 0.005336063861846924, 0.005332992076873779, 0.00536678409576416, 0.005332992076873779, 0.005341184139251709, 0.011826175689697266, 0.005362688064575195, 0.005344255924224854, 0.005326848030090332, 0.005381120204925537, 0.005351424217224121, 0.005347328186035156, 0.005377024173736572, 0.0053350400924682614, 0.005356544017791748, 0.005681151866912842, 0.005616640090942383, 0.005479423999786377, 0.005543935775756836, 0.00536575984954834, 0.0053350400924682614, 0.005381152153015136, 0.005346271991729736, 0.0053975038528442385, 0.005359615802764893, 0.00535859203338623, 0.005387263774871826, 0.005338111877441406, 0.0053637118339538575, 0.00537497615814209, 0.0053309440612792965, 0.005350399971008301, 0.0053678078651428224, 0.00536678409576416, 0.005349376201629639, 0.005351424217224121, 0.005401631832122803, 0.005354464054107666, 0.005369855880737305, 0.005359615802764893, 0.005343232154846191, 0.005413887977600097, 0.005489664077758789, 0.005633024215698243, 0.005542912006378174, 0.005676032066345215, 0.005535744190216065, 0.005511168003082275, 0.0054906878471374515, 0.005516287803649903, 0.0054271998405456545, 0.005356544017791748, 0.005399551868438721, 0.005347328186035156, 0.0053637118339538575, 0.005389311790466309, 0.005348351955413819, 0.0053637118339538575, 0.005377024173736572, 0.005375040054321289, 0.005405632019042969, 0.00536575984954834, 0.005356544017791748, 0.005385216236114502, 0.005359615802764893, 0.005344255924224854, 0.005353536128997803, 0.005365695953369141, 0.011812864303588867, 0.0053340158462524415, 0.005378047943115235, 0.005348351955413819, 0.005615647792816162, 0.005541855812072754, 0.0054568958282470706, 0.005359615802764893, 0.0054568958282470706, 0.005440512180328369, 0.005355584144592285, 0.005419968128204345, 0.005345280170440674, 0.005328896045684814, 0.00538316822052002, 0.00537395191192627, 0.0053821439743042, 0.005355519771575928, 0.0053340158462524415, 0.005381120204925537, 0.00536678409576416, 0.005341184139251709, 0.005378079891204834, 0.005346271991729736, 0.005340191841125488, 0.005371871948242187, 0.005347328186035156, 0.005336063861846924, 0.005392384052276611, 0.00535756778717041, 0.005343232154846191, 0.0053760318756103515, 0.005535711765289306, 0.005489664077758789, 0.005540863990783691, 0.00552345609664917, 0.0055285758972167965, 0.005506048202514649, 0.005440512180328369, 0.0053534722328186036, 0.005355519771575928, 0.00535859203338623, 0.005371903896331787, 0.005359615802764893, 0.005329919815063477, 0.005343232154846191, 0.0053719358444213865, 0.005359583854675293, 0.00537395191192627, 0.005370880126953125, 0.005347328186035156, 0.00537395191192627, 0.005371903896331787, 0.0053647360801696775, 0.005395455837249756, 0.005340159893035889, 0.005340159893035889, 0.0053647360801696775, 0.005344255924224854, 0.005332992076873779, 0.005355519771575928, 0.005636096000671387, 0.005393407821655274, 0.011786239624023438, 0.005387263774871826, 0.00535859203338623, 0.005337088108062744, 0.005352447986602784, 0.005337088108062744, 0.0053350400924682614, 0.005703680038452149, 0.005446656227111817, 0.00537395191192627, 0.005359615802764893, 0.005346303939819336, 0.00535756778717041, 0.0053637118339538575, 0.005344255924224854, 0.005380095958709717, 0.00535756778717041, 0.005329919815063477, 0.005368832111358642, 0.005362688064575195, 0.005380095958709717, 0.00535756778717041, 0.005356544017791748, 0.005339136123657226, 0.005387263774871826, 0.0053944320678710935, 0.005322751998901368, 0.005345280170440674, 0.0053340158462524415, 0.00535859203338623, 0.005372928142547607, 0.005356544017791748, 0.005337088108062744, 0.00535756778717041, 0.005344319820404053, 0.005366720199584961, 0.0053647360801696775, 0.005352447986602784, 0.005344255924224854, 0.0053678078651428224, 0.005340159893035889, 0.005347328186035156, 0.0053637118339538575, 0.005326848030090332, 0.005277696132659912, 0.005298175811767578, 0.005287936210632324, 0.005291007995605469, 0.005291007995605469, 0.005296127796173096, 0.005269536018371582, 0.005336031913757324, 0.0053340158462524415, 0.0053637118339538575, 0.005631999969482422, 0.005485568046569824, 0.0055316481590271, 0.0053678078651428224, 0.005325823783874512, 0.00537395191192627, 0.005336063861846924, 0.005328896045684814, 0.0053647360801696775, 0.011806719779968262, 0.005380095958709717, 0.005336063861846924, 0.005396480083465576, 0.005338111877441406, 0.0053647360801696775, 0.005395455837249756, 0.005346303939819336, 0.0053534722328186036, 0.005375999927520752, 0.0053770880699157714, 0.005363647937774658, 0.00536678409576416, 0.00537395191192627, 0.0053463678359985355, 0.005370816230773926, 0.005354496002197265, 0.005361663818359375, 0.005636096000671387, 0.005479423999786377, 0.005497920036315918, 0.005550015926361084, 0.005471231937408447, 0.005341184139251709, 0.005302271842956543, 0.005310463905334473, 0.0053350400924682614, 0.00530944013595581, 0.005307392120361328, 0.005329919815063477, 0.00541593599319458, 0.005443583965301513, 0.005381120204925537, 0.005337088108062744, 0.005340159893035889, 0.0053637118339538575, 0.00552243185043335, 0.005493760108947754, 0.0055101442337036136, 0.005390336036682129, 0.0053637118339538575, 0.005375999927520752, 0.005350399971008301, 0.005342207908630371, 0.005361663818359375, 0.0053637118339538575, 0.005596159934997558, 0.005495808124542236, 0.005430272102355957, 0.005354496002197265, 0.005359615802764893, 0.005352447986602784, 0.0053678078651428224, 0.005362751960754394, 0.005342144012451172, 0.0053678078651428224, 0.005337088108062744, 0.0053340158462524415, 0.005347328186035156, 0.005352447986602784, 0.005337088108062744, 0.0053760638236999515, 0.005345215797424316, 0.011821120262145997, 0.005352384090423584, 0.005349376201629639, 0.005348351955413819, 0.005356544017791748, 0.005368832111358642, 0.00535756778717041, 0.005351424217224121, 0.005349376201629639, 0.0053534722328186036, 0.005362688064575195, 0.0056217598915100095, 0.005480447769165039, 0.005482495784759522, 0.005500927925109863, 0.005336063861846924, 0.005361663818359375, 0.00535859203338623, 0.005343232154846191, 0.00536678409576416, 0.005343232154846191, 0.005446656227111817, 0.00536678409576416, 0.005375999927520752, 0.00536575984954834, 0.005357600212097168, 0.00536681604385376, 0.0053482880592346195, 0.005384191989898681, 0.005344255924224854, 0.005343232154846191, 0.0053821439743042, 0.005352447986602784, 0.0053534722328186036, 0.005378047943115235, 0.00537497615814209, 0.005351424217224121, 0.00537395191192627, 0.0053678078651428224, 0.0055582718849182125, 0.005505023956298828, 0.005493760108947754, 0.00550707197189331, 0.005368832111358642, 0.005342207908630371, 0.00536678409576416, 0.005354496002197265, 0.005341184139251709, 0.005419007778167725, 0.005459968090057373, 0.005478400230407715, 0.005401599884033203, 0.005346303939819336, 0.005341184139251709, 0.005377024173736572, 0.005336063861846924, 0.0055316481590271, 0.0053821439743042, 0.005482495784759522, 0.005430272102355957, 0.005379072189331055, 0.005344255924224854, 0.005340159893035889, 0.011783200263977051, 0.005725215911865234, 0.005380032062530518, 0.005381120204925537, 0.005359615802764893, 0.005359615802764893, 0.005440512180328369, 0.005586976051330566, 0.0055203518867492675, 0.00553984022140503, 0.005385216236114502, 0.005437439918518067, 0.005512191772460938, 0.005452799797058106, 0.00537497615814209, 0.005336063861846924, 0.005315584182739258, 0.0053534722328186036, 0.005344255924224854, 0.005338111877441406, 0.005362751960754394, 0.0053216638565063475, 0.005340159893035889, 0.00538316822052002, 0.005359615802764893, 0.005344255924224854, 0.005345280170440674, 0.005339136123657226, 0.005342207908630371, 0.00537497615814209, 0.005343232154846191, 0.005372928142547607, 0.005346303939819336, 0.005340159893035889, 0.005350399971008301, 0.005368832111358642, 0.005347328186035156, 0.005347328186035156, 0.0053606400489807126, 0.005354496002197265, 0.005329919815063477, 0.005378047943115235, 0.005349376201629639, 0.005349376201629639, 0.005347328186035156, 0.005348351955413819, 0.005341184139251709, 0.005350399971008301, 0.0053944320678710935, 0.005349376201629639, 0.005369855880737305, 0.005346303939819336, 0.0053309440612792965, 0.005351424217224121, 0.005348351955413819, 0.005369855880737305, 0.0053534722328186036, 0.0053606400489807126, 0.005343232154846191, 0.0053637118339538575, 0.00538316822052002, 0.005342207908630371, 0.005386303901672363, 0.011814847946166992, 0.0053534722328186036, 0.0053606400489807126, 0.005359615802764893, 0.00537395191192627, 0.00531660795211792, 0.00535756778717041, 0.0053340158462524415, 0.005342207908630371, 0.005339136123657226, 0.005325823783874512, 0.005327871799468994, 0.00532480001449585, 0.005347328186035156, 0.005328896045684814, 0.00537497615814209, 0.0053534722328186036, 0.0053678078651428224, 0.005357632160186768, 0.005333951950073242, 0.0053350400924682614, 0.005371903896331787, 0.005331999778747558, 0.005344223976135254, 0.005407743930816651, 0.00536575984954834, 0.00536575984954834, 0.005337088108062744, 0.005329919815063477, 0.005346303939819336, 0.005362688064575195, 0.0053309440612792965, 0.005357664108276367, 0.005351327896118164, 0.005318655967712403, 0.00538316822052002, 0.005333024024963379, 0.005321695804595948, 0.005352447986602784, 0.005336063861846924, 0.005344255924224854, 0.005387263774871826, 0.005326848030090332, 0.005307392120361328, 0.0053637118339538575, 0.005332992076873779, 0.005326848030090332, 0.005348351955413819, 0.005336063861846924, 0.0053350720405578615, 0.005342175960540771, 0.005310463905334473, 0.005307392120361328, 0.005340191841125488, 0.00534115219116211, 0.005339136123657226, 0.005345280170440674, 0.005331967830657959, 0.005313536167144775, 0.005379072189331055, 0.00531660795211792, 0.005321728229522705, 0.0053678078651428224, 0.011678720474243164, 0.005375999927520752, 0.0053350400924682614, 0.005349376201629639, 0.005349376201629639, 0.005327871799468994, 0.0053534722328186036, 0.005355519771575928, 0.005341184139251709, 0.0053647360801696775, 0.005347328186035156, 0.005315584182739258, 0.005348351955413819, 0.005346303939819336, 0.005306367874145508, 0.005381120204925537, 0.005321728229522705, 0.005346303939819336, 0.005379072189331055, 0.005320703983306885, 0.005311488151550293, 0.005349376201629639, 0.005355519771575928, 0.005319680213928223, 0.005369855880737305, 0.005338111877441406, 0.005337088108062744, 0.005371903896331787, 0.005359615802764893, 0.00530944013595581, 0.00535756778717041, 0.005337088108062744, 0.005319680213928223, 0.00535756778717041, 0.0053606400489807126, 0.005326911926269531, 0.005337056159973145, 0.005351391792297364, 0.00531660795211792, 0.005350399971008301, 0.005344255924224854, 0.005329919815063477, 0.005311488151550293, 0.005322751998901368, 0.005330976009368897, 0.00534115219116211, 0.005420032024383545, 0.005333055973052978, 0.005368768215179443, 0.005344255924224854, 0.005347328186035156, 0.0053647360801696775, 0.005370880126953125, 0.005356544017791748, 0.005346303939819336, 0.00536678409576416, 0.005319680213928223, 0.005381120204925537, 0.005338143825531006, 0.0055725760459899906, 0.005421055793762207, 0.005352447986602784, 0.005345280170440674, 0.011767807960510255, 0.005338111877441406, 0.005337088108062744, 0.005325823783874512, 0.0053678078651428224, 0.005421055793762207, 0.005440512180328369, 0.005402624130249023, 0.005344255924224854, 0.005340159893035889, 0.0053647360801696775, 0.00537395191192627, 0.005342207908630371, 0.005381120204925537, 0.0053340158462524415, 0.005336063861846924, 0.0053678078651428224, 0.005343232154846191, 0.005337088108062744, 0.005421055793762207, 0.005327871799468994, 0.005354496002197265, 0.005370880126953125, 0.005346303939819336, 0.005371903896331787, 0.005349376201629639, 0.0053637118339538575, 0.005384191989898681, 0.005349376201629639, 0.005348351955413819, 0.005350399971008301, 0.005371903896331787, 0.00535756778717041, 0.005369855880737305, 0.00538316822052002, 0.005406720161437988, 0.005362688064575195, 0.005354496002197265, 0.005359615802764893, 0.005404672145843506, 0.0053534722328186036, 0.006082560062408447, 0.005584928035736084, 0.005510111808776855, 0.005946368217468262, 0.006405119895935059, 0.005727231979370117, 0.005501952171325684, 0.005547008037567139, 0.005518335819244385, 0.005480447769165039, 0.00552243185043335, 0.005377024173736572, 0.00537395191192627, 0.005381120204925537, 0.005355519771575928, 0.0053534722328186036, 0.005388288021087646, 0.005370880126953125, 0.005337088108062744, 0.005388288021087646, 0.005331967830657959, 0.00536681604385376, 0.011834336280822753, 0.005361663818359375, 0.005349376201629639, 0.005387263774871826, 0.005378047943115235, 0.005351424217224121, 0.005331967830657959, 0.005389311790466309, 0.005342207908630371, 0.005341184139251709, 0.005368832111358642, 0.005371903896331787, 0.005449728012084961, 0.005362751960754394, 0.005346240043640137, 0.005375999927520752, 0.00536678409576416, 0.00535145616531372, 0.005387231826782226, 0.00537395191192627, 0.005372960090637207, 0.0053851838111877446, 0.005348351955413819, 0.005347328186035156, 0.005362688064575195, 0.005336063861846924, 0.005326848030090332, 0.005391359806060791, 0.00536575984954834, 0.005346303939819336, 0.00536678409576416, 0.005368832111358642, 0.005342207908630371, 0.005362688064575195, 0.005354496002197265, 0.005377024173736572, 0.0053606400489807126, 0.00538316822052002, 0.0053350400924682614, 0.005361663818359375, 0.005361663818359375, 0.005361663818359375, 0.005368832111358642, 0.0053647360801696775, 0.005340159893035889, 0.0053534722328186036, 0.005342207908630371, 0.005356544017791748, 0.005362688064575195, 0.005351424217224121, 0.005336063861846924, 0.005412896156311035, 0.005338079929351806, 0.005331967830657959, 0.0054271998405456545, 0.005340159893035889, 0.0053606400489807126, 0.005361663818359375, 0.0053309440612792965, 0.005368832111358642, 0.005384191989898681, 0.005350399971008301, 0.00538316822052002, 0.01176371192932129, 0.005386240005493164, 0.005328896045684814, 0.005362688064575195, 0.005359615802764893, 0.005331967830657959, 0.0053688640594482424, 0.00536572790145874, 0.005352447986602784, 0.00536678409576416, 0.005347328186035156, 0.005343232154846191, 0.005402624130249023, 0.0053534722328186036, 0.005356544017791748, 0.005398528099060058, 0.005339136123657226, 0.005331967830657959, 0.00536678409576416, 0.005338111877441406, 0.005345280170440674, 0.005378047943115235, 0.005356544017791748, 0.005350399971008301, 0.005385216236114502, 0.005390336036682129, 0.005356544017791748, 0.00537395191192627, 0.005354496002197265, 0.005345280170440674, 0.0053944320678710935, 0.0057784318923950195, 0.005426176071166992, 0.005399551868438721, 0.005359615802764893, 0.005384191989898681, 0.005456992149353027, 0.005542816162109375, 0.005990399837493897, 0.005540863990783691, 0.0054568958282470706, 0.005370880126953125, 0.005366847991943359, 0.005394368171691894, 0.00538316822052002, 0.005387263774871826, 0.005369855880737305, 0.0053678078651428224, 0.0053350400924682614, 0.00536575984954834, 0.005349376201629639, 0.005347328186035156, 0.005620736122131348, 0.005624832153320313, 0.005490719795227051, 0.005450719833374023, 0.005342207908630371, 0.005352447986602784, 0.005434368133544922, 0.005620736122131348, 0.005559296131134033, 0.005368832111358642, 0.005392384052276611, 0.011998208045959472, 0.00535964822769165, 0.005344223976135254, 0.005346303939819336, 0.005403647899627686, 0.005517312049865723, 0.005346303939819336, 0.005376095771789551, 0.005357471942901612, 0.005359615802764893, 0.0053975038528442385, 0.0053350400924682614, 0.00532480001449585, 0.005377056121826172, 0.005336031913757324, 0.005352447986602784, 0.005381120204925537, 0.005616640090942383, 0.005501952171325684, 0.005471231937408447, 0.005354496002197265, 0.005372928142547607, 0.005356544017791748, 0.005350399971008301, 0.0053647360801696775, 0.005361663818359375, 0.005343232154846191, 0.005391359806060791, 0.005361663818359375, 0.005347328186035156, 0.005362688064575195, 0.005371903896331787, 0.0053647360801696775, 0.005355519771575928, 0.005329919815063477, 0.005328927993774414, 0.005403647899627686, 0.005337056159973145, 0.005327871799468994, 0.005368832111358642, 0.005350399971008301, 0.0053350400924682614, 0.005342207908630371, 0.005359615802764893, 0.005331967830657959, 0.005561344146728516, 0.005470240116119384, 0.00551632022857666, 0.005362656116485596, 0.005343200206756592, 0.005386240005493164, 0.0053647360801696775, 0.00535756778717041, 0.005361663818359375, 0.005336063861846924, 0.0053534722328186036, 0.005372928142547607, 0.005355519771575928, 0.005336063861846924, 0.005361663818359375, 0.005349376201629639, 0.005344255924224854, 0.005343232154846191, 0.011752448081970214, 0.005341184139251709, 0.005337088108062744, 0.005342207908630371, 0.005351424217224121, 0.00535859203338623, 0.005356544017791748, 0.005340159893035889, 0.005329919815063477, 0.005395455837249756, 0.005640192031860352, 0.005587967872619629, 0.005519360065460205, 0.005470208168029785, 0.005533696174621582, 0.005469183921813964, 0.005451776027679443, 0.005395455837249756, 0.005344319820404053, 0.005337024211883545, 0.005349376201629639, 0.00536575984954834, 0.005346303939819336, 0.005371903896331787, 0.0053504319190979, 0.005340127944946289, 0.0053606400489807126, 0.0053606400489807126, 0.0053821439743042, 0.005371903896331787, 0.005347328186035156, 0.005347328186035156, 0.00536678409576416, 0.005369855880737305, 0.005328896045684814, 0.0053637118339538575, 0.0053350400924682614, 0.005348351955413819, 0.005638144016265869, 0.0054988799095153805, 0.005441535949707031, 0.005359615802764893, 0.005351424217224121, 0.005351424217224121, 0.005332992076873779, 0.0053340158462524415, 0.005370880126953125, 0.005345280170440674, 0.005344255924224854, 0.005369855880737305, 0.005369855880737305, 0.005338175773620606, 0.00537286376953125, 0.005343232154846191, 0.005342207908630371, 0.00536678409576416, 0.005347392082214355, 0.005326784133911132, 0.005359615802764893, 0.005336063861846924, 0.00532480001449585, 0.005346303939819336, 0.0053340158462524415, 0.011702272415161133, 0.005395455837249756, 0.005550079822540284, 0.005496831893920898, 0.0054917120933532714, 0.00553984022140503, 0.0055101442337036136, 0.005554175853729248, 0.005483520030975342, 0.00532480001449585, 0.005342207908630371, 0.005341184139251709, 0.005334047794342041, 0.0053636798858642575, 0.00535756778717041, 0.005347328186035156, 0.00537497615814209, 0.005346303939819336, 0.005343232154846191, 0.005369855880737305, 0.005338111877441406, 0.005350399971008301, 0.005375999927520752, 0.00536575984954834, 0.005352447986602784, 0.00535859203338623, 0.005350399971008301, 0.005341184139251709, 0.005369919776916504, 0.005350336074829102, 0.005329919815063477, 0.005606400012969971, 0.005347328186035156, 0.005347328186035156, 0.005361663818359375, 0.005361663818359375, 0.005346303939819336, 0.005379072189331055, 0.005381120204925537, 0.005339136123657226, 0.005368832111358642, 0.005336063861846924, 0.005332992076873779, 0.00535756778717041, 0.00532480001449585, 0.005322751998901368, 0.0058726401329040525, 0.005451776027679443, 0.005375999927520752, 0.005351424217224121, 0.005340159893035889, 0.00573747205734253, 0.005532671928405761, 0.00551423978805542, 0.00546611213684082, 0.005393407821655274, 0.005734399795532226, 0.005370880126953125, 0.005378047943115235, 0.006987775802612305, 0.005979135990142822, 0.005567488193511963, 0.0054876160621643065, 0.011932671546936035, 0.005455872058868408, 0.00536575984954834, 0.00535756778717041, 0.0054057598114013675, 0.0053861761093139645, 0.005362688064575195, 0.005407743930816651, 0.005388288021087646, 0.005362688064575195, 0.005419007778167725, 0.00535756778717041, 0.005384191989898681, 0.005395455837249756, 0.005339136123657226, 0.005387263774871826, 0.005349376201629639, 0.00535859203338623, 0.005361663818359375, 0.005380095958709717, 0.005339136123657226, 0.005371903896331787, 0.005387263774871826, 0.005341184139251709, 0.005387263774871826, 0.005387263774871826, 0.005340159893035889, 0.005375040054321289, 0.005351359844207764, 0.005361663818359375, 0.0053637118339538575, 0.0053637118339538575, 0.00535756778717041, 0.005380095958709717, 0.005348351955413819, 0.00536575984954834, 0.005375999927520752, 0.0053606400489807126, 0.005354496002197265, 0.0053821439743042, 0.0053637118339538575, 0.005593088150024414, 0.005469183921813964, 0.005352447986602784, 0.00540780782699585, 0.00547219181060791, 0.005387263774871826, 0.005519360065460205, 0.005359615802764893, 0.005354496002197265, 0.00536575984954834, 0.005351424217224121, 0.005351424217224121, 0.005398528099060058, 0.00536575984954834, 0.00535756778717041, 0.005362720012664795, 0.005349343776702881, 0.005347328186035156, 0.005375999927520752, 0.005344255924224854, 0.0053350400924682614, 0.0053606400489807126]",tokens/s,181.54611465520205,,,main,False,False,,,,,,,,,,,,,,,,,,,,,,, bfloat16-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,google/gemma-2b,google/gemma-2b,cuda,0,42,,,True,,,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.223-212.873.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.0,,,,1.21.4,,,,0.12.0,,,gemma,MB,1554.223104,6797.39392,0.0,6150.946816,5419.87328,s,10,0.31568086433410647,0.03156808643341064,4.172659623563964e-05,0.031570879936218266,0.031607007598876954,0.031624575424194334,0.03163862968444824,"[0.03147007942199707, 0.03158172798156738, 0.03154080009460449, 0.03164214324951172, 0.03156918334960938, 0.03155824089050293, 0.03157257652282715, 0.03157488059997558, 0.031603103637695314, 0.03156812858581543]",tokens/s,8109.455748608755,kWh,3.727353862074855e-07,2.042397252914246e-07,1.9012659794763286e-06,2.4782410909752386e-06,tokens/kWh,103299070.02682243,MB,1607.364608,6797.39392,0.0,6150.946816,5419.87584,s,10,14.299476806640623,1.4299476806640627,0.007636718609743986,1.429648193359375,1.4357959106445313,1.4415692810058593,1.4461879772949218,"[1.434512939453125, 1.4300867919921876, 1.4292095947265624, 1.4237501220703126, 1.428209716796875, 1.424857666015625, 1.4473426513671874, 1.416581787109375, 1.4329952392578125, 1.4319302978515625]",tokens/s,44.05755598746313,kWh,1.7508762621084808e-05,9.594742285843944e-06,5.432700122532346e-05,8.143050613225222e-05,tokens/kWh,773665.828598449,,s,630,14.29779354667665,0.022694910391550215,0.0004375520451270339,0.02244095993041992,0.02324111328125,0.023350476932525635,0.02409779167175293,"[0.022468608856201173, 0.02243891143798828, 0.022460416793823244, 0.022425600051879883, 0.022425600051879883, 0.022406143188476564, 0.02247987174987793, 0.022382591247558595, 0.022372352600097657, 0.02240716743469238, 0.02329804801940918, 0.023812095642089845, 0.023335935592651368, 0.02334115219116211, 0.023283615112304687, 0.023201791763305665, 0.02325606346130371, 0.023198720932006835, 0.023221248626708983, 0.023211008071899415, 0.02315673637390137, 0.023143423080444335, 0.023137279510498047, 0.02312704086303711, 0.023192575454711914, 0.023240703582763672, 0.023166976928710937, 0.022508544921875, 0.0224399356842041, 0.022427648544311524, 0.02240716743469238, 0.022432767868041992, 0.022411264419555665, 0.02243891143798828, 0.022401023864746093, 0.022393856048583984, 0.022378496170043945, 0.023465024948120118, 0.02318636894226074, 0.02290176010131836, 0.022846464157104493, 0.022931455612182617, 0.023196672439575194, 0.02306662368774414, 0.023060480117797853, 0.02310553550720215, 0.022377471923828125, 0.02231808090209961, 0.022432767868041992, 0.022363136291503907, 0.022313983917236328, 0.022338592529296875, 0.02233750343322754, 0.022378496170043945, 0.023104511260986327, 0.022526975631713866, 0.022460416793823244, 0.02232524871826172, 0.022452224731445314, 0.022425600051879883, 0.022386688232421875, 0.022840320587158205, 0.023371776580810546, 0.022433792114257813, 0.02249830436706543, 0.022254592895507814, 0.022775808334350587, 0.023218175888061524, 0.02253107261657715, 0.02246553611755371, 0.022510591506958007, 0.022483968734741212, 0.023348224639892577, 0.023257087707519532, 0.022477823257446287, 0.022404096603393556, 0.022466560363769532, 0.022445056915283205, 0.022397951126098634, 0.022874111175537108, 0.02326527976989746, 0.023244800567626952, 0.023145471572875977, 0.023227392196655275, 0.023421951293945312, 0.023197696685791015, 0.02326527976989746, 0.023179264068603517, 0.02313113594055176, 0.023158784866333007, 0.023192575454711914, 0.022955007553100586, 0.022968320846557616, 0.023174144744873046, 0.023248895645141602, 0.023234560012817384, 0.02327859115600586, 0.02260479927062988, 0.02247475242614746, 0.022394880294799805, 0.022552576065063477, 0.02248192024230957, 0.022411264419555665, 0.022494207382202147, 0.02244095993041992, 0.022467584609985353, 0.022413312911987306, 0.022376447677612304, 0.022380544662475587, 0.02228223991394043, 0.0224901123046875, 0.022443008422851563, 0.022441984176635742, 0.022424575805664062, 0.022367231369018553, 0.022421503067016603, 0.02226483154296875, 0.022551551818847656, 0.022441984176635742, 0.022441984176635742, 0.02244095993041992, 0.02241535949707031, 0.022412288665771486, 0.022624256134033204, 0.022395904541015626, 0.022434816360473633, 0.02245529556274414, 0.022396928787231447, 0.022408191680908202, 0.023196672439575194, 0.02329804801940918, 0.022477823257446287, 0.022459392547607423, 0.02249625587463379, 0.02249113655090332, 0.02245631980895996, 0.022424575805664062, 0.022466560363769532, 0.022401023864746093, 0.02246348762512207, 0.02266726493835449, 0.02328371238708496, 0.02334617614746094, 0.023177215576171875, 0.023182336807250976, 0.022445056915283205, 0.022444128036499023, 0.023586719512939454, 0.023211008071899415, 0.02256390380859375, 0.022376384735107422, 0.022378496170043945, 0.022322175979614257, 0.02239897537231445, 0.02230681610107422, 0.02250752067565918, 0.022352895736694335, 0.022410240173339844, 0.022419456481933595, 0.022560768127441407, 0.022411264419555665, 0.022408191680908202, 0.022380544662475587, 0.022402048110961914, 0.02229964828491211, 0.02230886459350586, 0.023017471313476562, 0.023166976928710937, 0.02288844871520996, 0.022845439910888672, 0.022363136291503907, 0.022363136291503907, 0.022437887191772463, 0.022376447677612304, 0.022375423431396483, 0.02234982490539551, 0.022362112045288086, 0.022383615493774413, 0.022387712478637696, 0.0223191032409668, 0.022363136291503907, 0.022525951385498046, 0.024303615570068358, 0.023610368728637695, 0.02388479995727539, 0.023292959213256834, 0.02315977668762207, 0.02309017562866211, 0.02312294387817383, 0.022432767868041992, 0.022323200225830078, 0.022388736724853517, 0.022377471923828125, 0.022378496170043945, 0.022380544662475587, 0.022344703674316405, 0.022382591247558595, 0.022408191680908202, 0.02242252731323242, 0.022375423431396483, 0.022345727920532226, 0.02243891143798828, 0.022379520416259766, 0.022366207122802736, 0.023213056564331053, 0.022347776412963868, 0.02232729530334473, 0.02234163284301758, 0.022424575805664062, 0.02323967933654785, 0.02330112075805664, 0.022666240692138673, 0.02409779167175293, 0.023383039474487305, 0.022569984436035157, 0.02253926467895508, 0.02331340789794922, 0.02305740737915039, 0.023197696685791015, 0.023204864501953124, 0.023193599700927735, 0.022409215927124023, 0.022382591247558595, 0.022451200485229493, 0.022785024642944338, 0.022803455352783202, 0.022692863464355468, 0.022575103759765625, 0.022421503067016603, 0.02245734405517578, 0.022433792114257813, 0.022425600051879883, 0.022380544662475587, 0.022377471923828125, 0.022363136291503907, 0.023299072265625, 0.02308915138244629, 0.02307481575012207, 0.022360063552856444, 0.022369279861450195, 0.022351871490478514, 0.02235699272155762, 0.02234982490539551, 0.02231500816345215, 0.02233344078063965, 0.022383615493774413, 0.022352895736694335, 0.022355968475341798, 0.022372352600097657, 0.022322175979614257, 0.022396928787231447, 0.022374399185180666, 0.023379968643188476, 0.022391807556152343, 0.02266726493835449, 0.023500799179077148, 0.025589759826660157, 0.024466432571411133, 0.022642688751220705, 0.022359039306640623, 0.022363136291503907, 0.022377471923828125, 0.022385663986206054, 0.022338560104370117, 0.022374399185180666, 0.02242255973815918, 0.022381536483764647, 0.022409215927124023, 0.022334463119506837, 0.023516159057617187, 0.02333286476135254, 0.022468608856201173, 0.022359039306640623, 0.022433792114257813, 0.022351871490478514, 0.02234163284301758, 0.02239897537231445, 0.022445056915283205, 0.022417407989501953, 0.022409215927124023, 0.022369279861450195, 0.02234982490539551, 0.022346752166748047, 0.022350847244262697, 0.022404096603393556, 0.02243891143798828, 0.0222740478515625, 0.022185983657836913, 0.022289407730102538, 0.02231091117858887, 0.02248192024230957, 0.022370304107666016, 0.022374399185180666, 0.022385663986206054, 0.022428672790527345, 0.022372352600097657, 0.023622655868530275, 0.023119871139526366, 0.02306355285644531, 0.02307072067260742, 0.022425600051879883, 0.022525951385498046, 0.023161855697631836, 0.02327244758605957, 0.022424575805664062, 0.02243174362182617, 0.02227916717529297, 0.023004159927368165, 0.023109632492065428, 0.023187456130981447, 0.022399999618530272, 0.022396928787231447, 0.023176191329956054, 0.022411264419555665, 0.022367231369018553, 0.02253824043273926, 0.022405120849609376, 0.022413312911987306, 0.022433792114257813, 0.022374399185180666, 0.02245529556274414, 0.022413312911987306, 0.02243174362182617, 0.0224399356842041, 0.022355968475341798, 0.022435840606689454, 0.02234163284301758, 0.022392831802368163, 0.022421503067016603, 0.022417407989501953, 0.022419456481933595, 0.022433792114257813, 0.022368255615234374, 0.022399999618530272, 0.022386688232421875, 0.022382591247558595, 0.022388736724853517, 0.022342655181884767, 0.022509567260742186, 0.023088127136230468, 0.02307891273498535, 0.023150592803955077, 0.023134208679199218, 0.02314035224914551, 0.023352319717407227, 0.023201791763305665, 0.02330521583557129, 0.02245631980895996, 0.02245529556274414, 0.02244095993041992, 0.023178239822387696, 0.023299072265625, 0.02309427261352539, 0.023130111694335938, 0.022377471923828125, 0.022383615493774413, 0.022993919372558593, 0.023227392196655275, 0.023146495819091797, 0.02311065673828125, 0.02311577606201172, 0.022915071487426757, 0.022199296951293947, 0.022451200485229493, 0.022345727920532226, 0.022388736724853517, 0.022386688232421875, 0.022379520416259766, 0.022393856048583984, 0.022336511611938475, 0.022433792114257813, 0.02247270393371582, 0.02243071937561035, 0.022380544662475587, 0.022372352600097657, 0.022371328353881836, 0.022330368041992187, 0.022339584350585938, 0.022392831802368163, 0.02228223991394043, 0.022416383743286132, 0.022576128005981445, 0.023171072006225587, 0.023121919631958008, 0.02305740737915039, 0.02215116882324219, 0.022360063552856444, 0.023113727569580078, 0.02311065673828125, 0.023359487533569336, 0.023180288314819338, 0.023157760620117186, 0.022424575805664062, 0.022350847244262697, 0.023392255783081056, 0.023432191848754884, 0.024328191757202147, 0.022697023391723633, 0.02263238334655762, 0.02290995216369629, 0.023405567169189453, 0.02244607925415039, 0.02246553611755371, 0.022411264419555665, 0.022405120849609376, 0.022715391159057616, 0.02332057571411133, 0.022413312911987306, 0.022393856048583984, 0.022443008422851563, 0.022403072357177735, 0.022373376846313478, 0.022391807556152343, 0.02310553550720215, 0.023189504623413085, 0.02310860824584961, 0.023161855697631836, 0.02310553550720215, 0.02314035224914551, 0.023165952682495116, 0.02365132713317871, 0.023144447326660156, 0.024953855514526366, 0.02409779167175293, 0.023348224639892577, 0.023361536026000978, 0.022938623428344726, 0.023196672439575194, 0.023182336807250976, 0.023199743270874023, 0.023388160705566406, 0.022444032669067384, 0.022392831802368163, 0.02242355155944824, 0.023007232666015624, 0.023224319458007812, 0.023144447326660156, 0.023335935592651368, 0.023215103149414062, 0.023157760620117186, 0.023181312561035155, 0.02429952049255371, 0.023207935333251953, 0.02315673637390137, 0.022397951126098634, 0.022641664505004884, 0.022379520416259766, 0.022305791854858398, 0.022362112045288086, 0.022280191421508787, 0.022350847244262697, 0.022433792114257813, 0.022372352600097657, 0.022363136291503907, 0.022373376846313478, 0.022390783309936522, 0.022344703674316405, 0.022379520416259766, 0.022330368041992187, 0.022339584350585938, 0.022372352600097657, 0.022323200225830078, 0.022347776412963868, 0.022384639739990234, 0.022391807556152343, 0.022366207122802736, 0.022386688232421875, 0.022444032669067384, 0.023022592544555662, 0.02288844871520996, 0.02311884880065918, 0.02247270393371582, 0.022328319549560546, 0.02223308753967285, 0.022369279861450195, 0.022410240173339844, 0.02226688003540039, 0.02229145622253418, 0.022346752166748047, 0.022339584350585938, 0.022420480728149415, 0.022391807556152343, 0.0224399356842041, 0.022303743362426756, 0.022335487365722655, 0.022387712478637696, 0.023422975540161133, 0.022288383483886717, 0.022367231369018553, 0.022373376846313478, 0.022372352600097657, 0.022399999618530272, 0.02241231918334961, 0.022439903259277343, 0.02247475242614746, 0.022436864852905275, 0.02246963119506836, 0.02241433525085449, 0.022408191680908202, 0.022367231369018553, 0.022387712478637696, 0.02244095993041992, 0.022451200485229493, 0.022397951126098634, 0.022459392547607423, 0.02263654327392578, 0.023185407638549805, 0.023178239822387696, 0.023152639389038086, 0.022458368301391602, 0.022388736724853517, 0.023216127395629883, 0.023141376495361327, 0.023134208679199218, 0.023136255264282226, 0.022387712478637696, 0.02244915199279785, 0.0224399356842041, 0.022424575805664062, 0.022409215927124023, 0.022379520416259766, 0.022365184783935548, 0.022385663986206054, 0.02332159996032715, 0.02311577606201172, 0.02306252861022949, 0.023085056304931642, 0.023059455871582032, 0.023162879943847657, 0.022396928787231447, 0.022328319549560546, 0.0222873592376709, 0.022377471923828125, 0.022347776412963868, 0.02230681610107422, 0.022369279861450195, 0.022359039306640623, 0.022392831802368163, 0.022373376846313478, 0.022334463119506837, 0.022390783309936522, 0.022379520416259766, 0.022379520416259766, 0.022773759841918945, 0.02310655975341797, 0.02310758399963379, 0.02307379150390625, 0.02305740737915039, 0.02330931282043457, 0.023189504623413085, 0.023189504623413085, 0.02305433654785156, 0.02306764793395996, 0.022372352600097657, 0.022418432235717774, 0.023195648193359376, 0.023130111694335938, 0.02324787139892578, 0.023159807205200195, 0.022373376846313478, 0.022354944229125977, 0.022421503067016603, 0.02290483283996582, 0.023191551208496093, 0.023170047760009766, 0.022416383743286132, 0.022379520416259766, 0.022441984176635742, 0.022345727920532226, 0.022383615493774413, 0.022358015060424806, 0.02283417510986328, 0.023199743270874023, 0.023143423080444335, 0.023201791763305665, 0.023218175888061524, 0.02311884880065918, 0.023212032318115236, 0.022437887191772463, 0.022359039306640623, 0.022353919982910156, 0.022406143188476564, 0.022365184783935548, 0.022419456481933595, 0.022337535858154296, 0.02242355155944824, 0.022436864852905275, 0.022551551818847656, 0.023274496078491212, 0.023236608505249022, 0.02245529556274414, 0.02246348762512207, 0.022399999618530272, 0.0224532470703125, 0.02246143913269043, 0.022468608856201173, 0.022418432235717774, 0.02269900894165039, 0.02304614448547363, 0.023161855697631836, 0.023163904190063478, 0.023202816009521485, 0.023154687881469727, 0.022957056045532227, 0.02312294387817383, 0.022436864852905275, 0.022475776672363282, 0.022946815490722656, 0.022717439651489257, 0.0228853759765625, 0.02313216018676758, 0.023171072006225587, 0.02353971290588379, 0.02364723205566406, 0.02329599952697754, 0.023426048278808592, 0.023178239822387696, 0.022424575805664062, 0.02246553611755371, 0.022412288665771486, 0.022373376846313478, 0.022379520416259766, 0.022379520416259766, 0.022437887191772463, 0.022378496170043945, 0.022386688232421875, 0.022390783309936522, 0.022401023864746093, 0.022419456481933595, 0.022371328353881836]",tokens/s,44.062742824149716,,,,,,MB,933.687296,6768.033792,0.0,6138.363904,6060.931072,s,1,6.9070419921875,6.9070419921875,0.0,6.9070419921875,6.9070419921875,6.9070419921875,6.9070419921875,[6.9070419921875],,kWh,1.7008939048644657e-06,8.946338036964958e-07,1.900557076002507e-06,4.496084784563468e-06, bfloat16-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,s,s,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/s/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66947355-76598b0f7f7d19526254e134;d40b0c79-a852-4447-8f2e-112429e3a992) Repository Not Found for url: https://huggingface.co/s/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: s is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-40b,tiiuae/falcon-40b,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ self.load_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 250, in load_model_with_no_weights self.pretrained_model = self.automodel_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 559, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3710, in from_pretrained model = cls(config, *model_args, **model_kwargs) File ""/root/.cache/huggingface/modules/transformers_modules/tiiuae/falcon-40b/4a70170c215b36a3cce4b4253f6d0612bb7d4146/modeling_falcon.py"", line 843, in __init__ self.transformer = FalconModel(config) File ""/root/.cache/huggingface/modules/transformers_modules/tiiuae/falcon-40b/4a70170c215b36a3cce4b4253f6d0612bb7d4146/modeling_falcon.py"", line 650, in __init__ self.h = nn.ModuleList([FalconDecoderLayer(config) for _ in range(config.num_hidden_layers)]) File ""/root/.cache/huggingface/modules/transformers_modules/tiiuae/falcon-40b/4a70170c215b36a3cce4b4253f6d0612bb7d4146/modeling_falcon.py"", line 650, in self.h = nn.ModuleList([FalconDecoderLayer(config) for _ in range(config.num_hidden_layers)]) File ""/root/.cache/huggingface/modules/transformers_modules/tiiuae/falcon-40b/4a70170c215b36a3cce4b4253f6d0612bb7d4146/modeling_falcon.py"", line 420, in __init__ self.mlp = FalconMLP(config) File ""/root/.cache/huggingface/modules/transformers_modules/tiiuae/falcon-40b/4a70170c215b36a3cce4b4253f6d0612bb7d4146/modeling_falcon.py"", line 403, in __init__ self.dense_h_to_4h = FalconLinear(hidden_size, 4 * hidden_size, bias=config.bias) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/linear.py"", line 98, in __init__ self.weight = Parameter(torch.empty((out_features, in_features), **factory_kwargs)) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_device.py"", line 78, in __torch_function__ return func(*args, **kwargs) torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 512.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,google/gemma-7b,google/gemma-7b,cuda,0,42,,,True,,,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.223-212.873.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.0,,,,1.21.4,,,,0.12.0,,,gemma,MB,1529.856,19300.614144,0.0,18654.16704,17596.665856,s,10,0.9555711288452149,0.09555711288452148,0.0002064769309982979,0.09549507141113281,0.09561405258178711,0.09588624229431153,0.09610399406433105,"[0.09539692687988281, 0.095459228515625, 0.0955525131225586, 0.09545030212402343, 0.09615843200683594, 0.09554508972167969, 0.09546492767333985, 0.09552438354492188, 0.09555356597900391, 0.09546575927734376]",tokens/s,2679.0261056690774,kWh,1.128451381157417e-06,6.183320255450474e-07,6.337864858647703e-06,8.084648265350167e-06,tokens/kWh,31664952.09163091,MB,1591.279616,19302.711296,0.0,18656.264192,17596.668416,s,10,27.79764233398437,2.7797642333984376,0.0011756250476684749,2.7794295654296874,2.78119287109375,2.781348876953125,2.781473681640625,"[2.78065576171875, 2.778869873046875, 2.779685546875, 2.7815048828125, 2.77833203125, 2.779173583984375, 2.781158203125, 2.77867822265625, 2.778438232421875, 2.78114599609375]",tokens/s,22.66379257746565,kWh,3.280903113655095e-05,1.7980651807925875e-05,0.00013816583407755148,0.00018895551702202828,tokens/kWh,333411.80502634123,,s,630,27.79577956771853,0.04412028502812461,0.00012884226956761606,0.044107263565063476,0.04424499130249023,0.04430597076416016,0.044676168212890624,"[0.04419891357421875, 0.04454195022583008, 0.043937793731689455, 0.04390092849731445, 0.04398387145996094, 0.043937793731689455, 0.04400332641601563, 0.04419379043579102, 0.04395622253417969, 0.04395315170288086, 0.04405964660644531, 0.04396748733520508, 0.04403814315795898, 0.04408422470092774, 0.04402995300292969, 0.044115966796875, 0.04404121780395508, 0.04401766586303711, 0.04404326248168945, 0.044063743591308595, 0.04406169509887695, 0.04400128173828125, 0.04406784057617188, 0.04409139251708984, 0.04405350494384765, 0.044109825134277345, 0.044065792083740236, 0.044434432983398435, 0.04414976119995117, 0.04406784057617188, 0.044063743591308595, 0.04411699295043945, 0.04407807922363281, 0.04403507232666016, 0.04411391830444336, 0.04404940795898438, 0.04413849639892578, 0.044148735046386715, 0.0440975341796875, 0.04415795135498047, 0.04413030242919922, 0.044088321685791014, 0.04415692901611328, 0.0446382064819336, 0.04420710372924805, 0.04423168182373047, 0.044262401580810545, 0.04422041702270508, 0.0447191047668457, 0.04411391830444336, 0.04418048095703125, 0.044104705810546874, 0.04420608139038086, 0.044131328582763675, 0.04419276809692383, 0.04420710372924805, 0.04421529769897461, 0.04415999984741211, 0.04426444625854492, 0.044170238494873046, 0.044240894317626955, 0.044181503295898435, 0.044281856536865234, 0.04420915222167969, 0.0439736328125, 0.043963390350341795, 0.04407603073120117, 0.04398591995239258, 0.0439736328125, 0.044448833465576175, 0.04413433456420898, 0.04410879898071289, 0.04398080062866211, 0.04398393630981445, 0.043985855102539065, 0.04400332641601563, 0.04403507232666016, 0.04399411010742187, 0.04408115386962891, 0.044044288635253906, 0.043990016937255856, 0.04399923324584961, 0.04398284912109375, 0.04408422470092774, 0.04398489761352539, 0.04400435256958008, 0.044077056884765625, 0.044028926849365234, 0.044036094665527346, 0.0440893440246582, 0.044022785186767575, 0.044044288635253906, 0.044055553436279295, 0.04415795135498047, 0.044333057403564455, 0.04406988906860351, 0.04409446334838867, 0.044028926849365234, 0.04402687835693359, 0.044085247039794925, 0.04405964660644531, 0.04405657577514648, 0.04413235092163086, 0.04410265731811523, 0.04409958267211914, 0.04409036636352539, 0.044093441009521485, 0.0441343994140625, 0.04411699295043945, 0.04413337707519531, 0.04418764877319336, 0.0441343994140625, 0.0441539192199707, 0.044163009643554685, 0.044286975860595705, 0.044200958251953124, 0.044701694488525394, 0.04418867111206055, 0.04419891357421875, 0.04417843246459961, 0.0441712646484375, 0.04416921615600586, 0.044153854370117186, 0.04423680114746094, 0.04419481658935547, 0.04417228698730469, 0.04416614532470703, 0.04390911865234375, 0.043963390350341795, 0.04392959976196289, 0.04393983840942383, 0.04392652893066406, 0.04406476974487305, 0.04405254364013672, 0.04399302291870117, 0.04417433547973633, 0.04410163116455078, 0.04422553634643555, 0.04412313461303711, 0.04396646499633789, 0.04405145645141602, 0.04409139251708984, 0.04400435256958008, 0.04398080062866211, 0.044014591217041016, 0.0439552001953125, 0.044045310974121094, 0.04403814315795898, 0.044052478790283206, 0.04407603073120117, 0.04408115386962891, 0.04401663970947266, 0.04410367965698242, 0.04402175903320313, 0.044085247039794925, 0.0441794548034668, 0.044186622619628906, 0.0441077766418457, 0.04412115097045898, 0.044012481689453126, 0.04413951873779297, 0.044453887939453124, 0.04413644790649414, 0.04404940795898438, 0.044104705810546874, 0.04411801528930664, 0.044109825134277345, 0.044080127716064454, 0.04417433547973633, 0.04405657577514648, 0.04415590286254883, 0.04409036636352539, 0.04415283203125, 0.04416614532470703, 0.04416921615600586, 0.044112895965576174, 0.044158977508544923, 0.044109825134277345, 0.044565502166748046, 0.04415590286254883, 0.04422348785400391, 0.04419686508178711, 0.04426137542724609, 0.0442081298828125, 0.04469145584106445, 0.04422348785400391, 0.044240894317626955, 0.04416921615600586, 0.044267520904541016, 0.04419891357421875, 0.044009471893310545, 0.04392755126953125, 0.0440186882019043, 0.04391939163208008, 0.04399305725097656, 0.043979774475097655, 0.044077056884765625, 0.0440002555847168, 0.044111873626708986, 0.044012542724609374, 0.04393267059326172, 0.04402483367919922, 0.044045310974121094, 0.04406476974487305, 0.04410572814941406, 0.044034046173095705, 0.04408115386962891, 0.04401971054077149, 0.04399718475341797, 0.04399513626098633, 0.044012542724609374, 0.04414976119995117, 0.04468121719360352, 0.04406784057617188, 0.04405759811401367, 0.044112895965576174, 0.044082176208496096, 0.04411801528930664, 0.044093441009521485, 0.0441343994140625, 0.044243968963623044, 0.04417433547973633, 0.044095489501953126, 0.04416716766357422, 0.0440893440246582, 0.0441416015625, 0.04411286544799805, 0.044148735046386715, 0.04420198440551758, 0.04471807861328125, 0.04412928009033203, 0.04411699295043945, 0.044114944458007815, 0.04415999984741211, 0.04413337707519531, 0.04419583892822266, 0.044246017456054686, 0.044243968963623044, 0.04415283203125, 0.04420403289794922, 0.0441712646484375, 0.0442081298828125, 0.04419071960449219, 0.04419583892822266, 0.044286975860595705, 0.0442501106262207, 0.04420608139038086, 0.04483071899414062, 0.044306430816650394, 0.04432179260253906, 0.04423372650146484, 0.044265472412109375, 0.04432691192626953, 0.044006401062011716, 0.0439552001953125, 0.04399718475341797, 0.04396748733520508, 0.04394496154785156, 0.043937793731689455, 0.044050430297851564, 0.044009471893310545, 0.044184574127197264, 0.04398694229125977, 0.04402073669433594, 0.044066814422607424, 0.04412416076660156, 0.04399923324584961, 0.04409036636352539, 0.04402073669433594, 0.0440975341796875, 0.04400230407714844, 0.04397568130493164, 0.0440893440246582, 0.04403302383422852, 0.04400537490844727, 0.044128257751464846, 0.044052478790283206, 0.04408627319335937, 0.04404121780395508, 0.04404019165039062, 0.04406784057617188, 0.044060672760009766, 0.04415590286254883, 0.04416716766357422, 0.04415999984741211, 0.04404633712768555, 0.04404838562011719, 0.044028926849365234, 0.04409036636352539, 0.044050430297851564, 0.044050430297851564, 0.044452865600585936, 0.044111873626708986, 0.044063743591308595, 0.044088321685791014, 0.044055553436279295, 0.04412416076660156, 0.04411084747314453, 0.044096511840820314, 0.04416409683227539, 0.044142593383789064, 0.04409958267211914, 0.04413951873779297, 0.04409139251708984, 0.044219390869140625, 0.04417228698730469, 0.04417331314086914, 0.044219390869140625, 0.04418560028076172, 0.04417228698730469, 0.044200958251953124, 0.04417331314086914, 0.04421836853027344, 0.04421529769897461, 0.044254207611083986, 0.04416204833984375, 0.04392652893066406, 0.04393983840942383, 0.04395315170288086, 0.04396953582763672, 0.04402073669433594, 0.04401049423217773, 0.044093441009521485, 0.04402995300292969, 0.04394291305541992, 0.044028926849365234, 0.04400435256958008, 0.04404326248168945, 0.04395622253417969, 0.044071937561035154, 0.04416211318969727, 0.04400735855102539, 0.04404940795898438, 0.044104705810546874, 0.044022785186767575, 0.044096511840820314, 0.04398694229125977, 0.04404019165039062, 0.04409446334838867, 0.04406272125244141, 0.04398284912109375, 0.044111873626708986, 0.04405452728271484, 0.04412313461303711, 0.044039169311523435, 0.04408729553222656, 0.04411391830444336, 0.044112895965576174, 0.044060672760009766, 0.0440975341796875, 0.044036094665527346, 0.04412723159790039, 0.044055553436279295, 0.0441610221862793, 0.04419385528564453, 0.04420294570922852, 0.0442429428100586, 0.04422348785400391, 0.04412108612060547, 0.0442081298828125, 0.044170238494873046, 0.04421017456054688, 0.04418867111206055, 0.044200958251953124, 0.04409958267211914, 0.04422553634643555, 0.044112895965576174, 0.04428902435302735, 0.04429619216918945, 0.04421017456054688, 0.04421222305297851, 0.044249088287353515, 0.04417331314086914, 0.044229633331298826, 0.04417433547973633, 0.044251136779785157, 0.044260353088378904, 0.04430745697021484, 0.04444672012329102, 0.04394905471801758, 0.04394803237915039, 0.043924480438232424, 0.04394803237915039, 0.04398694229125977, 0.04396851348876953, 0.04415692901611328, 0.04409446334838867, 0.04402175903320313, 0.044055553436279295, 0.04397260665893555, 0.044036094665527346, 0.04416716766357422, 0.0441343994140625, 0.0441077766418457, 0.04402175903320313, 0.04398899078369141, 0.04412313461303711, 0.04401971054077149, 0.044112895965576174, 0.044055553436279295, 0.044063743591308595, 0.04413542556762695, 0.044572673797607425, 0.044083198547363284, 0.04405657577514648, 0.04405452728271484, 0.044115966796875, 0.044060672760009766, 0.044203006744384765, 0.04418764877319336, 0.044440574645996093, 0.04413747024536133, 0.044112895965576174, 0.04404633712768555, 0.04415078353881836, 0.04413644790649414, 0.04413644790649414, 0.04417740631103516, 0.04419891357421875, 0.0441343994140625, 0.04422348785400391, 0.04409446334838867, 0.04424499130249023, 0.044148735046386715, 0.044158977508544923, 0.04419481658935547, 0.044142593383789064, 0.04414976119995117, 0.044153854370117186, 0.04413030242919922, 0.04423168182373047, 0.0441610221862793, 0.04417843246459961, 0.044410945892333985, 0.04424595260620117, 0.04435763168334961, 0.04422246551513672, 0.044251136779785157, 0.04431769561767578, 0.04424806213378906, 0.04424192047119141, 0.04417843246459961, 0.04399411010742187, 0.04394496154785156, 0.044077056884765625, 0.044028926849365234, 0.04394496154785156, 0.043940864562988284, 0.04409139251708984, 0.04400128173828125, 0.04401155090332031, 0.04396233749389648, 0.04401049423217773, 0.044077056884765625, 0.04400128173828125, 0.04397875213623047, 0.04410879898071289, 0.04405350494384765, 0.043990016937255856, 0.044009471893310545, 0.04398080062866211, 0.04412723159790039, 0.04400128173828125, 0.04398694229125977, 0.0440893440246582, 0.04437606430053711, 0.044077056884765625, 0.04403712081909179, 0.044050430297851564, 0.044068862915039066, 0.04406476974487305, 0.04402995300292969, 0.044111873626708986, 0.04407398223876953, 0.043994144439697264, 0.04404630279541016, 0.044055553436279295, 0.04408627319335937, 0.04404838562011719, 0.04407398223876953, 0.04417433547973633, 0.044663806915283204, 0.04411391830444336, 0.0440975341796875, 0.04413030242919922, 0.04415692901611328, 0.04416614532470703, 0.044142593383789064, 0.044211200714111325, 0.04421017456054688, 0.04427679824829102, 0.04412102508544922, 0.04411391830444336, 0.0441712646484375, 0.04412108612060547, 0.04412006378173828, 0.04420915222167969, 0.044181503295898435, 0.04416614532470703, 0.04420710372924805, 0.04417638397216797, 0.044219390869140625, 0.044270591735839845, 0.044286975860595705, 0.04415283203125, 0.04391731262207031, 0.043963390350341795, 0.04395929718017578, 0.04393983840942383, 0.0441077766418457, 0.044028926849365234, 0.04405350494384765, 0.04400844955444336, 0.04396134567260742, 0.04401049423217773, 0.04397772979736328, 0.044052478790283206, 0.04397055816650391, 0.04403814315795898, 0.044085247039794925, 0.04403712081909179, 0.04395929718017578, 0.04404940795898438, 0.0439736328125, 0.04406988906860351, 0.04400435256958008, 0.04404121780395508, 0.044104705810546874, 0.044131328582763675, 0.043993087768554685, 0.0440893440246582, 0.044028926849365234, 0.04441702270507813, 0.04408115386962891, 0.04408422470092774, 0.04425932693481445, 0.04421017456054688, 0.04403712081909179, 0.04412313461303711, 0.04407603073120117, 0.04412108612060547, 0.04403712081909179, 0.044104705810546874, 0.04410572814941406, 0.04412211227416992, 0.04416921615600586, 0.04418048095703125, 0.04405452728271484, 0.04415488052368164, 0.04409241485595703, 0.04416921615600586, 0.04415999984741211, 0.04415283203125, 0.04408422470092774, 0.04419379043579102, 0.044128257751464846, 0.0442429428100586, 0.04415590286254883, 0.044211200714111325, 0.044189697265625, 0.044246017456054686, 0.04415078353881836, 0.04422553634643555, 0.044165119171142575, 0.04424192047119141, 0.04417331314086914, 0.044229633331298826, 0.044249088287353515, 0.043921409606933595, 0.04394598388671875, 0.04393164825439453, 0.04392345428466797, 0.04403814315795898, 0.044930049896240234, 0.044181503295898435, 0.04403302383422852, 0.044045310974121094, 0.044014591217041016, 0.04394598388671875, 0.04402483367919922, 0.04399513626098633, 0.04399411010742187, 0.04409958267211914, 0.04401971054077149, 0.043979774475097655, 0.04400332641601563, 0.04399411010742187, 0.044006401062011716, 0.04402175903320313, 0.04404838562011719, 0.04424499130249023, 0.04409446334838867, 0.04406784057617188, 0.04407910537719727, 0.0440709114074707, 0.04413747024536133, 0.04406476974487305, 0.044104705810546874, 0.04417740631103516, 0.044305408477783206, 0.044147712707519535, 0.04412006378173828, 0.044077056884765625, 0.0441343994140625, 0.044096511840820314, 0.04415488052368164, 0.04424499130249023, 0.04418560028076172, 0.04413235092163086, 0.044148735046386715, 0.04410675048828125, 0.044165119171142575, 0.044609535217285154, 0.04419481658935547, 0.04432691192626953, 0.04417536163330078, 0.044254207611083986, 0.044211200714111325, 0.04416614532470703, 0.044306430816650394, 0.044306430816650394, 0.04424499130249023, 0.04424806213378906, 0.044230655670166014, 0.04420505523681641, 0.044221439361572266, 0.04416409683227539, 0.04423168182373047, 0.04421222305297851, 0.044252159118652344]",tokens/s,22.66531141769703,,,,,,MB,934.944768,19279.642624,0.0,18649.972736,18648.241152,s,1,6.9921083984375,6.9921083984375,0.0,6.9921083984375,6.9921083984375,6.9921083984375,6.9921083984375,[6.9921083984375],,kWh,2.41459732568943e-06,1.2876565152083306e-06,8.864173757991423e-06,1.2566427598889183e-05, bfloat16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-125m,,cuda,0,42,,,,,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,d35829e539df8480b726c647eeabf91e41eae047,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,,MB,1168.5888,1019.74016,0.0,373.293056,302.523392,s,23,0.19000934505462644,0.008261275871940279,0.0007920114248349487,0.008010656356811523,0.008298675346374513,0.008638550853729247,0.011176230468750005,"[0.01188156795501709, 0.008675488471984863, 0.008249183654785157, 0.007968255996704102, 0.007941023826599121, 0.007941728115081786, 0.00795145606994629, 0.008010656356811523, 0.008005184173583984, 0.00790112018585205, 0.008009856224060058, 0.007931551933288573, 0.0079901762008667, 0.008261919975280762, 0.008121184349060059, 0.00804096031188965, 0.007987391948699952, 0.00826643180847168, 0.008268927574157715, 0.008000703811645507, 0.008251520156860351, 0.008046943664550782, 0.008306112289428711]",tokens/s,30987.949557466447,kWh,9.650255059018548e-08,5.2878837633376375e-08,2.207636116143768e-07,3.7014499983793865e-07,tokens/kWh,691620851.5908226,MB,1168.5888,1019.74016,0.0,373.293056,312.415744,s,23,10.075949432373047,0.43808475792926294,0.00941723845187448,0.4384620666503906,0.44165304565429686,0.44281994323730467,0.4696373205566407,"[0.47716641235351565, 0.4414251403808594, 0.43919485473632813, 0.4317526550292969, 0.4301205749511719, 0.43231777954101563, 0.44171002197265624, 0.43642147827148436, 0.4300735168457031, 0.43406436157226563, 0.4279365234375, 0.4340263977050781, 0.44080706787109375, 0.433282958984375, 0.4332001342773438, 0.4310262451171875, 0.4384620666503906, 0.43951202392578126, 0.44036474609375, 0.4429432678222656, 0.4407447509765625, 0.43988232421875, 0.43951412963867187]",tokens/s,143.8077880129592,kWh,5.09665186246492e-06,2.7927161006653693e-06,8.918048047256495e-06,1.6807416010386787e-05,tokens/kWh,3748345.370940229,,s,1448,10.244387838840469,0.007074853479862212,0.0010543028958638788,0.006966271877288818,0.007110963058471679,0.007475507140159607,0.015209052209854126,"[0.008376319885253907, 0.007783423900604248, 0.007759871959686279, 0.007724031925201416, 0.007566336154937744, 0.007345151901245117, 0.007518208026885987, 0.007511040210723877, 0.007442431926727295, 0.007398399829864502, 0.0074967041015625, 0.007469056129455566, 0.007488512039184571, 0.007502848148345947, 0.007546879768371582, 0.007494656085968018, 0.007460864067077637, 0.007582719802856445, 0.00754585599899292, 0.007565408229827881, 0.007557024002075195, 0.007465983867645264, 0.007539711952209473, 0.007524352073669433, 0.007529471874237061, 0.007413760185241699, 0.007457791805267334, 0.007601151943206787, 0.00744652795791626, 0.007490560054779053, 0.007557119846343994, 0.007715839862823487, 0.007705599784851074, 0.007590911865234375, 0.007600128173828125, 0.007515135765075683, 0.007515135765075683, 0.0075673599243164065, 0.007350272178649903, 0.00749567985534668, 0.0074291200637817386, 0.0074301438331604, 0.007476223945617676, 0.007502848148345947, 0.007602176189422607, 0.007357439994812012, 0.00738918399810791, 0.007481344223022461, 0.007434239864349365, 0.0074741759300231934, 0.0074670081138610836, 0.007525375843048096, 0.007610367774963379, 0.0075642881393432615, 0.0076605439186096195, 0.007602176189422607, 0.007589920043945313, 0.007448544025421142, 0.007498816013336181, 0.007378880023956299, 0.007372799873352051, 0.00719974422454834, 0.016752639770507814, 0.007117824077606201, 0.0070594558715820314, 0.007147520065307617, 0.007138304233551026, 0.007299071788787842, 0.007222271919250488, 0.007202816009521484, 0.007394303798675537, 0.007517183780670166, 0.007285791873931885, 0.007130112171173096, 0.007139296054840088, 0.007259136199951172, 0.007207935810089112, 0.007185440063476562, 0.007046112060546875, 0.006985727787017822, 0.007047167778015137, 0.007027711868286133, 0.007028736114501953, 0.007117824077606201, 0.00709222412109375, 0.0069918718338012695, 0.006974463939666748, 0.0069816322326660156, 0.007009280204772949, 0.006978559970855713, 0.006993919849395752, 0.007069695949554444, 0.007006207942962647, 0.0070256638526916505, 0.0069847040176391605, 0.007008255958557129, 0.007014400005340577, 0.007080992221832275, 0.006968287944793701, 0.007054336071014404, 0.007024640083312988, 0.006993919849395752, 0.006862847805023193, 0.006755328178405762, 0.006797311782836914, 0.006783999919891357, 0.0067758078575134275, 0.006811647891998291, 0.006796288013458252, 0.006756351947784424, 0.00678604793548584, 0.006922239780426025, 0.006792191982269287, 0.00677785587310791, 0.006788095951080322, 0.006772736072540283, 0.006778880119323731, 0.0068577280044555666, 0.00678604793548584, 0.006749184131622315, 0.006790143966674805, 0.006776832103729248, 0.006780928134918213, 0.006969344139099121, 0.006987775802612305, 0.01571020793914795, 0.0069959678649902345, 0.006931456089019775, 0.006986752033233643, 0.007019519805908203, 0.006985727787017822, 0.006967296123504638, 0.006994944095611572, 0.007045119762420654, 0.007017471790313721, 0.006978559970855713, 0.007014400005340577, 0.007013376235961914, 0.0070348801612854, 0.006965248107910156, 0.0070891518592834475, 0.00709119987487793, 0.00702569580078125, 0.0070184640884399414, 0.007013408184051514, 0.0070471358299255375, 0.0069928960800170895, 0.007017471790313721, 0.007005184173583984, 0.007005184173583984, 0.007006207942962647, 0.007014400005340577, 0.0069847040176391605, 0.007107583999633789, 0.007036928176879883, 0.007008255958557129, 0.006975488185882568, 0.0071209602355957035, 0.007091135978698731, 0.006961152076721191, 0.0070266880989074704, 0.007019519805908203, 0.0070225920677185055, 0.007021567821502686, 0.007016448020935059, 0.007173120021820068, 0.007028736114501953, 0.007010303974151612, 0.007062528133392334, 0.007068672180175781, 0.00709222412109375, 0.006894591808319092, 0.006782976150512696, 0.006792191982269287, 0.006783999919891357, 0.006741024017333984, 0.006812640190124512, 0.006774784088134766, 0.006806528091430664, 0.006782976150512696, 0.006799359798431396, 0.006797311782836914, 0.006750207901000976, 0.006849535942077637, 0.00690176010131836, 0.006821951866149902, 0.006796224117279053, 0.006796319961547851, 0.015191007614135742, 0.00676966381072998, 0.006833151817321777, 0.006816768169403077, 0.006789120197296142, 0.006774784088134766, 0.006853631973266602, 0.0067707200050354, 0.006784992218017578, 0.006797311782836914, 0.0068055038452148435, 0.0067686400413513184, 0.0068280320167541505, 0.006856704235076904, 0.0067870721817016606, 0.006770688056945801, 0.006810624122619629, 0.006788095951080322, 0.006780928134918213, 0.006815743923187256, 0.006808576107025147, 0.006767615795135498, 0.006814720153808594, 0.006820864200592041, 0.0067983360290527345, 0.006803455829620361, 0.006811647891998291, 0.006778880119323731, 0.006802432060241699, 0.006929408073425293, 0.007108607769012451, 0.006823935985565186, 0.0067870721817016606, 0.006855679988861084, 0.006809599876403808, 0.006808576107025147, 0.00682700777053833, 0.006811647891998291, 0.0068392958641052244, 0.006814720153808594, 0.006801472187042236, 0.006781888008117676, 0.006778880119323731, 0.006778880119323731, 0.006833151817321777, 0.0067983360290527345, 0.006783999919891357, 0.006802432060241699, 0.006771743774414063, 0.0070184640884399414, 0.007154687881469727, 0.007051263809204102, 0.006994944095611572, 0.007067647933959961, 0.007007232189178467, 0.007166016101837158, 0.007027647972106934, 0.006976511955261231, 0.006975488185882568, 0.006800415992736816, 0.006769631862640381, 0.006790143966674805, 0.0068055038452148435, 0.015136768341064453, 0.006783999919891357, 0.006788095951080322, 0.006797311782836914, 0.006751232147216797, 0.006792191982269287, 0.006803455829620361, 0.006762495994567871, 0.00678604793548584, 0.006773759841918945, 0.00692633581161499, 0.007003136157989502, 0.006986752033233643, 0.007011328220367432, 0.007023615837097168, 0.007024640083312988, 0.006989823818206787, 0.00704307222366333, 0.006978559970855713, 0.006979584217071533, 0.007010303974151612, 0.006848512172698974, 0.006793216228485107, 0.006751232147216797, 0.006810624122619629, 0.0067645440101623535, 0.006766592025756836, 0.0067870721817016606, 0.006790143966674805, 0.00674508810043335, 0.006780928134918213, 0.006774784088134766, 0.006819839954376221, 0.006782976150512696, 0.006785024166107178, 0.0067870721817016606, 0.006762495994567871, 0.006780928134918213, 0.006761472225189209, 0.006782976150512696, 0.006796288013458252, 0.006780928134918213, 0.0067420477867126466, 0.006779871940612793, 0.006820864200592041, 0.006791168212890625, 0.006759424209594727, 0.006780928134918213, 0.006739007949829102, 0.006787007808685303, 0.006785056114196777, 0.006813663959503174, 0.006747136116027832, 0.006756351947784424, 0.006780928134918213, 0.0067573761940002445, 0.006841343879699707, 0.0067686400413513184, 0.006872064113616944, 0.006756351947784424, 0.006811647891998291, 0.0067686400413513184, 0.006830080032348633, 0.015417344093322754, 0.007006207942962647, 0.007007232189178467, 0.006938687801361084, 0.0070430078506469726, 0.007019519805908203, 0.007064608097076416, 0.007019487857818604, 0.007019519805908203, 0.007015423774719238, 0.00698367977142334, 0.006965248107910156, 0.0069550080299377445, 0.006855679988861084, 0.006700032234191895, 0.006672383785247803, 0.006823935985565186, 0.006725632190704346, 0.006677504062652588, 0.00672054386138916, 0.006726655960083008, 0.006748127937316895, 0.006803455829620361, 0.00678604793548584, 0.006773759841918945, 0.0068055038452148435, 0.006762495994567871, 0.006762495994567871, 0.006821887969970703, 0.006812672138214112, 0.006781951904296875, 0.006774784088134766, 0.006758399963378906, 0.006774784088134766, 0.00683622407913208, 0.006834176063537598, 0.006754303932189941, 0.006793216228485107, 0.006783999919891357, 0.006814720153808594, 0.006751232147216797, 0.0067645440101623535, 0.006772736072540283, 0.006763519763946534, 0.006752255916595459, 0.00676966381072998, 0.006931456089019775, 0.007070720195770264, 0.006976511955261231, 0.006993919849395752, 0.006967296123504638, 0.0069632000923156735, 0.006948863983154297, 0.006990848064422607, 0.006969344139099121, 0.0069918718338012695, 0.006892543792724609, 0.006810624122619629, 0.00676966381072998, 0.006781951904296875, 0.0067758078575134275, 0.006777919769287109, 0.006734784126281738, 0.015156224250793457, 0.006793216228485107, 0.0067840957641601566, 0.007388063907623291, 0.007328767776489258, 0.007408639907836914, 0.00754585599899292, 0.0071526398658752445, 0.007030784130096435, 0.006965248107910156, 0.007050240039825439, 0.007028736114501953, 0.007003168106079102, 0.007015391826629639, 0.007047167778015137, 0.007065599918365479, 0.007041024208068848, 0.007591936111450195, 0.006964223861694336, 0.006987775802612305, 0.006986752033233643, 0.00694271993637085, 0.006949888229370117, 0.007057407855987549, 0.0069847040176391605, 0.006985727787017822, 0.007002111911773682, 0.006986752033233643, 0.006961152076721191, 0.006947840213775635, 0.006976511955261231, 0.007000063896179199, 0.006980607986450196, 0.006941696166992187, 0.006989823818206787, 0.006976511955261231, 0.00698367977142334, 0.006939712047576904, 0.006996928215026855, 0.00748748779296875, 0.007041024208068848, 0.006964223861694336, 0.006951935768127441, 0.006980607986450196, 0.007155712127685547, 0.006988800048828125, 0.006960127830505371, 0.007027711868286133, 0.006968319892883301, 0.0070225920677185055, 0.006952960014343262, 0.007039999961853028, 0.007002111911773682, 0.006970367908477783, 0.006939648151397705, 0.006949888229370117, 0.006780928134918213, 0.006754303932189941, 0.006765567779541016, 0.006763519763946534, 0.006760479927062989, 0.006759391784667969, 0.006776832103729248, 0.015210495948791505, 0.007038976192474365, 0.007014400005340577, 0.0070225920677185055, 0.007041024208068848, 0.006975488185882568, 0.007036928176879883, 0.006987775802612305, 0.006968319892883301, 0.007068704128265381, 0.007023583889007568, 0.007011328220367432, 0.007054336071014404, 0.0069632000923156735, 0.007023615837097168, 0.007011328220367432, 0.006999040126800537, 0.0069283838272094726, 0.0069928960800170895, 0.007015423774719238, 0.007007232189178467, 0.00695091199874878, 0.007007232189178467, 0.006987775802612305, 0.00697760009765625, 0.006938560009002686, 0.006975488185882568, 0.007011328220367432, 0.007009280204772949, 0.0069621758460998535, 0.0069621758460998535, 0.006990848064422607, 0.0069816322326660156, 0.006957056045532227, 0.006977536201477051, 0.006968319892883301, 0.007017471790313721, 0.007039999961853028, 0.007003136157989502, 0.006996992111206054, 0.006842368125915528, 0.006765567779541016, 0.006917119979858399, 0.006812672138214112, 0.006759424209594727, 0.006785024166107178, 0.006794271945953369, 0.006789087772369385, 0.006765567779541016, 0.006892543792724609, 0.006859776020050049, 0.006795328140258789, 0.006775743961334228, 0.006791168212890625, 0.006739967823028564, 0.006816768169403077, 0.006785024166107178, 0.00676966381072998, 0.0067717118263244626, 0.006783999919891357, 0.006759424209594727, 0.006754303932189941, 0.006825984001159668, 0.015217663764953614, 0.0067573761940002445, 0.006762495994567871, 0.006793216228485107, 0.0067768640518188475, 0.006770688056945801, 0.0067870402336120605, 0.006766592025756836, 0.006763519763946534, 0.006793216228485107, 0.006821951866149902, 0.006732736110687256, 0.006818816184997558, 0.006803487777709961, 0.006769631862640381, 0.006765567779541016, 0.006783999919891357, 0.006763519763946534, 0.006731776237487793, 0.006923264026641846, 0.006986752033233643, 0.007020544052124023, 0.007071743965148926, 0.00698367977142334, 0.00695091199874878, 0.006968319892883301, 0.006952960014343262, 0.00704307222366333, 0.006887423992156983, 0.006748159885406494, 0.006761472225189209, 0.00678710412979126, 0.006746079921722412, 0.00676966381072998, 0.006772736072540283, 0.006755328178405762, 0.006763519763946534, 0.006799359798431396, 0.006766592025756836, 0.006770688056945801, 0.00679423999786377, 0.006760479927062989, 0.0067645120620727535, 0.007035903930664063, 0.006994944095611572, 0.007045152187347412, 0.0068853440284729, 0.006858751773834228, 0.006811647891998291, 0.006762495994567871, 0.006755392074584961, 0.006777791976928711, 0.006738944053649902, 0.006740992069244385, 0.006767615795135498, 0.006779903888702392, 0.006770688056945801, 0.00677785587310791, 0.006789120197296142, 0.006752255916595459, 0.006788095951080322, 0.0067758078575134275, 0.0067758078575134275, 0.015070207595825195, 0.0068392958641052244, 0.006770688056945801, 0.006799359798431396, 0.00677785587310791, 0.0067645440101623535, 0.0068351998329162595, 0.00704204797744751, 0.007008255958557129, 0.0070042238235473635, 0.006937535762786865, 0.0069928960800170895, 0.007009280204772949, 0.007028736114501953, 0.006947840213775635, 0.00698367977142334, 0.006980607986450196, 0.006776832103729248, 0.006746111869812011, 0.006773759841918945, 0.006831103801727295, 0.006761472225189209, 0.006797311782836914, 0.0068280320167541505, 0.006774784088134766, 0.006759424209594727, 0.006860799789428711, 0.007112703800201416, 0.0069550080299377445, 0.007067647933959961, 0.006993919849395752, 0.007032832145690918, 0.006966335773468018, 0.006994880199432373, 0.006980607986450196, 0.007012351989746094, 0.0069550080299377445, 0.006994944095611572, 0.006996992111206054, 0.007002111911773682, 0.006929408073425293, 0.006996992111206054, 0.006958079814910889, 0.006979584217071533, 0.006961184024810791, 0.006993887901306152, 0.0069959678649902345, 0.00688640022277832, 0.006767615795135498, 0.0067686400413513184, 0.0067870721817016606, 0.006729728221893311, 0.006758399963378906, 0.006797311782836914, 0.006752255916595459, 0.006758399963378906, 0.006768703937530518, 0.006770624160766602, 0.006765567779541016, 0.006780928134918213, 0.006770688056945801, 0.006744063854217529, 0.006799359798431396, 0.015263744354248047, 0.006760447978973389, 0.006760447978973389, 0.006790143966674805, 0.006767615795135498, 0.006772736072540283, 0.006801407814025879, 0.0067983360290527345, 0.006765567779541016, 0.006767648220062256, 0.006779871940612793, 0.0067983360290527345, 0.0067983360290527345, 0.006813695907592773, 0.00678604793548584, 0.006788095951080322, 0.006909952163696289, 0.006810624122619629, 0.006752255916595459, 0.006793216228485107, 0.006811647891998291, 0.0067573761940002445, 0.006783999919891357, 0.006796288013458252, 0.006755328178405762, 0.0067738242149353025, 0.006792128086090088, 0.006809599876403808, 0.006744063854217529, 0.00677785587310791, 0.0067983360290527345, 0.006767615795135498, 0.006781951904296875, 0.006774784088134766, 0.006792191982269287, 0.006752255916595459, 0.006797311782836914, 0.00675328016281128, 0.006841343879699707, 0.006797311782836914, 0.006834176063537598, 0.006834176063537598, 0.006823935985565186, 0.00682700777053833, 0.006772736072540283, 0.006778880119323731, 0.006797311782836914, 0.0067686400413513184, 0.006744063854217529, 0.006780928134918213, 0.006778880119323731, 0.006751232147216797, 0.006795263767242431, 0.0068280320167541505, 0.006748159885406494, 0.006788095951080322, 0.006776832103729248, 0.006763519763946534, 0.006766592025756836, 0.006785024166107178, 0.006770688056945801, 0.006737919807434082, 0.006793216228485107, 0.015129599571228027, 0.007038976192474365, 0.006933504104614257, 0.007116799831390381, 0.007024640083312988, 0.006990848064422607, 0.006988800048828125, 0.006970367908477783, 0.006989823818206787, 0.006978559970855713, 0.006967296123504638, 0.007013376235961914, 0.007041024208068848, 0.0069704318046569826, 0.006979519844055176, 0.007004159927368164, 0.007016448020935059, 0.006964223861694336, 0.006937600135803222, 0.006941696166992187, 0.006797311782836914, 0.0067717118263244626, 0.006822912216186523, 0.006795263767242431, 0.006750207901000976, 0.006778880119323731, 0.006785024166107178, 0.00677785587310791, 0.006761472225189209, 0.006761472225189209, 0.006767615795135498, 0.0067717118263244626, 0.006783999919891357, 0.0067983360290527345, 0.006737919807434082, 0.006791168212890625, 0.0067686400413513184, 0.006792191982269287, 0.006754303932189941, 0.006808576107025147, 0.006758399963378906, 0.006770688056945801, 0.006778880119323731, 0.006807551860809326, 0.0067758078575134275, 0.006978559970855713, 0.007409664154052734, 0.007066624164581299, 0.006933504104614257, 0.007077888011932373, 0.0070830078125, 0.006882304191589355, 0.006904831886291504, 0.006853631973266602, 0.006815743923187256, 0.006814720153808594, 0.006907904148101806, 0.006796351909637451, 0.006788032054901123, 0.0067573761940002445, 0.006822912216186523, 0.006820864200592041, 0.00677785587310791, 0.015207424163818359, 0.006807551860809326, 0.006818816184997558, 0.0067686400413513184, 0.006791232109069824, 0.006775743961334228, 0.006758463859558105, 0.006798272132873535, 0.006873087882995605, 0.006793216228485107, 0.007070720195770264, 0.007020544052124023, 0.007005184173583984, 0.006953983783721924, 0.0069816322326660156, 0.007038976192474365, 0.0070225920677185055, 0.007044095993041993, 0.006975488185882568, 0.006980607986450196, 0.007039999961853028, 0.006988800048828125, 0.007020544052124023, 0.007030784130096435, 0.007076863765716553, 0.007093247890472412, 0.006961152076721191, 0.0069959678649902345, 0.0069959678649902345, 0.006975488185882568, 0.0069621758460998535, 0.007023615837097168, 0.007012351989746094, 0.007053376197814942, 0.006973375797271728, 0.006978559970855713, 0.007068672180175781, 0.007011328220367432, 0.0069621758460998535, 0.007080959796905518, 0.006985727787017822, 0.006979584217071533, 0.006961152076721191, 0.0069816322326660156, 0.00698367977142334, 0.007017471790313721, 0.006966271877288818, 0.0069959678649902345, 0.007107583999633789, 0.006998015880584717, 0.006965248107910156, 0.007027711868286133, 0.007016448020935059, 0.007816192150115966, 0.007049215793609619, 0.007072768211364746, 0.007057407855987549, 0.007000063896179199, 0.006986752033233643, 0.006969344139099121, 0.0070891518592834475, 0.006990848064422607, 0.007032832145690918, 0.015631360054016115, 0.007007232189178467, 0.006974463939666748, 0.00704307222366333, 0.006947840213775635, 0.007311359882354736, 0.007070720195770264, 0.007038976192474365, 0.007023615837097168, 0.006980607986450196, 0.0070266880989074704, 0.007013376235961914, 0.007029759883880615, 0.0069928960800170895, 0.007057407855987549, 0.007050240039825439, 0.0070225920677185055, 0.006964223861694336, 0.006919167995452881, 0.0067348480224609375, 0.006713344097137451, 0.006702079772949219, 0.006737919807434082, 0.006756351947784424, 0.006783999919891357, 0.006785024166107178, 0.006873087882995605, 0.006752255916595459, 0.006804480075836182, 0.006801407814025879, 0.006762495994567871, 0.006778880119323731, 0.0067870721817016606, 0.006797311782836914, 0.0067758078575134275, 0.006822912216186523, 0.006761472225189209, 0.006762495994567871, 0.00679423999786377, 0.006791168212890625, 0.007384064197540283, 0.007155712127685547, 0.006834176063537598, 0.006799359798431396, 0.00678604793548584, 0.006762495994567871, 0.00678604793548584, 0.00679423999786377, 0.006758399963378906, 0.006785024166107178, 0.006797311782836914, 0.006772736072540283, 0.006788095951080322, 0.00679423999786377, 0.00676966381072998, 0.006792191982269287, 0.006808576107025147, 0.0067870721817016606, 0.006749184131622315, 0.006802432060241699, 0.006796288013458252, 0.0067983360290527345, 0.006871039867401123, 0.015475775718688965, 0.006807487964630127, 0.006765567779541016, 0.006795263767242431, 0.006814720153808594, 0.006840320110321045, 0.006873087882995605, 0.006873087882995605, 0.006820864200592041, 0.006872064113616944, 0.006830080032348633, 0.006809599876403808, 0.0067717118263244626, 0.006819839954376221, 0.006853631973266602, 0.006797311782836914, 0.006858751773834228, 0.006858751773834228, 0.0069550080299377445, 0.006739967823028564, 0.006804480075836182, 0.006781951904296875, 0.0067573761940002445, 0.006791168212890625, 0.006758399963378906, 0.006782976150512696, 0.006761504173278809, 0.006810592174530029, 0.006789120197296142, 0.006749184131622315, 0.006763519763946534, 0.006780928134918213, 0.006783999919891357, 0.006801407814025879, 0.006838272094726563, 0.00677785587310791, 0.006770688056945801, 0.00679423999786377, 0.006856704235076904, 0.00678604793548584, 0.0067758078575134275, 0.00675328016281128, 0.0067573761940002445, 0.006762495994567871, 0.006765567779541016, 0.006778880119323731, 0.00677894401550293, 0.006795199871063232, 0.007123968124389648, 0.0070891518592834475, 0.007277567863464355, 0.0072120318412780765, 0.007577600002288819, 0.007151616096496582, 0.007010303974151612, 0.007078911781311035, 0.0070860800743103025, 0.007039999961853028, 0.007027711868286133, 0.006975488185882568, 0.006988800048828125, 0.006801407814025879, 0.006772736072540283, 0.015268927574157714, 0.0068023681640625, 0.006751232147216797, 0.006770751953125, 0.006799295902252197, 0.006936575889587402, 0.007035903930664063, 0.007024640083312988, 0.007005184173583984, 0.007045119762420654, 0.006975488185882568, 0.007012351989746094, 0.0069959678649902345, 0.0069816322326660156, 0.006982656002044678, 0.006961152076721191, 0.007002111911773682, 0.00695091199874878, 0.006939648151397705, 0.0069928960800170895, 0.006944767951965332, 0.0067041277885437015, 0.006709248065948486, 0.0068392958641052244, 0.00676966381072998, 0.006759424209594727, 0.006789120197296142, 0.0067983360290527345, 0.006871039867401123, 0.0067758078575134275, 0.00678604793548584, 0.006792191982269287, 0.006837247848510742, 0.0068055038452148435, 0.006841343879699707, 0.006770688056945801, 0.006792191982269287, 0.0067983360290527345, 0.006785088062286377, 0.0067788162231445315, 0.006772736072540283, 0.00675328016281128, 0.0067768959999084475, 0.006804416179656983, 0.006773759841918945, 0.006748159885406494, 0.00676966381072998, 0.00678604793548584, 0.006763519763946534, 0.006813695907592773, 0.006782976150512696, 0.0067717118263244626, 0.006821887969970703, 0.006801407814025879, 0.006776832103729248, 0.006754303932189941, 0.00676966381072998, 0.006795263767242431, 0.006752255916595459, 0.006792191982269287, 0.0067963838577270504, 0.006778783798217773, 0.0067870721817016606, 0.015178751945495606, 0.006781951904296875, 0.00675328016281128, 0.006800384044647217, 0.006843391895294189, 0.006749184131622315, 0.0067993922233581545, 0.006800352096557617, 0.006755328178405762, 0.006729728221893311, 0.006763552188873291, 0.006791135787963867, 0.006816768169403077, 0.006801407814025879, 0.0068618240356445315, 0.006967296123504638, 0.007064576148986816, 0.007090176105499267, 0.007058432102203369, 0.006965248107910156, 0.006977536201477051, 0.007014400005340577, 0.007017471790313721, 0.007038976192474365, 0.0069621758460998535, 0.007001088142395019, 0.006979584217071533, 0.007005184173583984, 0.006945792198181152, 0.006989823818206787, 0.006972415924072266, 0.006974463939666748, 0.0069816322326660156, 0.007032832145690918, 0.007013376235961914, 0.006980671882629395, 0.006985663890838623, 0.007053311824798584, 0.007044095993041993, 0.006978559970855713, 0.007062528133392334, 0.007001088142395019, 0.007007232189178467, 0.006975488185882568, 0.006980607986450196, 0.007022624015808106, 0.00699286413192749, 0.007019584178924561, 0.006980544090270996, 0.007015423774719238, 0.007021567821502686, 0.006990848064422607, 0.006989823818206787, 0.007000063896179199, 0.007009280204772949, 0.007014400005340577, 0.006971392154693603, 0.007011328220367432, 0.007005184173583984, 0.007004159927368164, 0.007009280204772949, 0.006988800048828125, 0.007071743965148926, 0.0157009916305542, 0.007169023990631103, 0.007023615837097168, 0.007168000221252442, 0.0071485438346862796, 0.0070256638526916505, 0.0069959678649902345, 0.00703385591506958, 0.0070225920677185055, 0.006956031799316406, 0.006929408073425293, 0.0069928960800170895, 0.007006207942962647, 0.007000063896179199, 0.006987808227539062, 0.006995935916900634, 0.007012351989746094, 0.006920191764831543, 0.006944767951965332, 0.007001088142395019, 0.006993919849395752, 0.006959104061126709, 0.007184383869171143, 0.007085055828094483, 0.007005184173583984, 0.007036928176879883, 0.006952960014343262, 0.007010303974151612, 0.00698367977142334, 0.006996992111206054, 0.007079936027526855, 0.006959104061126709, 0.006842368125915528, 0.006783999919891357, 0.0067758078575134275, 0.0067758078575134275, 0.006756351947784424, 0.0067686400413513184, 0.006790143966674805, 0.006780928134918213, 0.006731776237487793, 0.006999040126800537, 0.0069847040176391605, 0.006985727787017822, 0.006952960014343262, 0.006976511955261231, 0.006982656002044678, 0.006939648151397705, 0.006957056045532227, 0.007046144008636474, 0.00698367977142334, 0.006947840213775635, 0.006973440170288086, 0.006987775802612305, 0.007008255958557129, 0.006957056045532227, 0.006953983783721924, 0.007009280204772949, 0.0069632000923156735, 0.006966271877288818, 0.006961152076721191, 0.0069928960800170895, 0.006988800048828125, 0.01569279956817627, 0.006936639785766602, 0.006968255996704102, 0.0070563840866088865, 0.007008255958557129, 0.006976511955261231, 0.007000063896179199, 0.007019519805908203, 0.007129087924957276, 0.006978559970855713, 0.007060480117797851, 0.00703385591506958, 0.007023615837097168, 0.007023615837097168, 0.006989823818206787, 0.007051263809204102, 0.007036928176879883, 0.006972415924072266, 0.007009280204772949, 0.006987775802612305, 0.0069928960800170895, 0.0069847040176391605, 0.006988800048828125, 0.00709222412109375, 0.007006207942962647, 0.007024640083312988, 0.0069816322326660156, 0.0070348801612854, 0.007050240039825439, 0.0071198720932006835, 0.006998015880584717, 0.007068672180175781, 0.007037951946258545, 0.007095295906066895, 0.0070011520385742185, 0.007003071784973144, 0.007004159927368164, 0.006998015880584717, 0.0069918718338012695, 0.007038976192474365, 0.007021567821502686, 0.006996992111206054, 0.006979584217071533, 0.007016511917114257, 0.007009280204772949, 0.006985663890838623, 0.006972415924072266, 0.007106560230255127, 0.007008255958557129, 0.006994944095611572, 0.006967296123504638, 0.007014400005340577, 0.006976511955261231, 0.006825984001159668, 0.00678604793548584, 0.006815743923187256, 0.00678604793548584, 0.00676966381072998, 0.00694271993637085, 0.00683622407913208, 0.006801407814025879, 0.006760447978973389, 0.006830080032348633, 0.015218688011169433, 0.006790143966674805, 0.006801407814025879, 0.006820864200592041, 0.0068689918518066405, 0.00677785587310791, 0.006825984001159668, 0.006783999919891357, 0.006766592025756836, 0.006803455829620361, 0.00677894401550293, 0.00679417610168457, 0.006969344139099121, 0.007493631839752197, 0.007039999961853028, 0.007004159927368164, 0.007018496036529541, 0.00719155216217041, 0.006993919849395752, 0.006990848064422607, 0.007051263809204102, 0.007020607948303222, 0.007022528171539306, 0.006985727787017822, 0.0071823358535766605, 0.00709939193725586, 0.007008255958557129, 0.007007232189178467, 0.006968319892883301, 0.00703385591506958, 0.006994944095611572, 0.007038976192474365, 0.00740556812286377, 0.007042079925537109, 0.007548895835876465, 0.007911424160003662, 0.007294047832489014, 0.007163807868957519, 0.007087103843688965, 0.007004159927368164, 0.00724070405960083, 0.0070594558715820314, 0.0069847040176391605, 0.006999104022979736, 0.006993855953216553, 0.007088128089904785, 0.007006207942962647, 0.0069928960800170895, 0.0070522880554199216, 0.007009280204772949, 0.007001088142395019, 0.006977536201477051, 0.007012351989746094, 0.007009280204772949, 0.007020544052124023, 0.006964223861694336, 0.007016448020935059, 0.007001088142395019, 0.007013376235961914, 0.0069621758460998535, 0.006980607986450196, 0.006999040126800537, 0.006972415924072266, 0.01568051242828369, 0.006966271877288818, 0.006999040126800537, 0.007008255958557129, 0.007003136157989502, 0.007006207942962647, 0.006998047828674316, 0.007010272026062012, 0.006976511955261231, 0.006951935768127441, 0.007036992073059082, 0.006988736152648925, 0.006977536201477051, 0.006971392154693603, 0.006999040126800537, 0.006999040126800537, 0.006908927917480469, 0.00694374418258667, 0.00704204797744751, 0.006977536201477051, 0.006989823818206787, 0.0069918718338012695, 0.007050240039825439, 0.007002111911773682, 0.006951935768127441, 0.006965248107910156, 0.0069847040176391605, 0.006973440170288086, 0.00704307222366333, 0.007021599769592285, 0.007188447952270508, 0.006990848064422607, 0.006944767951965332, 0.006971392154693603, 0.006998015880584717, 0.007020544052124023, 0.007047167778015137, 0.0069847040176391605, 0.007006207942962647, 0.007024640083312988, 0.00698367977142334, 0.0069847040176391605, 0.0070266880989074704, 0.007018496036529541, 0.007065599918365479, 0.007049215793609619, 0.006989823818206787, 0.007142399787902832, 0.0070348801612854, 0.006993919849395752, 0.007071743965148926, 0.007003136157989502, 0.007015423774719238, 0.0069550080299377445, 0.006964223861694336, 0.007110655784606934, 0.007010303974151612, 0.00693555212020874, 0.006914048194885254, 0.0067983360290527345, 0.006817791938781738, 0.00678604793548584, 0.006778880119323731, 0.015246335983276366, 0.006754303932189941, 0.0067686400413513184, 0.007263232231140137, 0.007006207942962647, 0.007090176105499267, 0.007069695949554444, 0.007207935810089112, 0.007222271919250488, 0.0069918718338012695, 0.006966271877288818, 0.007046144008636474, 0.007095295906066895, 0.0070522880554199216, 0.007004159927368164, 0.007090176105499267, 0.007085055828094483, 0.007003136157989502, 0.006947840213775635, 0.006971392154693603, 0.006988800048828125, 0.006959104061126709, 0.006956031799316406, 0.007006271839141846, 0.00701638412475586, 0.0069550080299377445, 0.006962240219116211, 0.006986688137054444, 0.007027711868286133, 0.007079936027526855, 0.0069632000923156735, 0.007060480117797851, 0.006970367908477783, 0.007000063896179199, 0.006881279945373535, 0.0067758078575134275, 0.006843391895294189, 0.0067420158386230465, 0.006834176063537598, 0.006772736072540283, 0.006809599876403808, 0.0067573761940002445, 0.006814720153808594, 0.006766592025756836, 0.006762495994567871, 0.006816768169403077, 0.0069027838706970214, 0.007111680030822754, 0.0069816322326660156, 0.006994944095611572, 0.006974463939666748, 0.007569407939910889, 0.007007232189178467, 0.00710041618347168, 0.007008255958557129, 0.007001120090484619, 0.00694268798828125, 0.006989823818206787, 0.006985727787017822, 0.006980607986450196, 0.006964223861694336, 0.007035903930664063, 0.007004159927368164, 0.01575014400482178, 0.006979584217071533, 0.0069550080299377445, 0.0069928960800170895, 0.007011328220367432, 0.00694271993637085, 0.006915071964263916, 0.006803455829620361, 0.006804480075836182, 0.00673689603805542, 0.00679423999786377, 0.006772736072540283, 0.006783999919891357, 0.0069918718338012695, 0.0070563840866088865, 0.006979584217071533, 0.007008255958557129, 0.006976511955261231, 0.007006207942962647, 0.007045119762420654, 0.007014400005340577, 0.006939648151397705, 0.006975488185882568, 0.0069816322326660156, 0.0069283838272094726, 0.006980607986450196, 0.006985727787017822, 0.007003136157989502, 0.006960127830505371, 0.006994944095611572, 0.006979584217071533, 0.007016448020935059, 0.0069550080299377445, 0.007061503887176514, 0.007011328220367432, 0.0069928960800170895, 0.0069632000923156735, 0.006990911960601806, 0.007001023769378662, 0.007003136157989502, 0.006947840213775635, 0.006982656002044678, 0.007000063896179199, 0.006990848064422607, 0.006961152076721191, 0.00698367977142334, 0.0069959678649902345, 0.00698367977142334, 0.006989823818206787, 0.007016448020935059, 0.007010303974151612, 0.007003136157989502, 0.006945792198181152, 0.007011328220367432, 0.0069816322326660156, 0.006988800048828125, 0.006936575889587402, 0.006961152076721191, 0.006990848064422607, 0.0069621758460998535, 0.0070522880554199216, 0.00704307222366333, 0.007071743965148926]",tokens/s,141.34568339067224,,,main,False,False,,,,,,,,,,,,,,,,,,,,,,, bfloat16-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,databricks/dbrx-base,databricks/dbrx-base,cuda,0,42,,,True,,,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",dbrx,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-350m,,cuda,0,42,,,,,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,d35829e539df8480b726c647eeabf91e41eae047,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,,MB,1174.433792,1455.947776,0.0,809.500672,723.533824,s,13,0.18676848125457762,0.014366806250352126,0.0005655927627597173,0.01455446434020996,0.014721689414978027,0.01518812141418457,0.01572413932800293,"[0.01585814380645752, 0.014642687797546386, 0.014741439819335938, 0.014609312057495117, 0.013956352233886719, 0.013896032333374024, 0.013839424133300781, 0.013809280395507812, 0.01460409641265869, 0.014568863868713379, 0.013815808296203613, 0.01455446434020996, 0.013872575759887696]",tokens/s,17818.852397604063,kWh,1.672100464858931e-07,9.160921237509845e-08,4.6943853365206314e-07,7.282577925130547e-07,tokens/kWh,351523873.32046425,MB,1174.433792,1455.947776,0.0,809.500672,749.21984,s,13,10.515557861328123,0.8088890662560095,0.01598588227034598,0.8103314819335937,0.8233277099609375,0.831214794921875,0.840103408203125,"[0.8423255615234375, 0.8213235473632813, 0.8238076171875, 0.8103518676757813, 0.782132568359375, 0.803363037109375, 0.7837216186523438, 0.7996719360351563, 0.8214080810546875, 0.81534326171875, 0.8048592529296875, 0.8103314819335937, 0.7969180297851562]",tokens/s,77.88459830666174,kWh,9.496202768999352e-06,5.203489430023765e-06,1.7370255225268167e-05,3.206994742429129e-05,tokens/kWh,1964455.980126766,,s,818,10.672301052093502,0.0130468228020703,0.0018048516405457077,0.01297974395751953,0.013116723251342773,0.013315583896636961,0.026699346485137943,"[0.014223360061645507, 0.014007295608520508, 0.015222784042358398, 0.014003199577331543, 0.013818880081176758, 0.015129599571228027, 0.014413824081420898, 0.014791680335998534, 0.013922304153442382, 0.01409126377105713, 0.014049280166625976, 0.013950976371765136, 0.01466476821899414, 0.01347372817993164, 0.013149184226989746, 0.013238271713256837, 0.013000703811645508, 0.013118464469909668, 0.013089792251586914, 0.013172736167907715, 0.013143039703369141, 0.01305292797088623, 0.013071359634399414, 0.01306214427947998, 0.013190143585205078, 0.013019136428833008, 0.013095935821533204, 0.013046784400939941, 0.01307033634185791, 0.013423616409301758, 0.01305292797088623, 0.013020159721374512, 0.01305395221710205, 0.013109248161315918, 0.013044735908508302, 0.013036543846130372, 0.013006848335266113, 0.013083647727966309, 0.01297715187072754, 0.012917759895324708, 0.012932095527648926, 0.013006848335266113, 0.013328384399414063, 0.013268992424011231, 0.013270015716552735, 0.013496319770812988, 0.013110272407531738, 0.01318297576904297, 0.013117440223693847, 0.01297926425933838, 0.012971967697143554, 0.013061152458190919, 0.01304368019104004, 0.013007871627807617, 0.012980223655700684, 0.013025279998779296, 0.012989439964294434, 0.01298534393310547, 0.01296895980834961, 0.013030400276184082, 0.013002752304077148, 0.01297715187072754, 0.028115968704223632, 0.013014016151428222, 0.013172736167907715, 0.013099007606506348, 0.013074432373046875, 0.013047807693481446, 0.013002752304077148, 0.013033472061157226, 0.012983296394348144, 0.013022208213806152, 0.01303756809234619, 0.012976127624511719, 0.012957695960998536, 0.012974080085754394, 0.012986368179321289, 0.012958720207214355, 0.012991488456726074, 0.012995648384094238, 0.013151167869567871, 0.013116415977478027, 0.013014080047607422, 0.013004735946655273, 0.012995583534240723, 0.013029376029968261, 0.013083647727966309, 0.013011967658996582, 0.013022208213806152, 0.012980223655700684, 0.012967935562133789, 0.013273088455200196, 0.01306214427947998, 0.012969984054565429, 0.01296281623840332, 0.012981247901916505, 0.01296895980834961, 0.013031423568725586, 0.012980223655700684, 0.013023232460021973, 0.013040639877319337, 0.012999679565429688, 0.01298739242553711, 0.013038592338562012, 0.012984319686889649, 0.013030400276184082, 0.013000703811645508, 0.012994560241699218, 0.01303756809234619, 0.013058048248291015, 0.013004799842834473, 0.013073408126831054, 0.013034496307373047, 0.012991488456726074, 0.012975104331970215, 0.013081600189208984, 0.012945407867431641, 0.01296281623840332, 0.013157376289367676, 0.013058048248291015, 0.012981247901916505, 0.013083647727966309, 0.013314047813415527, 0.012993535995483398, 0.013092864036560058, 0.028242944717407226, 0.013030400276184082, 0.013048831939697265, 0.01305395221710205, 0.013033472061157226, 0.013058048248291015, 0.013022208213806152, 0.013064255714416504, 0.01308255958557129, 0.013124608039855956, 0.013118464469909668, 0.012998656272888183, 0.013011967658996582, 0.013036543846130372, 0.013022208213806152, 0.013128704071044921, 0.013054976463317871, 0.013126655578613282, 0.013066240310668945, 0.013033503532409669, 0.012939231872558594, 0.013085696220397949, 0.013234175682067872, 0.013249600410461426, 0.013135807991027832, 0.013026304244995117, 0.01307033634185791, 0.013116415977478027, 0.013030400276184082, 0.013033503532409669, 0.013102047920227051, 0.013033472061157226, 0.01303756809234619, 0.013109248161315918, 0.012986368179321289, 0.013240384101867677, 0.013071359634399414, 0.013055935859680175, 0.01305292797088623, 0.013175807952880859, 0.013057024002075195, 0.013056063652038574, 0.013079487800598144, 0.01305292797088623, 0.01307545566558838, 0.013128704071044921, 0.013109248161315918, 0.013074463844299316, 0.013017120361328125, 0.012999615669250487, 0.013179903984069824, 0.013029376029968261, 0.013040639877319337, 0.013011967658996582, 0.013112319946289062, 0.01306726360321045, 0.01305907154083252, 0.012999679565429688, 0.013097984313964844, 0.013029376029968261, 0.013017087936401368, 0.01306828784942627, 0.013087743759155274, 0.02812313652038574, 0.01306726360321045, 0.013007871627807617, 0.01305907154083252, 0.013093888282775879, 0.01296281623840332, 0.013049856185913086, 0.013153280258178711, 0.013021183967590331, 0.013083647727966309, 0.013009920120239257, 0.013063167572021485, 0.01313587188720703, 0.013107199668884278, 0.01307033634185791, 0.013054976463317871, 0.013090815544128418, 0.012999679565429688, 0.013054976463317871, 0.01305395221710205, 0.013029376029968261, 0.012959744453430176, 0.012958720207214355, 0.012980223655700684, 0.013061152458190919, 0.013029343605041504, 0.013011967658996582, 0.013049856185913086, 0.01354854393005371, 0.01326694393157959, 0.0132925443649292, 0.01343283176422119, 0.013219840049743652, 0.013042688369750977, 0.013019136428833008, 0.013055999755859376, 0.013085696220397949, 0.012998656272888183, 0.013035519599914551, 0.013041664123535156, 0.013005824089050292, 0.013023232460021973, 0.012781567573547363, 0.012198944091796875, 0.012355551719665528, 0.012399616241455079, 0.01232588768005371, 0.012409855842590332, 0.012333056449890138, 0.012379136085510254, 0.012327936172485352, 0.012355584144592285, 0.012368895530700684, 0.012318719863891601, 0.012306431770324706, 0.012410880088806153, 0.012956671714782715, 0.012585984230041505, 0.012337151527404786, 0.012339200019836426, 0.012453887939453125, 0.012298239707946777, 0.012349439620971679, 0.02674892807006836, 0.012324864387512208, 0.012374015808105468, 0.012386303901672363, 0.012370944023132324, 0.012207103729248046, 0.01232588768005371, 0.012340224266052247, 0.012311552047729492, 0.012361791610717774, 0.012392383575439453, 0.01236684799194336, 0.012313599586486817, 0.012305407524108887, 0.012309503555297852, 0.012368895530700684, 0.012305472373962402, 0.012352448463439941, 0.012357631683349609, 0.012285951614379884, 0.012339200019836426, 0.012306431770324706, 0.012374015808105468, 0.012315648078918457, 0.012355615615844726, 0.012303327560424805, 0.012316672325134278, 0.012331007957458496, 0.012309503555297852, 0.012275712013244629, 0.01233407974243164, 0.012346367835998535, 0.012338175773620605, 0.012332032203674317, 0.012326911926269531, 0.012314623832702636, 0.01238425636291504, 0.01233510398864746, 0.012309503555297852, 0.012302335739135742, 0.012167231559753418, 0.012144607543945312, 0.012191712379455566, 0.012248064041137695, 0.012620800018310547, 0.012620800018310547, 0.012785663604736328, 0.012713024139404297, 0.012696512222290039, 0.012643327713012695, 0.012707839965820313, 0.012731391906738282, 0.012685312271118163, 0.012676095962524414, 0.012711935997009278, 0.012380160331726075, 0.012308480262756348, 0.012718079566955566, 0.012755968093872071, 0.012724224090576173, 0.012666879653930664, 0.012277759552001954, 0.012290047645568849, 0.027765760421752928, 0.012851200103759765, 0.0127457275390625, 0.012516351699829101, 0.012830719947814942, 0.012534784317016602, 0.012319744110107422, 0.012383232116699219, 0.01275494384765625, 0.012527615547180175, 0.012541952133178711, 0.012452863693237304, 0.012560383796691894, 0.012712960243225097, 0.012547072410583495, 0.012734527587890626, 0.01278048038482666, 0.01325062370300293, 0.013735872268676758, 0.012928000450134277, 0.012864543914794921, 0.012786656379699708, 0.012743680000305176, 0.012713983535766601, 0.012776448249816894, 0.012483584403991698, 0.012587008476257324, 0.012990464210510254, 0.012790847778320312, 0.012723135948181152, 0.012750847816467285, 0.012669952392578124, 0.012699647903442383, 0.012701696395874023, 0.012693504333496093, 0.012744704246520995, 0.012510208129882813, 0.012728320121765137, 0.012812288284301757, 0.012718079566955566, 0.012695551872253418, 0.012702752113342285, 0.012743712425231934, 0.013584320068359374, 0.012843008041381837, 0.012808192253112792, 0.01274880027770996, 0.012492799758911132, 0.0123023681640625, 0.012484607696533203, 0.012739551544189454, 0.012728320121765137, 0.012811264038085938, 0.012743680000305176, 0.012733440399169921, 0.01267302417755127, 0.01266585636138916, 0.012747776031494141, 0.01294438362121582, 0.012803071975708008, 0.012718079566955566, 0.012802047729492188, 0.012767231941223145, 0.026677248001098632, 0.01233510398864746, 0.012399616241455079, 0.012340224266052247, 0.01234227180480957, 0.012347392082214356, 0.012338175773620605, 0.012318719863891601, 0.012597248077392579, 0.01276416015625, 0.012697600364685058, 0.012747776031494141, 0.012697600364685058, 0.012760064125061036, 0.012676095962524414, 0.012666879653930664, 0.012536895751953126, 0.012476351737976074, 0.012791808128356934, 0.012881919860839844, 0.012755968093872071, 0.012694527626037597, 0.01273036766052246, 0.012691455841064453, 0.012704768180847169, 0.012584959983825684, 0.012279808044433594, 0.012285951614379884, 0.012317695617675782, 0.012290047645568849, 0.012299263954162597, 0.012317695617675782, 0.012283904075622559, 0.012333056449890138, 0.012694527626037597, 0.012721152305603027, 0.012320768356323243, 0.012329983711242675, 0.012291071891784668, 0.012314623832702636, 0.012403712272644044, 0.012311552047729492, 0.012311552047729492, 0.012315648078918457, 0.012322815895080566, 0.012337151527404786, 0.012316672325134278, 0.012316672325134278, 0.012309503555297852, 0.012301312446594239, 0.012277759552001954, 0.012327936172485352, 0.012253184318542481, 0.012318719863891601, 0.012386303901672363, 0.012343359947204589, 0.012355520248413085, 0.012326911926269531, 0.012302335739135742, 0.012288000106811523, 0.012263423919677734, 0.01216102409362793, 0.01215385627746582, 0.02656153678894043, 0.012348447799682617, 0.012369888305664063, 0.01235148811340332, 0.012339200019836426, 0.012312576293945313, 0.012298239707946777, 0.012348416328430176, 0.012303359985351562, 0.012333056449890138, 0.012321791648864745, 0.012340224266052247, 0.012338208198547364, 0.012369888305664063, 0.012312576293945313, 0.012354559898376465, 0.012323840141296387, 0.012315648078918457, 0.012307456016540527, 0.012324864387512208, 0.012324864387512208, 0.012316672325134278, 0.012356608390808106, 0.012294143676757812, 0.012301312446594239, 0.012297216415405274, 0.012285951614379884, 0.012377087593078612, 0.012300288200378418, 0.01228492832183838, 0.012285951614379884, 0.012311552047729492, 0.01226956844329834, 0.01248256015777588, 0.01307545566558838, 0.01325158405303955, 0.013030400276184082, 0.01297715187072754, 0.013220864295959473, 0.015228927612304688, 0.013443072319030762, 0.013102080345153809, 0.01306009578704834, 0.012984319686889649, 0.013091839790344239, 0.012957695960998536, 0.012953599929809571, 0.013028351783752442, 0.01303660774230957, 0.012982208251953125, 0.013002752304077148, 0.013020159721374512, 0.012996607780456543, 0.012988415718078614, 0.012908543586730957, 0.013014016151428222, 0.012984319686889649, 0.012951552391052246, 0.012943360328674316, 0.012975104331970215, 0.012999744415283203, 0.013018048286437989, 0.012899328231811523, 0.028044288635253906, 0.012953599929809571, 0.013091839790344239, 0.01304371166229248, 0.012972031593322754, 0.013033472061157226, 0.013017087936401368, 0.013030400276184082, 0.012959744453430176, 0.01297107219696045, 0.013031359672546386, 0.012981247901916505, 0.012950528144836425, 0.013038592338562012, 0.013040639877319337, 0.013007871627807617, 0.013009920120239257, 0.013024255752563477, 0.013074432373046875, 0.012963839530944824, 0.012983296394348144, 0.01298739242553711, 0.013008895874023438, 0.0129617919921875, 0.013113344192504883, 0.013194239616394043, 0.013031423568725586, 0.013039615631103516, 0.01303756809234619, 0.013022208213806152, 0.013096960067749023, 0.013009920120239257, 0.013005824089050292, 0.012964863777160645, 0.013024255752563477, 0.013145088195800781, 0.013058048248291015, 0.012984319686889649, 0.013088768005371093, 0.012981247901916505, 0.013007871627807617, 0.012974080085754394, 0.012992511749267579, 0.012940287590026855, 0.013108223915100097, 0.013048831939697265, 0.013044735908508302, 0.013011967658996582, 0.012949503898620606, 0.013010944366455078, 0.013050880432128906, 0.013435903549194337, 0.013109248161315918, 0.013085696220397949, 0.012995583534240723, 0.013046784400939941, 0.013028351783752442, 0.013064191818237305, 0.013047807693481446, 0.012988415718078614, 0.013049856185913086, 0.013050880432128906, 0.013023232460021973, 0.027986944198608397, 0.013011967658996582, 0.01308572769165039, 0.013063136100769044, 0.013027327537536621, 0.01305907154083252, 0.01305907154083252, 0.013063167572021485, 0.012978240013122559, 0.013069248199462891, 0.013006848335266113, 0.012994560241699218, 0.01303756809234619, 0.013050880432128906, 0.0129617919921875, 0.012963839530944824, 0.012956671714782715, 0.01297920036315918, 0.013063167572021485, 0.013036543846130372, 0.013066240310668945, 0.01298534393310547, 0.013001728057861327, 0.012942336082458495, 0.013001728057861327, 0.012949503898620606, 0.013008895874023438, 0.012924927711486817, 0.012967935562133789, 0.012945407867431641, 0.013012991905212403, 0.013001728057861327, 0.012991488456726074, 0.01305292797088623, 0.013029376029968261, 0.013027327537536621, 0.01298739242553711, 0.013308927536010743, 0.012399616241455079, 0.012290047645568849, 0.01226854419708252, 0.012279808044433594, 0.01236684799194336, 0.012346367835998535, 0.012294143676757812, 0.013014016151428222, 0.013021183967590331, 0.012988415718078614, 0.012942367553710938, 0.012983263969421387, 0.012990464210510254, 0.013075519561767578, 0.012958656311035157, 0.013017087936401368, 0.01298739242553711, 0.012951552391052246, 0.012986368179321289, 0.013030400276184082, 0.013031423568725586, 0.013029376029968261, 0.013004799842834473, 0.012983296394348144, 0.013027327537536621, 0.026658815383911134, 0.012326911926269531, 0.012370944023132324, 0.012357631683349609, 0.012401663780212402, 0.012301312446594239, 0.012289024353027344, 0.012345343589782716, 0.01234124755859375, 0.012340224266052247, 0.01235865592956543, 0.012290047645568849, 0.012332032203674317, 0.012308480262756348, 0.012302335739135742, 0.012155903816223144, 0.01256447982788086, 0.012445695877075195, 0.012326911926269531, 0.012345343589782716, 0.012348416328430176, 0.012308480262756348, 0.012354559898376465, 0.01266380786895752, 0.013042688369750977, 0.012975104331970215, 0.013011967658996582, 0.013011967658996582, 0.013005824089050292, 0.01305907154083252, 0.01297920036315918, 0.012951552391052246, 0.013012991905212403, 0.012963839530944824, 0.013010944366455078, 0.012988415718078614, 0.01308672046661377, 0.012988415718078614, 0.013019136428833008, 0.013014016151428222, 0.013021183967590331, 0.013032447814941407, 0.01298739242553711, 0.013016063690185547, 0.013002752304077148, 0.012993535995483398, 0.01305907154083252, 0.013041664123535156, 0.013065216064453124, 0.01303756809234619, 0.01298739242553711, 0.01297100830078125, 0.013324288368225098, 0.013092864036560058, 0.013030400276184082, 0.013009920120239257, 0.013025279998779296, 0.013011967658996582, 0.01298534393310547, 0.01306726360321045, 0.01307545566558838, 0.01297920036315918, 0.01294438362121582, 0.028052480697631835, 0.012974080085754394, 0.013095935821533204, 0.013039615631103516, 0.012986368179321289, 0.01298534393310547, 0.012939264297485351, 0.013031423568725586, 0.013142016410827637, 0.013030400276184082, 0.013020159721374512, 0.012976160049438476, 0.012978143692016601, 0.013171711921691894, 0.013093888282775879, 0.013016063690185547, 0.013008895874023438, 0.012967935562133789, 0.013001728057861327, 0.013021183967590331, 0.012945407867431641, 0.013028351783752442, 0.012909567832946778, 0.013049856185913086, 0.013009984016418457, 0.013238207817077637, 0.01295257568359375, 0.0129617919921875, 0.012969984054565429, 0.012990464210510254, 0.012967935562133789, 0.012988479614257812, 0.012958656311035157, 0.013046784400939941, 0.013035519599914551, 0.012996607780456543, 0.013031423568725586, 0.013010944366455078, 0.01296281623840332, 0.012992511749267579, 0.012935168266296386, 0.013081664085388184, 0.013012928009033203, 0.012912639617919922, 0.012860416412353515, 0.012777471542358398, 0.01286348819732666, 0.012805120468139648, 0.012726271629333496, 0.012781567573547363, 0.01276518440246582, 0.01227673625946045, 0.012314623832702636, 0.012329983711242675, 0.012315648078918457, 0.012312576293945313, 0.012346367835998535, 0.012312576293945313, 0.012282879829406738, 0.012322815895080566, 0.012355648040771484, 0.01234937572479248, 0.012306431770324706, 0.026703872680664063, 0.012537856101989747, 0.012406784057617188, 0.012346367835998535, 0.013423616409301758, 0.013212672233581543, 0.013048831939697265, 0.013064191818237305, 0.01306828784942627, 0.012991488456726074, 0.01305190372467041, 0.012982272148132324, 0.012990464210510254, 0.013031423568725586, 0.013014016151428222, 0.013667327880859375, 0.013643775939941406, 0.01397555160522461, 0.012969984054565429, 0.01287782382965088, 0.01286143970489502, 0.012811264038085938, 0.012800000190734863, 0.012703743934631348, 0.012715007781982422, 0.012696576118469239, 0.01266585636138916, 0.01237606430053711, 0.012340224266052247, 0.012355584144592285, 0.012337151527404786, 0.012338239669799805, 0.012320704460144043, 0.012310527801513671, 0.012346400260925292, 0.01231663990020752, 0.01237299156188965, 0.012289088249206543, 0.01235654354095459, 0.012388352394104005, 0.012314623832702636, 0.012311552047729492, 0.01227673625946045, 0.012338175773620605, 0.012363776206970215, 0.0127457275390625, 0.012875776290893554, 0.01274880027770996, 0.012710911750793457, 0.012705792427062988, 0.012588031768798828, 0.012382207870483398, 0.012307456016540527, 0.012422207832336427, 0.01236780834197998, 0.012340224266052247, 0.012315648078918457, 0.012404735565185548, 0.012314623832702636, 0.012368895530700684, 0.012321791648864745, 0.012285951614379884, 0.012305407524108887]",tokens/s,76.64701323615108,,,main,False,False,,,,,,,,,,,,,,,,,,,,,,, bfloat16-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,-,-,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 106, in _inner_fn validate_repo_id(arg_value) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 160, in validate_repo_id raise HFValidationError( huggingface_hub.errors.HFValidationError: Repo id must use alphanumeric chars or '-', '_', '.', '--' and '..' are forbidden, '-' and '.' cannot start or end the name, max length is 96: '-'. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 466, in cached_file raise EnvironmentError( OSError: Incorrect path_or_model_id: '-'. Please provide either the path to a local folder or the repo_id of a model on the Hub. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,stabilityai/stablelm-2-12b,stabilityai/stablelm-2-12b,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ self.load_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 250, in load_model_with_no_weights self.pretrained_model = self.automodel_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3710, in from_pretrained model = cls(config, *model_args, **model_kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/stablelm/modeling_stablelm.py"", line 1188, in __init__ self.model = StableLmModel(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/stablelm/modeling_stablelm.py"", line 966, in __init__ [StableLmDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/stablelm/modeling_stablelm.py"", line 966, in [StableLmDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/stablelm/modeling_stablelm.py"", line 748, in __init__ self.mlp = StableLmMLP(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/stablelm/modeling_stablelm.py"", line 206, in __init__ self.down_proj = nn.Linear(self.intermediate_size, self.hidden_size, bias=False) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/linear.py"", line 98, in __init__ self.weight = Parameter(torch.empty((out_features, in_features), **factory_kwargs)) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_device.py"", line 78, in __torch_function__ return func(*args, **kwargs) torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 136.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,v,v,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/v/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66932562-37203d382238f61e4707f03a;000bab14-a8e5-4590-9614-42fbb8e3c1fd) Repository Not Found for url: https://huggingface.co/v/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: v is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,openai-community/gpt2,openai-community/gpt2,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1234.038784,1032.323072,0.0,385.875968,318.907904,s,19,0.1761198081970215,0.00926946358931692,0.0002835937887193509,0.00926963233947754,0.009363161277770995,0.009465558624267576,0.010105310363769531,"[0.01026524829864502, 0.009332096099853516, 0.009309375762939453, 0.009376704216003418, 0.009312447547912597, 0.009260831832885742, 0.009246496200561524, 0.00935977554321289, 0.008928544044494629, 0.009299424171447754, 0.009354207992553711, 0.009302592277526856, 0.00926483154296875, 0.00926963233947754, 0.009247872352600097, 0.008929023742675781, 0.008908320426940917, 0.008900192260742188, 0.009252191543579102]",tokens/s,27617.563576714474,kWh,1.0802043333616341e-07,5.918964731942071e-08,2.6133247597623025e-07,4.2854255663181443e-07,tokens/kWh,597373577.1123061,MB,1234.038784,1032.323072,0.0,385.875968,328.80384,s,19,10.278541442871092,0.5409758654142681,0.00678675710984656,0.54367724609375,0.545797119140625,0.5463724487304688,0.5490663842773438,"[0.5497398681640625, 0.545746826171875, 0.5439652709960937, 0.5398739013671875, 0.543837158203125, 0.5437479858398437, 0.54564892578125, 0.5441216430664062, 0.5408333129882813, 0.545998291015625, 0.5440383911132812, 0.5428787231445312, 0.5430189208984375, 0.5387953491210937, 0.5398588256835938, 0.526919677734375, 0.5200580444335937, 0.5357830810546875, 0.54367724609375]",tokens/s,116.45621187140374,kWh,6.337918272713282e-06,3.4728926074871085e-06,1.1043004052865835e-05,2.0853814933066228e-05,tokens/kWh,3021029.9747172846,,s,1196,10.43495627975463,0.008724879832570769,0.0011961206501586723,0.00860262393951416,0.00870143985748291,0.008812983751296997,0.018294545173645014,"[0.00960921573638916, 0.009563136100769042, 0.009516032218933105, 0.00981503963470459, 0.009378815650939941, 0.008648703575134278, 0.008649727821350098, 0.008677375793457032, 0.008624128341674805, 0.008626175880432128, 0.008616000175476075, 0.008636351585388183, 0.008646656036376953, 0.008670207977294921, 0.008650752067565918, 0.008689663887023925, 0.008629247665405274, 0.008630271911621093, 0.008577024459838867, 0.00861695957183838, 0.008600576400756836, 0.008594431877136231, 0.008606719970703124, 0.00869478416442871, 0.008638463973999023, 0.008730624198913574, 0.008641535758972169, 0.008581119537353516, 0.008572928428649903, 0.008614912033081054, 0.008606719970703124, 0.008599552154541015, 0.00858521556854248, 0.008596480369567871, 0.008640512466430664, 0.008639488220214844, 0.008577024459838867, 0.00859340763092041, 0.008613887786865235, 0.008633343696594239, 0.008679424285888672, 0.00859545612335205, 0.00858732795715332, 0.008583104133605957, 0.008677375793457032, 0.008597503662109375, 0.008678463935852051, 0.008670144081115723, 0.00860268783569336, 0.008646592140197754, 0.008651776313781738, 0.008606719970703124, 0.008600576400756836, 0.008600576400756836, 0.008828927993774414, 0.008631296157836914, 0.008586239814758301, 0.008588288307189941, 0.008606719970703124, 0.008619008064270019, 0.008634367942810058, 0.00860979175567627, 0.01861734390258789, 0.008725503921508788, 0.008714240074157715, 0.008846336364746094, 0.008684543609619141, 0.008775679588317872, 0.008758272171020508, 0.008814592361450196, 0.008697855949401855, 0.008681471824645997, 0.008720383644104004, 0.008751104354858399, 0.008645631790161134, 0.008679424285888672, 0.008753151893615722, 0.008663040161132812, 0.008672256469726563, 0.008627200126647949, 0.008599552154541015, 0.008652799606323243, 0.008640512466430664, 0.008686592102050781, 0.008624128341674805, 0.008614912033081054, 0.00861087989807129, 0.008591296195983887, 0.00860364818572998, 0.008615936279296875, 0.008679424285888672, 0.008624128341674805, 0.008686592102050781, 0.00860159969329834, 0.008612863540649414, 0.00861081600189209, 0.008606719970703124, 0.008583168029785156, 0.008549375534057617, 0.008621055603027344, 0.008566783905029298, 0.008592384338378906, 0.008669183731079102, 0.008658944129943847, 0.008626175880432128, 0.00860979175567627, 0.008626175880432128, 0.008614912033081054, 0.008631296157836914, 0.008623200416564942, 0.008594335556030274, 0.008588288307189941, 0.008633343696594239, 0.008597503662109375, 0.008641535758972169, 0.008774656295776367, 0.00862003231048584, 0.008674304008483886, 0.00860364818572998, 0.008615936279296875, 0.008730624198913574, 0.00869375991821289, 0.008614944458007813, 0.008651743888854981, 0.008619008064270019, 0.018497535705566406, 0.00869478416442871, 0.00862003231048584, 0.008921088218688965, 0.008733695983886718, 0.009038847923278808, 0.008803327560424805, 0.009194496154785157, 0.009391103744506836, 0.008847359657287598, 0.008753151893615722, 0.008700927734375, 0.008655872344970703, 0.008652799606323243, 0.008627200126647949, 0.008913920402526856, 0.008684543609619141, 0.008644607543945313, 0.00861299228668213, 0.008557439804077149, 0.008605695724487305, 0.008619008064270019, 0.008514559745788575, 0.008493056297302246, 0.008482912063598632, 0.008566847801208496, 0.00850108814239502, 0.008480768203735351, 0.008501248359680176, 0.008581119537353516, 0.00851865577697754, 0.00849612808227539, 0.008474623680114746, 0.00852070426940918, 0.008570879936218261, 0.008571904182434082, 0.008498175621032715, 0.00859545612335205, 0.008623104095458984, 0.008492128372192383, 0.008515551567077636, 0.008512448310852051, 0.008858624458312989, 0.008569855690002442, 0.008532992362976074, 0.008581119537353516, 0.008529919624328614, 0.008468480110168456, 0.008531968116760253, 0.008484864234924316, 0.008621055603027344, 0.008539135932922364, 0.008488960266113281, 0.008506367683410645, 0.008863743782043456, 0.008530976295471192, 0.008571871757507324, 0.008551424026489257, 0.008511487960815429, 0.00850432014465332, 0.0085032958984375, 0.0086364164352417, 0.008644607543945313, 0.01839321517944336, 0.008479616165161134, 0.008769536018371582, 0.008564895629882812, 0.008478591918945313, 0.00854422378540039, 0.008560640335083008, 0.008645631790161134, 0.008503392219543457, 0.008634271621704102, 0.008517631530761719, 0.008549504280090332, 0.008546175956726075, 0.008557567596435547, 0.00851353645324707, 0.008596480369567871, 0.008539135932922364, 0.008495264053344727, 0.008476608276367188, 0.008514464378356934, 0.008493056297302246, 0.008507391929626466, 0.008493056297302246, 0.008491007804870606, 0.008546303749084473, 0.008525919914245606, 0.008828831672668456, 0.008557567596435547, 0.008629247665405274, 0.00851465606689453, 0.00848697566986084, 0.008502240180969239, 0.008502143859863281, 0.008560640335083008, 0.008599552154541015, 0.008566783905029298, 0.008479743957519532, 0.008496159553527833, 0.008467488288879395, 0.008479711532592773, 0.0084716157913208, 0.008487839698791504, 0.008525823593139649, 0.008508416175842285, 0.0085698881149292, 0.008614879608154297, 0.008631296157836914, 0.00862003231048584, 0.008639488220214844, 0.00862822437286377, 0.008554592132568359, 0.008516511917114257, 0.008501248359680176, 0.008656895637512207, 0.008624128341674805, 0.008635392189025879, 0.008613887786865235, 0.008623104095458984, 0.00868556785583496, 0.00861695957183838, 0.008600576400756836, 0.008658944129943847, 0.008662015914916991, 0.01845043182373047, 0.008548352241516113, 0.008648703575134278, 0.008770560264587402, 0.008599552154541015, 0.008632320404052735, 0.00861695957183838, 0.008502304077148437, 0.008612832069396973, 0.00858726406097412, 0.008626175880432128, 0.008614912033081054, 0.008612863540649414, 0.008600576400756836, 0.008586239814758301, 0.008626175880432128, 0.008632320404052735, 0.008659968376159668, 0.00859545612335205, 0.008600576400756836, 0.008586239814758301, 0.008592384338378906, 0.00870809555053711, 0.008632320404052735, 0.008629247665405274, 0.008629247665405274, 0.008582176208496094, 0.0085513916015625, 0.008529919624328614, 0.00861184024810791, 0.008631296157836914, 0.008788991928100585, 0.008653823852539062, 0.008623104095458984, 0.00860159969329834, 0.00870297622680664, 0.008625151634216309, 0.008619008064270019, 0.00862822437286377, 0.008621055603027344, 0.008592384338378906, 0.008624128341674805, 0.008684543609619141, 0.008638463973999023, 0.00861695957183838, 0.00861695957183838, 0.008625151634216309, 0.008630271911621093, 0.008646656036376953, 0.008599552154541015, 0.008571904182434082, 0.008586239814758301, 0.008663040161132812, 0.008896512031555176, 0.008625151634216309, 0.008598527908325194, 0.008591360092163085, 0.008599552154541015, 0.008621055603027344, 0.008542207717895508, 0.00869478416442871, 0.008641535758972169, 0.008627200126647949, 0.018366464614868162, 0.008639488220214844, 0.008650752067565918, 0.008634367942810058, 0.008612863540649414, 0.0086364164352417, 0.00861695957183838, 0.008606719970703124, 0.008622079849243165, 0.008630271911621093, 0.008604767799377442, 0.008617888450622559, 0.00860979175567627, 0.008645631790161134, 0.008675328254699707, 0.008605695724487305, 0.008597503662109375, 0.008583168029785156, 0.008623104095458984, 0.008625151634216309, 0.008573951721191407, 0.00860262393951416, 0.008597503662109375, 0.00858521556854248, 0.008572928428649903, 0.008604672431945801, 0.008613887786865235, 0.008641535758972169, 0.008627200126647949, 0.008572928428649903, 0.008559616088867187, 0.008574975967407226, 0.008722432136535644, 0.008671232223510742, 0.00860979175567627, 0.00859340763092041, 0.008564736366271973, 0.008547327995300292, 0.008666111946105956, 0.008614912033081054, 0.008644607543945313, 0.008652799606323243, 0.008573951721191407, 0.008878111839294434, 0.00868553638458252, 0.00860262393951416, 0.00861184024810791, 0.008626175880432128, 0.008600576400756836, 0.008599552154541015, 0.008639488220214844, 0.00868556785583496, 0.00861081600189209, 0.008627200126647949, 0.008697855949401855, 0.008635392189025879, 0.008641535758972169, 0.008641535758972169, 0.008586239814758301, 0.008550399780273438, 0.008642560005187988, 0.008572928428649903, 0.008695808410644532, 0.018362367630004883, 0.008586239814758301, 0.008633343696594239, 0.008608799934387207, 0.008598496437072754, 0.008687616348266602, 0.008660991668701172, 0.008594431877136231, 0.008591360092163085, 0.008758272171020508, 0.008675328254699707, 0.008668160438537598, 0.008675328254699707, 0.008608768463134766, 0.008598527908325194, 0.008642560005187988, 0.008649727821350098, 0.00860159969329834, 0.008626175880432128, 0.00889350414276123, 0.008676287651062011, 0.0086692476272583, 0.008631232261657715, 0.008707072257995606, 0.00861695957183838, 0.008606719970703124, 0.008588288307189941, 0.00859545612335205, 0.008766464233398438, 0.00860364818572998, 0.00860364818572998, 0.008645631790161134, 0.00859340763092041, 0.008573951721191407, 0.008622079849243165, 0.008597503662109375, 0.00860262393951416, 0.00909721565246582, 0.00884438419342041, 0.008812447547912598, 0.008582143783569337, 0.008646656036376953, 0.00862217617034912, 0.008614879608154297, 0.008728511810302734, 0.00860159969329834, 0.00862822437286377, 0.00874294376373291, 0.008622048377990722, 0.008660991668701172, 0.008674304008483886, 0.00861184024810791, 0.008721407890319824, 0.00861081600189209, 0.008573984146118164, 0.008600543975830079, 0.0086179838180542, 0.00860159969329834, 0.008668160438537598, 0.008713215827941894, 0.00870297622680664, 0.008607744216918945, 0.00860979175567627, 0.018431999206542968, 0.008613887786865235, 0.008761343955993652, 0.008674304008483886, 0.008592384338378906, 0.00863548755645752, 0.008608672142028808, 0.008647680282592773, 0.00863542366027832, 0.008628191947937012, 0.008604672431945801, 0.008557567596435547, 0.008641535758972169, 0.008631296157836914, 0.008656895637512207, 0.008639488220214844, 0.008836128234863282, 0.00866710376739502, 0.00859340763092041, 0.00861695957183838, 0.008607744216918945, 0.008666111946105956, 0.008599552154541015, 0.008596480369567871, 0.008545280456542969, 0.008604672431945801, 0.008604672431945801, 0.008577024459838867, 0.00858521556854248, 0.008619008064270019, 0.008604672431945801, 0.008590335845947266, 0.008598527908325194, 0.008679424285888672, 0.008605695724487305, 0.008607744216918945, 0.009479167938232422, 0.008707072257995606, 0.008621055603027344, 0.008563712120056152, 0.008769536018371582, 0.008626175880432128, 0.008640576362609863, 0.008529855728149413, 0.008521727561950684, 0.008519680023193359, 0.008491007804870606, 0.008591360092163085, 0.008580096244812012, 0.008676511764526367, 0.008531807899475097, 0.008543231964111327, 0.00851353645324707, 0.008501248359680176, 0.008613887786865235, 0.008619008064270019, 0.008644607543945313, 0.00861081600189209, 0.008650752067565918, 0.008753151893615722, 0.008604672431945801, 0.0086179838180542, 0.008615936279296875, 0.017687551498413084, 0.008210432052612305, 0.008233983993530274, 0.008246272087097169, 0.008204287528991699, 0.00840499210357666, 0.008238112449645996, 0.008816608428955078, 0.008642560005187988, 0.008674304008483886, 0.00861081600189209, 0.008608768463134766, 0.008586239814758301, 0.008664064407348633, 0.008605695724487305, 0.008803327560424805, 0.008681471824645997, 0.008854528427124024, 0.008634367942810058, 0.00859340763092041, 0.008598527908325194, 0.008491007804870606, 0.008597503662109375, 0.00860159969329834, 0.008590335845947266, 0.00857907199859619, 0.00860159969329834, 0.009358336448669433, 0.008696831703186036, 0.008651776313781738, 0.008608768463134766, 0.008589344024658203, 0.008601568222045898, 0.00859545612335205, 0.008660991668701172, 0.008583168029785156, 0.00858521556854248, 0.008589311599731446, 0.008578047752380372, 0.008597503662109375, 0.00860262393951416, 0.008591360092163085, 0.008634367942810058, 0.008553471565246582, 0.008621055603027344, 0.008576000213623047, 0.00862934398651123, 0.008628128051757812, 0.008577024459838867, 0.00858726406097412, 0.008586336135864257, 0.008451999664306641, 0.008608768463134766, 0.008604767799377442, 0.008504223823547364, 0.008637503623962402, 0.008584128379821778, 0.008641535758972169, 0.008486911773681641, 0.008455167770385743, 0.00849612808227539, 0.008499199867248536, 0.008499296188354492, 0.019571615219116212, 0.008855551719665527, 0.008785920143127441, 0.008935423851013183, 0.00912281608581543, 0.008796159744262694, 0.008659968376159668, 0.00862822437286377, 0.008816639900207519, 0.008637439727783204, 0.00862822437286377, 0.008658944129943847, 0.008589311599731446, 0.008622079849243165, 0.008624223709106446, 0.008592288017272949, 0.008615936279296875, 0.008632320404052735, 0.00872447967529297, 0.008774656295776367, 0.008627200126647949, 0.008612863540649414, 0.0087326717376709, 0.00860979175567627, 0.008615936279296875, 0.00858726406097412, 0.008583168029785156, 0.008774656295776367, 0.008605695724487305, 0.008639488220214844, 0.009695232391357422, 0.008706048011779785, 0.00931942367553711, 0.008755200386047364, 0.008649727821350098, 0.008646656036376953, 0.008619008064270019, 0.008596480369567871, 0.00862003231048584, 0.008265727996826172, 0.008257535934448243, 0.008247296333312988, 0.008230912208557128, 0.008282112121582032, 0.008240127563476562, 0.008223744392395019, 0.00841932773590088, 0.008633343696594239, 0.00862003231048584, 0.00858726406097412, 0.008612863540649414, 0.008591360092163085, 0.008596480369567871, 0.008568832397460938, 0.008580096244812012, 0.008600576400756836, 0.008577024459838867, 0.008604672431945801, 0.008647680282592773, 0.008705023765563966, 0.008664064407348633, 0.008664064407348633, 0.008615936279296875, 0.018513919830322266, 0.008660991668701172, 0.008626175880432128, 0.00861695957183838, 0.008649727821350098, 0.008605695724487305, 0.008591360092163085, 0.008646656036376953, 0.008643584251403809, 0.008599552154541015, 0.008651776313781738, 0.008608799934387207, 0.008613856315612792, 0.008641599655151367, 0.008740799903869629, 0.008658944129943847, 0.008648703575134278, 0.008644607543945313, 0.00862822437286377, 0.008647680282592773, 0.008797183990478515, 0.008650752067565918, 0.008652799606323243, 0.00861184024810791, 0.008630271911621093, 0.008649727821350098, 0.008633343696594239, 0.008642560005187988, 0.008646656036376953, 0.00858521556854248, 0.00862003231048584, 0.008607775688171386, 0.008592351913452149, 0.008606719970703124, 0.0086179838180542, 0.008606719970703124, 0.008578047752380372, 0.008584192276000976, 0.008631296157836914, 0.00861695957183838, 0.00871116828918457, 0.008665087699890137, 0.008643584251403809, 0.008679424285888672, 0.008599552154541015, 0.008638463973999023, 0.008621055603027344, 0.008584192276000976, 0.008622079849243165, 0.00860364818572998, 0.00859340763092041, 0.008581119537353516, 0.008580096244812012, 0.008641535758972169, 0.008614912033081054, 0.008567808151245117, 0.008586239814758301, 0.008574975967407226, 0.008584192276000976, 0.008698880195617676, 0.008624128341674805, 0.008652799606323243, 0.008581119537353516, 0.018361343383789062, 0.008582143783569337, 0.008637439727783204, 0.00861184024810791, 0.00860159969329834, 0.008555520057678222, 0.008588288307189941, 0.008622079849243165, 0.008656895637512207, 0.008580096244812012, 0.008570879936218261, 0.008600576400756836, 0.008600576400756836, 0.00860159969329834, 0.008728575706481934, 0.008646656036376953, 0.008612863540649414, 0.008612863540649414, 0.0086179838180542, 0.00868556785583496, 0.008594431877136231, 0.00860159969329834, 0.008591360092163085, 0.00861081600189209, 0.008597503662109375, 0.008566783905029298, 0.008615936279296875, 0.00861695957183838, 0.00859545612335205, 0.008550399780273438, 0.008612863540649414, 0.008583231925964356, 0.008604607582092284, 0.00868556785583496, 0.008625151634216309, 0.008560640335083008, 0.008590335845947266, 0.00860159969329834, 0.008622079849243165, 0.00857907199859619, 0.00857907199859619, 0.008604672431945801, 0.008597536087036133, 0.008594400405883789, 0.008714240074157715, 0.008639488220214844, 0.008765439987182617, 0.008697855949401855, 0.008582143783569337, 0.00858521556854248, 0.008597536087036133, 0.00857084846496582, 0.008668160438537598, 0.008552448272705078, 0.008581119537353516, 0.008566783905029298, 0.00859340763092041, 0.008583168029785156, 0.008581119537353516, 0.00858726406097412, 0.00862003231048584, 0.008668160438537598, 0.008599583625793456, 0.018345951080322265, 0.008599616050720215, 0.008660927772521972, 0.008613887786865235, 0.008606719970703124, 0.008594431877136231, 0.008652799606323243, 0.008779775619506837, 0.008623104095458984, 0.00858726406097412, 0.008656895637512207, 0.008615936279296875, 0.00860159969329834, 0.00859545612335205, 0.008583168029785156, 0.008589311599731446, 0.008614912033081054, 0.008584192276000976, 0.008623104095458984, 0.008600576400756836, 0.00861184024810791, 0.008651776313781738, 0.008626175880432128, 0.008591360092163085, 0.008539135932922364, 0.008612863540649414, 0.008700927734375, 0.008632320404052735, 0.008563712120056152, 0.008544256210327148, 0.008594431877136231, 0.008580096244812012, 0.008583168029785156, 0.008558591842651368, 0.008604672431945801, 0.008598527908325194, 0.00861695957183838, 0.008605695724487305, 0.008578047752380372, 0.008584192276000976, 0.008581119537353516, 0.008669183731079102, 0.008577024459838867, 0.008586239814758301, 0.0086179838180542, 0.00860262393951416, 0.008672256469726563, 0.008641535758972169, 0.00862003231048584, 0.008586272239685059, 0.008597472190856933, 0.008574975967407226, 0.00859545612335205, 0.00858726406097412, 0.008573951721191407, 0.008580096244812012, 0.008559616088867187, 0.00858521556854248, 0.008816639900207519, 0.008751104354858399, 0.008647680282592773, 0.008561663627624512, 0.008616991996765137, 0.018291839599609373, 0.008631135940551758, 0.008653823852539062, 0.008612863540649414, 0.008466431617736817, 0.008551424026489257, 0.008605695724487305, 0.008470527648925781, 0.008495103836059571, 0.008524800300598144, 0.008592384338378906, 0.008525823593139649, 0.008489983558654785, 0.008482815742492676, 0.008600576400756836, 0.008571904182434082, 0.00854537582397461, 0.00847043228149414, 0.008525823593139649, 0.008940544128417969, 0.008872960090637207, 0.00851865577697754, 0.008604672431945801, 0.008555520057678222, 0.008457216262817382, 0.00847878360748291, 0.008506303787231445, 0.008551424026489257, 0.00849715232849121, 0.008477696418762207, 0.008525823593139649, 0.008648703575134278, 0.008497183799743652, 0.008490976333618163, 0.008516608238220215, 0.008785951614379882, 0.008578016281127929, 0.008500224113464355, 0.008482815742492676, 0.00870195198059082, 0.008458239555358887, 0.00860159969329834, 0.008501248359680176, 0.00850227165222168, 0.008471551895141602, 0.008524800300598144, 0.008447999954223634, 0.008543231964111327, 0.008495103836059571, 0.008574975967407226, 0.008470527648925781, 0.008488960266113281, 0.008477824211120605, 0.008491904258728027, 0.0086364164352417, 0.008530943870544434, 0.008537088394165039, 0.008506367683410645, 0.008481792449951172, 0.008525856018066406, 0.008484831809997559, 0.00856991958618164, 0.008487872123718262, 0.018192384719848635, 0.008523776054382324, 0.008588288307189941, 0.00850534439086914, 0.008482815742492676, 0.008519680023193359, 0.008455167770385743, 0.00848793601989746, 0.008476672172546386, 0.008488991737365722, 0.008496255874633788, 0.008473440170288086, 0.008488960266113281, 0.008495103836059571, 0.008566816329956055, 0.008496095657348633, 0.008525823593139649, 0.008466591835021972, 0.008514399528503419, 0.008522751808166504, 0.008523776054382324, 0.008465408325195312, 0.008499199867248536, 0.008571904182434082, 0.008631296157836914, 0.008663040161132812, 0.008570879936218261, 0.008516736030578613, 0.008466303825378418, 0.008524800300598144, 0.008578047752380372, 0.0086179838180542, 0.008629247665405274, 0.008622079849243165, 0.0086179838180542, 0.00862822437286377, 0.008572928428649903, 0.008625151634216309, 0.008604672431945801, 0.008562687873840333, 0.008598527908325194, 0.008584192276000976, 0.008572928428649903, 0.008589311599731446, 0.008588288307189941, 0.008605695724487305, 0.008606719970703124, 0.008583168029785156, 0.008663040161132812, 0.008591360092163085, 0.00857907199859619, 0.008600576400756836, 0.00860159969329834, 0.008598527908325194, 0.008594431877136231, 0.008625184059143066, 0.008588255882263184, 0.008607744216918945, 0.008632320404052735, 0.008564736366271973, 0.008599552154541015, 0.008647680282592773, 0.008648703575134278, 0.01756879997253418, 0.00823087978363037, 0.008239104270935058, 0.008239104270935058, 0.008252448081970214, 0.008231904029846191, 0.008230912208557128, 0.008208383560180664, 0.008254464149475099, 0.008227840423583984, 0.008236031532287597, 0.008238080024719239, 0.008251392364501953, 0.008266752243041992, 0.008210463523864746, 0.008242143630981445, 0.008265727996826172, 0.008195103645324707, 0.0082390718460083, 0.008223744392395019, 0.008291328430175781, 0.008223744392395019, 0.008257535934448243, 0.008236031532287597, 0.008293375968933106, 0.008239104270935058, 0.008243200302124023, 0.008246272087097169, 0.008295424461364746, 0.008249343872070313, 0.008249343872070313, 0.008241151809692383, 0.008269824028015137, 0.008238080024719239, 0.008250368118286134, 0.008256511688232422, 0.008254464149475099, 0.008258560180664062, 0.008282112121582032, 0.008232959747314453, 0.008286208152770995, 0.008138751983642578, 0.008111104011535645, 0.008159232139587403, 0.008158207893371582, 0.008156160354614257, 0.008212479591369629, 0.009135104179382325, 0.009143296241760255, 0.008757247924804687, 0.008656895637512207, 0.008614912033081054, 0.008695808410644532, 0.008671296119689942, 0.008644543647766113, 0.008625151634216309, 0.00861081600189209, 0.00860159969329834, 0.008639488220214844, 0.008622079849243165, 0.008765439987182617, 0.008695808410644532, 0.00861081600189209, 0.01762816047668457, 0.008187904357910156, 0.008258560180664062, 0.008238080024719239, 0.008229887962341309, 0.008261631965637208, 0.008258624076843262, 0.008235967636108399, 0.008255488395690918, 0.008233983993530274, 0.008244223594665527, 0.008210432052612305, 0.008274944305419921, 0.008220671653747558, 0.008229887962341309, 0.008245247840881348, 0.008267775535583496, 0.008196096420288086, 0.008261631965637208, 0.008228896141052246, 0.008219679832458497, 0.008235967636108399, 0.008202239990234375, 0.008203264236450195, 0.008246272087097169, 0.008217599868774414, 0.008246272087097169, 0.008224767684936523, 0.008227840423583984, 0.008338432312011718, 0.008295424461364746, 0.008226816177368163, 0.008265727996826172, 0.008285183906555176, 0.008243200302124023, 0.008245247840881348, 0.008333312034606934, 0.008259584426879883, 0.008263680458068847, 0.008284159660339355, 0.008286208152770995, 0.008266752243041992, 0.008262656211853027, 0.008245247840881348, 0.008300543785095215, 0.008239104270935058, 0.008231936454772949, 0.008235008239746093, 0.008259584426879883, 0.008242176055908204, 0.008249343872070313, 0.008216575622558593, 0.008233983993530274, 0.008228863716125488, 0.008241151809692383, 0.00822169589996338, 0.008272895812988281, 0.008324095726013184, 0.008265727996826172, 0.008244223594665527, 0.008226816177368163, 0.008257535934448243, 0.008216575622558593, 0.0175861759185791, 0.008181759834289551, 0.008225791931152344, 0.008239104270935058, 0.008209407806396484, 0.008219648361206054, 0.008249343872070313, 0.008232959747314453, 0.008255488395690918, 0.008243200302124023, 0.008238080024719239, 0.008238080024719239, 0.008249343872070313, 0.00831283187866211, 0.008235008239746093, 0.00829030418395996, 0.008261631965637208, 0.008256511688232422, 0.008227840423583984, 0.008226816177368163, 0.008252415657043457, 0.00821555233001709, 0.008227840423583984, 0.008219648361206054, 0.008942591667175292, 0.008948736190795899, 0.009385984420776367, 0.008733695983886718, 0.008668160438537598, 0.008678400039672851, 0.008720383644104004, 0.008633343696594239, 0.008651776313781738, 0.008631296157836914, 0.008662015914916991, 0.00862003231048584, 0.008633343696594239, 0.00860979175567627, 0.008638463973999023, 0.008616991996765137, 0.008580063819885254, 0.008599552154541015, 0.008598527908325194, 0.008597503662109375, 0.008626175880432128, 0.008570879936218261, 0.008640512466430664, 0.008550399780273438, 0.008629247665405274, 0.008597503662109375, 0.008597503662109375, 0.008578047752380372, 0.00861184024810791, 0.008572928428649903, 0.008588288307189941, 0.008618047714233398, 0.008571840286254883, 0.008563712120056152, 0.008680447578430176, 0.008572928428649903, 0.008588352203369141, 0.008663999557495116, 0.00862003231048584, 0.01839411163330078, 0.00861184024810791, 0.008652799606323243, 0.00862003231048584, 0.008589311599731446, 0.008625151634216309, 0.008624128341674805, 0.008615936279296875, 0.008637439727783204, 0.008630271911621093, 0.008645631790161134, 0.008623104095458984, 0.00861081600189209, 0.008606719970703124, 0.00860262393951416, 0.008651776313781738, 0.008597503662109375, 0.008583168029785156, 0.008632320404052735, 0.008612863540649414, 0.008582176208496094, 0.008625120162963866, 0.008604672431945801, 0.008590335845947266, 0.008582143783569337, 0.008606719970703124, 0.00870809555053711, 0.008592448234558105, 0.008605631828308105, 0.008622079849243165, 0.008631296157836914, 0.008625151634216309, 0.008581119537353516, 0.008600576400756836, 0.008613887786865235, 0.008641535758972169, 0.008607744216918945, 0.00860262393951416, 0.008600576400756836, 0.008572928428649903, 0.008556544303894043, 0.00858726406097412, 0.008565759658813477, 0.008574975967407226, 0.008600576400756836, 0.008747008323669434, 0.008608768463134766, 0.00859340763092041, 0.008775679588317872, 0.008589311599731446, 0.008599552154541015, 0.00861184024810791, 0.008706048011779785, 0.008571904182434082, 0.00862003231048584, 0.008613887786865235, 0.008599552154541015, 0.008588288307189941, 0.008599552154541015, 0.008668160438537598, 0.008604672431945801, 0.008939519882202148, 0.00869478416442871]",tokens/s,114.61475907861896,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Salesforce/codegen-6B-nl,,cuda,0,42,,,,,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,d35829e539df8480b726c647eeabf91e41eae047,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,,MB,1297.416192,15339.094016,0.0,14692.646912,14512.889344,s,10,0.8911674957275391,0.08911674957275391,0.001237067046647154,0.0887234230041504,0.08921959457397462,0.09102051811218262,0.09246125694274902,"[0.09282144165039062, 0.08859903717041015, 0.08870169830322265, 0.0886430435180664, 0.08881938934326172, 0.08879456329345703, 0.08875440216064454, 0.08858707427978516, 0.0887088623046875, 0.08873798370361329]",tokens/s,2872.6361904728637,kWh,1.0506706237206887e-06,5.757208763457813e-07,5.455488631840576e-06,7.081880131907046e-06,tokens/kWh,36148592.63807717,MB,1297.416192,15339.094016,0.0,14692.646912,14645.240832,s,10,25.91175341796875,2.591175341796875,0.007274593906932008,2.5885311279296874,2.602458203125,2.6051935791015626,2.6073818798828126,"[2.601850341796875, 2.584159912109375, 2.58542578125, 2.607928955078125, 2.588089111328125, 2.591695068359375, 2.58709326171875, 2.58897314453125, 2.589697509765625, 2.58684033203125]",tokens/s,24.313290954795846,kWh,3.0617880261665244e-05,1.677915287357284e-05,0.00011977661155375825,0.00016717364468899634,tokens/kWh,376853.6608578635,,s,629,26.66847028732299,0.04239820395440858,0.010574408221097849,0.041076736450195314,0.04147527618408203,0.04160983047485352,0.1300556561279297,"[0.04151193618774414, 0.041899009704589846, 0.04166963195800781, 0.04209151840209961, 0.0419788818359375, 0.041695232391357424, 0.042039295196533204, 0.04196659088134766, 0.041714687347412106, 0.041644031524658204, 0.041599998474121096, 0.041296897888183595, 0.041022464752197264, 0.040959999084472655, 0.04093030548095703, 0.04088524627685547, 0.04094771194458008, 0.04096409606933594, 0.040971263885498044, 0.04093337631225586, 0.04094259262084961, 0.04098252868652344, 0.04096716690063477, 0.04101529693603516, 0.04101017761230469, 0.0409876480102539, 0.040981502532958985, 0.04107571029663086, 0.04100198364257813, 0.041047039031982424, 0.04128768157958984, 0.04102761459350586, 0.0410173110961914, 0.04102963256835938, 0.041106433868408204, 0.04172697448730469, 0.04131737518310547, 0.041059326171875, 0.04113612747192383, 0.04105830383300781, 0.04127129745483398, 0.04105420684814453, 0.04138905715942383, 0.041265151977539063, 0.041194496154785154, 0.04106854248046875, 0.04109721755981445, 0.04117299270629883, 0.04115660858154297, 0.0411228141784668, 0.04115763092041016, 0.04156415939331055, 0.04146585464477539, 0.04125798416137695, 0.04135424041748047, 0.04124364852905273, 0.04123852920532227, 0.041169921875, 0.041619457244873044, 0.042172416687011716, 0.041747455596923826, 0.04160921478271484, 0.13001216125488282, 0.04081356811523437, 0.040831039428710934, 0.04083808135986328, 0.04081356811523437, 0.04084428787231445, 0.04088422393798828, 0.04094464111328125, 0.04084735870361328, 0.04101529693603516, 0.040866817474365234, 0.04092211151123047, 0.04091596984863281, 0.040904705047607424, 0.04088729476928711, 0.04093030548095703, 0.04094464111328125, 0.040880126953125, 0.04090675354003906, 0.04091904067993164, 0.040937473297119144, 0.040927230834960936, 0.040959999084472655, 0.040948734283447266, 0.04096102523803711, 0.04096102523803711, 0.04092211151123047, 0.040948734283447266, 0.04096819305419922, 0.040965118408203126, 0.0409620475769043, 0.04100505447387695, 0.04103680038452148, 0.04098252868652344, 0.040948734283447266, 0.041024513244628906, 0.04109209442138672, 0.04106752014160156, 0.04103475189208984, 0.04111052703857422, 0.04102963256835938, 0.04109209442138672, 0.04103577423095703, 0.04108595275878906, 0.04104806518554688, 0.04113100814819336, 0.04112998580932617, 0.04118425750732422, 0.04102963256835938, 0.041090049743652345, 0.04107468795776367, 0.04111974334716797, 0.04112076950073242, 0.04118937683105469, 0.04115763092041016, 0.04119756698608398, 0.04113510513305664, 0.041163776397705076, 0.041188350677490236, 0.04122937774658203, 0.04118214416503906, 0.041289726257324216, 0.041219070434570314, 0.13010330200195314, 0.040850433349609375, 0.040784896850585936, 0.04090675354003906, 0.04077363204956055, 0.04085760116577149, 0.04087295913696289, 0.040853504180908204, 0.04084633636474609, 0.04088627243041992, 0.04085964965820312, 0.040976383209228515, 0.04088524627685547, 0.04088524627685547, 0.04103577423095703, 0.04148019027709961, 0.04099379348754883, 0.04095590209960937, 0.04097945785522461, 0.04095078277587891, 0.04099379348754883, 0.04092825698852539, 0.040978431701660156, 0.04105420684814453, 0.0409804801940918, 0.040981502532958985, 0.04099686431884766, 0.04098457717895508, 0.04098463821411133, 0.04102137756347656, 0.04100710296630859, 0.041014270782470705, 0.041049087524414066, 0.040978431701660156, 0.04106444931030274, 0.04104806518554688, 0.04097945785522461, 0.041078784942626956, 0.04105215835571289, 0.04108595275878906, 0.0409815673828125, 0.04100806427001953, 0.041133056640625, 0.0410880012512207, 0.04111360168457031, 0.04107164764404297, 0.041124832153320315, 0.04109823989868164, 0.04112793731689453, 0.041186336517333985, 0.04114332962036133, 0.04113919830322266, 0.04117504119873047, 0.041184192657470704, 0.04112588882446289, 0.041145343780517575, 0.041169921875, 0.04117299270629883, 0.04122726440429687, 0.041212928771972655, 0.041186302185058594, 0.04117913436889648, 0.041148414611816404, 0.13007257080078125, 0.041032703399658206, 0.0409354248046875, 0.04086476898193359, 0.04084326553344726, 0.04153139114379883, 0.04157440185546875, 0.041319423675537106, 0.041245697021484375, 0.0412334098815918, 0.04125593566894531, 0.041373695373535156, 0.04137267303466797, 0.04124262237548828, 0.04132659149169922, 0.04137779235839844, 0.04140544128417969, 0.04135935974121094, 0.041365505218505856, 0.04137472152709961, 0.041286655426025394, 0.04134502410888672, 0.04135116958618164, 0.04132761764526367, 0.04133375930786133, 0.04133990478515625, 0.04135321426391601, 0.04148940658569336, 0.041336830139160154, 0.04131840133666992, 0.041606143951416014, 0.04153855895996094, 0.04132460784912109, 0.04137363052368164, 0.04105215835571289, 0.041032703399658206, 0.04142182540893555, 0.0414648323059082, 0.0413675537109375, 0.04143820953369141, 0.041403392791748046, 0.04144332885742188, 0.04151603317260742, 0.04135935974121094, 0.04150783920288086, 0.04145151901245117, 0.04147507095336914, 0.04151910400390625, 0.04146790313720703, 0.04149248123168945, 0.041559040069580076, 0.04156108856201172, 0.04159078216552734, 0.04155801773071289, 0.041842689514160154, 0.041594879150390625, 0.04151193618774414, 0.04155596923828125, 0.041545726776123046, 0.041608192443847655, 0.04157235336303711, 0.04159385681152344, 0.04164710235595703, 0.1302097930908203, 0.040839168548583986, 0.04091801452636719, 0.040858623504638675, 0.04090163040161133, 0.04086579132080078, 0.0409620475769043, 0.04091596984863281, 0.040894462585449216, 0.04084428787231445, 0.04090777587890625, 0.04096921539306641, 0.04090572738647461, 0.04091904067993164, 0.040994815826416016, 0.04098355102539063, 0.04091187286376953, 0.04093030548095703, 0.04094668960571289, 0.04095078277587891, 0.040943614959716795, 0.04094771194458008, 0.040954879760742184, 0.040946720123291015, 0.04094870376586914, 0.04091596984863281, 0.04098355102539063, 0.04093132781982422, 0.040978431701660156, 0.04098559951782227, 0.04103577423095703, 0.040989696502685545, 0.041025535583496094, 0.04104294586181641, 0.04107571029663086, 0.041057281494140625, 0.04111360168457031, 0.0410665283203125, 0.04114838409423828, 0.04110438537597656, 0.04104499053955078, 0.04103680038452148, 0.04108390426635742, 0.041422847747802735, 0.04132352066040039, 0.04112486267089844, 0.0412303352355957, 0.04112998580932617, 0.04116582489013672, 0.04133785629272461, 0.04115561676025391, 0.041141216278076174, 0.04114636611938476, 0.04114739227294922, 0.04120678329467774, 0.04136345672607422, 0.04122623825073242, 0.04128870391845703, 0.041812992095947264, 0.04167679977416992, 0.04122323226928711, 0.04124671936035156, 0.0415610237121582, 0.13032447814941406, 0.040941566467285154, 0.040823806762695314, 0.04089651107788086, 0.040989696502685545, 0.040953857421875, 0.040925182342529294, 0.04103577423095703, 0.04096614456176758, 0.040894462585449216, 0.04103372955322265, 0.04108390426635742, 0.04084633636474609, 0.040910846710205076, 0.041059326171875, 0.04104191970825195, 0.04092006301879883, 0.040992767333984374, 0.04114944076538086, 0.040992767333984374, 0.04098457717895508, 0.041188350677490236, 0.04118527984619141, 0.040941566467285154, 0.0409620475769043, 0.04116582489013672, 0.04099174499511719, 0.04096716690063477, 0.041181182861328124, 0.0411228141784668, 0.041046016693115236, 0.04103680038452148, 0.041207809448242184, 0.04112998580932617, 0.04106547164916992, 0.04122521591186523, 0.04106137466430664, 0.04110233688354492, 0.04119244766235351, 0.041285633087158206, 0.0410695686340332, 0.041060352325439455, 0.041204734802246096, 0.041253887176513675, 0.041076736450195314, 0.04112486267089844, 0.041455615997314454, 0.04121702575683594, 0.04107161712646484, 0.04127743911743164, 0.04127743911743164, 0.041111553192138675, 0.04113510513305664, 0.04132863998413086, 0.04117504119873047, 0.04124467086791992, 0.04147609710693359, 0.0413573112487793, 0.041229312896728515, 0.04140544128417969, 0.04140031814575195, 0.04169830322265625, 0.041984001159667966, 0.13029280090332032, 0.04085446548461914, 0.040855552673339846, 0.04077875137329102, 0.040973312377929685, 0.04106444931030274, 0.0409354248046875, 0.04088115310668945, 0.040855552673339846, 0.04097740936279297, 0.04092927932739258, 0.040910846710205076, 0.0409354248046875, 0.04093548965454102, 0.04100396728515625, 0.04091904067993164, 0.0409354248046875, 0.040925182342529294, 0.04096409606933594, 0.04094668960571289, 0.040941566467285154, 0.04095590209960937, 0.04096416091918945, 0.041028545379638674, 0.040973312377929685, 0.04099071884155273, 0.04102041625976562, 0.040940544128417966, 0.04092211151123047, 0.04111052703857422, 0.04110847854614258, 0.04104499053955078, 0.04099993515014649, 0.04107161712646484, 0.041057281494140625, 0.04105318450927734, 0.041024513244628906, 0.04117299270629883, 0.04117708969116211, 0.04105215835571289, 0.0410695686340332, 0.04110745620727539, 0.04108492660522461, 0.04106854248046875, 0.04101836776733398, 0.04117913436889648, 0.04114022445678711, 0.04109721755981445, 0.04108390426635742, 0.04114227294921875, 0.041133056640625, 0.04117299270629883, 0.041191425323486325, 0.04174643325805664, 0.04155596923828125, 0.041207809448242184, 0.04114432144165039, 0.041183231353759765, 0.041224193572998044, 0.04120883178710937, 0.041181182861328124, 0.04124671936035156, 0.04133171081542969, 0.12998757934570312, 0.040970241546630856, 0.04105420684814453, 0.04087091064453125, 0.0408350715637207, 0.040921089172363284, 0.04095897674560547, 0.040978431701660156, 0.04086579132080078, 0.04096716690063477, 0.04104294586181641, 0.041009151458740234, 0.04097945785522461, 0.04139212799072266, 0.04119244766235351, 0.04095078277587891, 0.0409620475769043, 0.040976383209228515, 0.040940544128417966, 0.04094464111328125, 0.0409169921875, 0.04103475189208984, 0.040925182342529294, 0.04091289520263672, 0.04092211151123047, 0.04097433471679687, 0.04095590209960937, 0.040982593536376954, 0.04105625534057617, 0.04119136047363281, 0.04099379348754883, 0.04095795059204101, 0.04116070556640625, 0.0410511360168457, 0.041169921875, 0.041215999603271485, 0.04149964904785156, 0.04123756790161133, 0.04111148834228515, 0.04113510513305664, 0.04125593566894531, 0.04116889572143555, 0.04113100814819336, 0.04109414291381836, 0.04105318450927734, 0.04126310348510742, 0.041145343780517575, 0.04110745620727539, 0.04106854248046875, 0.04109414291381836, 0.041171966552734376, 0.04112076950073242, 0.04115353775024414, 0.04122828674316406, 0.04119551849365234, 0.04119043350219727, 0.04116169738769531, 0.04120371246337891, 0.0413306884765625, 0.041237503051757815, 0.04117401504516602, 0.04130099105834961, 0.04151603317260742, 0.13013401794433593, 0.041038848876953124, 0.04085760116577149, 0.040840190887451173, 0.04089753723144531, 0.04101939010620117, 0.040888320922851565, 0.040910846710205076, 0.04091187286376953, 0.041022464752197264, 0.040941566467285154, 0.04090265655517578, 0.0409804801940918, 0.040976383209228515, 0.040986625671386716, 0.040890369415283206, 0.04098252868652344, 0.04097740936279297, 0.04183039855957031, 0.04127231979370117, 0.04108697509765625, 0.041134078979492186, 0.04108083343505859, 0.041003009796142575, 0.040992767333984374, 0.04091494369506836, 0.04159283065795898, 0.041219070434570314, 0.0409620475769043, 0.04095283126831055, 0.04106854248046875, 0.041040897369384766, 0.04096307373046875, 0.040976383209228515, 0.04108595275878906, 0.04099686431884766, 0.040981502532958985, 0.04106649780273437, 0.04105625534057617, 0.0410511360168457, 0.041047039031982424, 0.0416102409362793, 0.04168294525146484, 0.04123852920532227, 0.04109619140625, 0.04129894256591797, 0.04116275024414062, 0.041128959655761715, 0.0411228141784668, 0.041150463104248046, 0.04117299270629883, 0.04110540771484375, 0.0410777587890625, 0.041148414611816404, 0.04118425750732422, 0.041191425323486325, 0.04112588882446289, 0.04117299270629883, 0.04122521591186523, 0.041210880279541014, 0.04117299270629883, 0.041356353759765624, 0.041287616729736326, 0.13017701721191408, 0.04085657501220703, 0.04082175827026367, 0.041047039031982424, 0.04100403213500976, 0.04096409606933594, 0.040875007629394534, 0.04092927932739258, 0.040890369415283206, 0.040880126953125, 0.040892414093017575, 0.040899585723876954, 0.04093030548095703, 0.04090675354003906, 0.04099795150756836, 0.040944576263427734, 0.04089344024658203, 0.04096102523803711, 0.040976383209228515, 0.04094566345214844, 0.04090572738647461, 0.04099379348754883, 0.040971263885498044, 0.04093132781982422, 0.04094668960571289, 0.04096819305419922, 0.04109107208251953, 0.04098355102539063, 0.04094566345214844, 0.040995838165283204, 0.041078784942626956, 0.041016319274902346, 0.04097740936279297, 0.0410849609375, 0.041024478912353515, 0.04105318450927734, 0.04101529693603516, 0.04109414291381836, 0.041134078979492186, 0.04114432144165039, 0.041073665618896485, 0.041093120574951174, 0.04111667251586914, 0.04110438537597656, 0.041219070434570314, 0.041452545166015625, 0.04125491333007812, 0.04115558242797852, 0.041133056640625, 0.04110847854614258, 0.04115558242797852, 0.04116889572143555, 0.041218048095703126, 0.041215999603271485, 0.04118841552734375, 0.04119750213623047, 0.04117606353759766, 0.041215999603271485, 0.04122832107543945, 0.04120163345336914, 0.041224193572998044, 0.041319423675537106, 0.04124467086791992]",tokens/s,23.58590474906236,,,main,False,False,,,,,,,,,,,,,,,,,,,,,,, bfloat16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-410m,,cuda,0,42,,,,,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,217063f5c507ed7cc255df7e1f64c4333a0b4dfe,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,,MB,1277.681664,1711.80032,0.0,1065.353216,1018.226688,s,10,0.19355452919006347,0.01935545291900635,0.0005977656089848024,0.019076640129089356,0.019849494552612305,0.020449354934692382,0.020929243240356447,"[0.021049215316772462, 0.019716192245483398, 0.019075616836547852, 0.019038528442382813, 0.019128671646118166, 0.01907747268676758, 0.019044479370117186, 0.019075807571411134, 0.019059680938720704, 0.019288864135742188]",tokens/s,13226.246942979946,kWh,2.2911734102513646e-07,1.2554492998809343e-07,6.189835916550509e-07,9.736458626682808e-07,tokens/kWh,262929274.20081758,MB,1277.681664,1711.80032,0.0,1065.353216,1058.88,s,10,11.631632934570314,1.1631632934570313,0.005588300337279548,1.1620383911132812,1.1683673583984375,1.1721897216796875,1.1752476123046876,"[1.1760120849609375, 1.16357421875, 1.1675179443359376, 1.167354736328125, 1.163515869140625, 1.1561873779296874, 1.1590982666015626, 1.1599454345703124, 1.1605609130859376, 1.1578660888671874]",tokens/s,54.16264453528106,kWh,1.3755305475145452e-05,7.537566972796201e-06,2.363948025954468e-05,4.4932352707486335e-05,tokens/kWh,1402107.750959218,,s,629,11.783421957015998,0.018733580217831468,0.0023476004409559346,0.018391040802001952,0.018720768737792968,0.018957926177978517,0.03788746841430664,"[0.020666368484497072, 0.019337215423583985, 0.018974720001220705, 0.01846681594848633, 0.018428928375244142, 0.018456575393676757, 0.018391040802001952, 0.018716672897338867, 0.0189040641784668, 0.018515968322753908, 0.01840332794189453, 0.01841459274291992, 0.01838387107849121, 0.01842995262145996, 0.01841155242919922, 0.01869513511657715, 0.018958335876464845, 0.018621471405029295, 0.018400224685668945, 0.01839923286437988, 0.01845248031616211, 0.018415615081787108, 0.01840025520324707, 0.01846886444091797, 0.019115007400512696, 0.01882931137084961, 0.018440191268920898, 0.01841459274291992, 0.018364416122436524, 0.018894847869873048, 0.019141632080078123, 0.019273727416992188, 0.019169279098510742, 0.018382848739624022, 0.01841459274291992, 0.01836953544616699, 0.018516992568969725, 0.018944000244140623, 0.018655231475830078, 0.01836031913757324, 0.018921472549438476, 0.018924543380737305, 0.018390016555786134, 0.01841049575805664, 0.018343936920166014, 0.018396160125732423, 0.018471935272216796, 0.01842995262145996, 0.018815999984741212, 0.018938880920410156, 0.018832384109497072, 0.01849760055541992, 0.018367424011230468, 0.01840127944946289, 0.01840332794189453, 0.01904332733154297, 0.018621503829956056, 0.018237375259399415, 0.018311168670654295, 0.018421760559082033, 0.01846784019470215, 0.018340864181518556, 0.03871744155883789, 0.01836031913757324, 0.018267135620117187, 0.018919424057006837, 0.01848320007324219, 0.0184268798828125, 0.018367488861083983, 0.018347007751464844, 0.018320383071899413, 0.018354175567626953, 0.01889795112609863, 0.01828758430480957, 0.018380800247192384, 0.018378751754760742, 0.018334720611572267, 0.018375680923461913, 0.018345983505249023, 0.018317312240600587, 0.018667520523071288, 0.01882931137084961, 0.01860403251647949, 0.018379776000976563, 0.01835212707519531, 0.01866444778442383, 0.01840127944946289, 0.018357248306274415, 0.01842790412902832, 0.018942975997924806, 0.018743295669555664, 0.018339839935302735, 0.018367488861083983, 0.018413568496704103, 0.01847500801086426, 0.018348031997680665, 0.018350080490112306, 0.018852863311767578, 0.018791423797607423, 0.018315263748168945, 0.018386943817138672, 0.018404352188110353, 0.018449407577514648, 0.01846067237854004, 0.018361343383789062, 0.018751487731933594, 0.018917375564575196, 0.018548736572265623, 0.018711551666259766, 0.018421760559082033, 0.018359296798706053, 0.018330623626708984, 0.018354175567626953, 0.018386943817138672, 0.01836953544616699, 0.01836755180358887, 0.018406335830688476, 0.018372608184814454, 0.018322431564331054, 0.018304000854492186, 0.018320383071899413, 0.018397184371948243, 0.018408447265625, 0.018328575134277342, 0.018378751754760742, 0.03801702499389648, 0.01879449653625488, 0.020534271240234374, 0.018957311630249024, 0.018502656936645507, 0.018412544250488282, 0.01841663932800293, 0.018371583938598633, 0.018337791442871093, 0.018380800247192384, 0.018341888427734376, 0.01842278480529785, 0.018380800247192384, 0.018437120437622072, 0.01843609619140625, 0.018374719619750977, 0.018373567581176756, 0.018325504302978517, 0.018397184371948243, 0.018341888427734376, 0.018367488861083983, 0.018355199813842774, 0.018437120437622072, 0.018358272552490236, 0.01840640068054199, 0.018378751754760742, 0.018362367630004883, 0.018336767196655272, 0.018359296798706053, 0.018316287994384766, 0.018377792358398436, 0.018415552139282226, 0.018379776000976563, 0.018356224060058594, 0.01838489532470703, 0.01842790412902832, 0.019042303085327148, 0.020187135696411132, 0.019148799896240236, 0.018917375564575196, 0.018967552185058592, 0.01882316780090332, 0.018400287628173827, 0.01870025634765625, 0.018549760818481444, 0.018413568496704103, 0.018461696624755858, 0.01839820861816406, 0.018404352188110353, 0.018565120697021483, 0.018318336486816408, 0.018363391876220703, 0.018405376434326173, 0.018313215255737304, 0.01839923286437988, 0.018370559692382812, 0.018371583938598633, 0.018365503311157227, 0.01871455955505371, 0.0183767032623291, 0.018340864181518556, 0.018355199813842774, 0.018359296798706053, 0.03789004898071289, 0.018387968063354493, 0.018348031997680665, 0.018518016815185546, 0.018372608184814454, 0.018311168670654295, 0.018344959259033202, 0.018324480056762696, 0.01839411163330078, 0.01840742492675781, 0.018326528549194337, 0.01840332794189453, 0.01841049575805664, 0.018396160125732423, 0.018306047439575195, 0.018396160125732423, 0.018379776000976563, 0.01841049575805664, 0.018387968063354493, 0.01836751937866211, 0.018369504928588867, 0.018465791702270508, 0.018501632690429686, 0.018471935272216796, 0.019573759078979493, 0.0190699520111084, 0.01881804847717285, 0.018549760818481444, 0.018374656677246092, 0.018405439376831055, 0.018399168014526367, 0.018448383331298827, 0.0184268798828125, 0.01840230369567871, 0.018626560211181642, 0.020023296356201172, 0.01943142318725586, 0.018778112411499022, 0.018354175567626953, 0.018780160903930664, 0.01843507194519043, 0.01843507194519043, 0.018449440002441406, 0.018387935638427735, 0.018370559692382812, 0.01843404769897461, 0.018685951232910156, 0.018421760559082033, 0.018737152099609376, 0.018447359085083007, 0.01847500801086426, 0.018397184371948243, 0.01843609619140625, 0.01879756736755371, 0.018545663833618165, 0.018463743209838866, 0.01842790412902832, 0.018508800506591795, 0.018516992568969725, 0.01837772750854492, 0.018557952880859374, 0.018397184371948243, 0.01838591957092285, 0.037897216796875, 0.01840947151184082, 0.018512895584106445, 0.01836031913757324, 0.018345983505249023, 0.018440191268920898, 0.018361343383789062, 0.018412544250488282, 0.018349056243896485, 0.018333696365356447, 0.01817190361022949, 0.018283519744873047, 0.018472959518432617, 0.01840332794189453, 0.018364416122436524, 0.018378751754760742, 0.018601984024047852, 0.018457599639892578, 0.01839411163330078, 0.018345983505249023, 0.018353151321411132, 0.018349056243896485, 0.018362367630004883, 0.01846886444091797, 0.01844223976135254, 0.01839820861816406, 0.018387968063354493, 0.018295808792114256, 0.019204095840454103, 0.018714624404907225, 0.01886412811279297, 0.018185216903686522, 0.018256895065307616, 0.01836953544616699, 0.018557952880859374, 0.01883545684814453, 0.018428928375244142, 0.018330623626708984, 0.0184586238861084, 0.018321407318115233, 0.018367488861083983, 0.01862246322631836, 0.018340864181518556, 0.01838489532470703, 0.018405376434326173, 0.018378751754760742, 0.018540544509887694, 0.01841971206665039, 0.018456575393676757, 0.018359296798706053, 0.018363391876220703, 0.018513952255249023, 0.01849648094177246, 0.018397184371948243, 0.018189311981201172, 0.018159616470336915, 0.01846272087097168, 0.01844326400756836, 0.018176000595092775, 0.018991104125976564, 0.01977446365356445, 0.019125247955322267, 0.018304000854492186, 0.038005760192871094, 0.01841868782043457, 0.018449472427368163, 0.01818822479248047, 0.01818009567260742, 0.018164735794067383, 0.018323455810546875, 0.01838591957092285, 0.018372608184814454, 0.01839923286437988, 0.018306047439575195, 0.018387968063354493, 0.018319360733032225, 0.01840127944946289, 0.01836751937866211, 0.01832547187805176, 0.018375680923461913, 0.018378751754760742, 0.018372608184814454, 0.018413568496704103, 0.018334720611572267, 0.01843612861633301, 0.018380767822265626, 0.01838489532470703, 0.018345983505249023, 0.018373632431030275, 0.018313215255737304, 0.0181790714263916, 0.018370559692382812, 0.018416704177856444, 0.018422719955444335, 0.018412544250488282, 0.018388992309570314, 0.018366464614868162, 0.018329599380493163, 0.01839308738708496, 0.018280448913574218, 0.01841152000427246, 0.018317312240600587, 0.018428928375244142, 0.018359296798706053, 0.018356224060058594, 0.018345983505249023, 0.018356224060058594, 0.018268159866333008, 0.018351104736328124, 0.01846886444091797, 0.01836031913757324, 0.018342912673950194, 0.018311168670654295, 0.018359296798706053, 0.018275327682495117, 0.018371583938598633, 0.018350080490112306, 0.01827840042114258, 0.01837772750854492, 0.018408447265625, 0.01844633674621582, 0.01818623924255371, 0.018137088775634767, 0.018155519485473632, 0.018361343383789062, 0.01836953544616699, 0.03784601593017578, 0.01839411163330078, 0.018324480056762696, 0.018343936920166014, 0.018461696624755858, 0.018326528549194337, 0.018395135879516602, 0.018334720611572267, 0.018340864181518556, 0.018316287994384766, 0.018320383071899413, 0.01902284812927246, 0.018523136138916017, 0.018367488861083983, 0.018404352188110353, 0.018282495498657226, 0.018347007751464844, 0.018365440368652345, 0.01833990478515625, 0.018400192260742188, 0.018323455810546875, 0.018327552795410155, 0.01835113525390625, 0.0184616641998291, 0.018373632431030275, 0.018349056243896485, 0.018358272552490236, 0.018322431564331054, 0.018481151580810547, 0.018315263748168945, 0.018453504562377928, 0.018345983505249023, 0.018354175567626953, 0.018355199813842774, 0.018362367630004883, 0.0182794246673584, 0.01845145606994629, 0.018647071838378906, 0.01848111915588379, 0.018379776000976563, 0.01842278480529785, 0.01835212707519531, 0.01829478454589844, 0.018513919830322266, 0.01844121551513672, 0.01836851119995117, 0.018469888687133788, 0.01840230369567871, 0.018457599639892578, 0.018456575393676757, 0.01837772750854492, 0.018397184371948243, 0.018397184371948243, 0.01836953544616699, 0.018307104110717773, 0.018311136245727538, 0.018408447265625, 0.018336767196655272, 0.01836031913757324, 0.018338815689086914, 0.018357248306274415, 0.018362367630004883, 0.018363391876220703, 0.03804569625854492, 0.018342912673950194, 0.01838591957092285, 0.018374656677246092, 0.01836031913757324, 0.018349056243896485, 0.018324480056762696, 0.01836953544616699, 0.01842790412902832, 0.018472959518432617, 0.018428928375244142, 0.01842790412902832, 0.018370559692382812, 0.018381824493408205, 0.018333696365356447, 0.01843404769897461, 0.018423807144165038, 0.01841766357421875, 0.01843507194519043, 0.018361343383789062, 0.018357248306274415, 0.018339839935302735, 0.01841868782043457, 0.01839923286437988, 0.018396160125732423, 0.018437183380126954, 0.018374591827392577, 0.018437120437622072, 0.018371583938598633, 0.01838595199584961, 0.018382816314697265, 0.018455551147460936, 0.01840742492675781, 0.01837772750854492, 0.018330623626708984, 0.01841152000427246, 0.018374656677246092, 0.01842995262145996, 0.018295839309692384, 0.018826208114624023, 0.01867366409301758, 0.018693119049072265, 0.018490367889404297, 0.018447359085083007, 0.018340864181518556, 0.01843814468383789, 0.01841766357421875, 0.018356224060058594, 0.01838489532470703, 0.01841868782043457, 0.01839308738708496, 0.018362367630004883, 0.018373632431030275, 0.018592767715454102, 0.01841152000427246, 0.018440191268920898, 0.0184770565032959, 0.01842585563659668, 0.01823846435546875, 0.018206720352172853, 0.01825279998779297, 0.01819545555114746, 0.01842585563659668, 0.03788083267211914, 0.01844633674621582, 0.01842995262145996, 0.018349056243896485, 0.018354175567626953, 0.01844326400756836, 0.018382848739624022, 0.018482175827026368, 0.018416671752929686, 0.018434015274047852, 0.018391040802001952, 0.018388992309570314, 0.01844633674621582, 0.01840947151184082, 0.018707456588745116, 0.018741247177124023, 0.018423807144165038, 0.01846886444091797, 0.01840742492675781, 0.01841049575805664, 0.018453504562377928, 0.01838387107849121, 0.0184770565032959, 0.01841152000427246, 0.01843609619140625, 0.018446367263793947, 0.018409439086914062, 0.01847603225708008, 0.018461696624755858, 0.018404352188110353, 0.01843916893005371, 0.018373632431030275, 0.01842995262145996, 0.018375680923461913, 0.018448383331298827, 0.018372608184814454, 0.01845043182373047, 0.01841663932800293, 0.01841663932800293, 0.018336767196655272, 0.01841766357421875, 0.018338815689086914, 0.018373632431030275, 0.018333696365356447, 0.01840742492675781, 0.01840025520324707, 0.01841766357421875, 0.018378751754760742, 0.01843404769897461, 0.018382848739624022, 0.018372608184814454, 0.018336767196655272, 0.01841766357421875, 0.018356224060058594, 0.01844326400756836, 0.01840332794189453, 0.018351104736328124, 0.018397184371948243, 0.018349056243896485, 0.018363391876220703, 0.018350080490112306, 0.018353151321411132, 0.018438207626342774, 0.038718399047851564, 0.018365440368652345, 0.018356224060058594, 0.018354175567626953, 0.018331680297851562, 0.01830601692199707, 0.018322431564331054, 0.018358272552490236, 0.018327552795410155, 0.018372608184814454, 0.018341888427734376, 0.018340864181518556, 0.01835212707519531, 0.018366464614868162, 0.018292736053466797, 0.018318336486816408, 0.018347007751464844, 0.018359296798706053, 0.018376768112182616, 0.018367424011230468, 0.018275327682495117, 0.018296831130981444, 0.018347007751464844, 0.018363391876220703, 0.018343936920166014, 0.018367488861083983, 0.018405376434326173, 0.018439231872558595, 0.01833568000793457, 0.018343936920166014, 0.018341888427734376, 0.01839411163330078, 0.018386943817138672, 0.018473983764648438, 0.01843507194519043, 0.01841766357421875, 0.018336767196655272, 0.018299903869628906, 0.018306047439575195, 0.01837772750854492, 0.0183767032623291, 0.01841459274291992, 0.01832863998413086, 0.01837664031982422, 0.018342912673950194, 0.018481151580810547, 0.01840233612060547, 0.018336736679077148, 0.01836031913757324, 0.018495487213134765, 0.018322431564331054, 0.018353151321411132, 0.018367488861083983, 0.018397184371948243, 0.018336767196655272, 0.0184268798828125, 0.018344959259033202, 0.018361343383789062, 0.018388992309570314, 0.018326559066772462, 0.018369504928588867, 0.018367488861083983, 0.018348031997680665]",tokens/s,53.38007942807189,,,main,False,False,,,,,,,,,,,,,,,,,,,,,,, float32-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,.,.,cuda,0,42,,,True,True,True,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 373, in cached_file raise EnvironmentError( OSError: . does not appear to have a file named config.json. Checkout 'https://huggingface.co/./tree/None' for available files. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,a,a,cuda,0,42,,,True,True,True,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/a/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66947484-653353db0d98d04951dd1a65;9f97ec32-b6a0-4739-9cde-33681cb7411f) Repository Not Found for url: https://huggingface.co/a/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: a is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-7B,Qwen/Qwen-7B,cuda,0,42,,,True,,,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",qwen,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-66b,facebook/opt-66b,cuda,0,42,,,True,,,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",opt,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,stabilityai/stablelm-base-alpha-7b,stabilityai/stablelm-base-alpha-7b,cuda,0,42,,,True,True,True,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ self.load_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 250, in load_model_with_no_weights self.pretrained_model = self.automodel_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3710, in from_pretrained model = cls(config, *model_args, **model_kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 1028, in __init__ self.gpt_neox = GPTNeoXModel(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 853, in __init__ self.layers = nn.ModuleList([GPTNeoXLayer(config) for _ in range(config.num_hidden_layers)]) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 853, in self.layers = nn.ModuleList([GPTNeoXLayer(config) for _ in range(config.num_hidden_layers)]) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 740, in __init__ self.attention = GPT_NEOX_ATTENTION_CLASSES[config._attn_implementation](config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 508, in __init__ super().__init__(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 118, in __init__ self.query_key_value = nn.Linear(config.hidden_size, 3 * config.hidden_size, bias=config.attention_bias) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/linear.py"", line 98, in __init__ self.weight = Parameter(torch.empty((out_features, in_features), **factory_kwargs)) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_device.py"", line 78, in __torch_function__ return func(*args, **kwargs) torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 432.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,1,1,cuda,0,42,,,True,True,True,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/1/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66932602-39db661c609f613d014e89bc;564bfd11-feaa-4f50-bff1-5c6096aa94c3) Repository Not Found for url: https://huggingface.co/1/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: 1 is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-7b,tiiuae/falcon-7b,cuda,0,42,,,True,True,True,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ self.load_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 250, in load_model_with_no_weights self.pretrained_model = self.automodel_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 559, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3704, in from_pretrained config = cls._autoset_attn_implementation( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1490, in _autoset_attn_implementation config = cls._check_and_enable_sdpa( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1656, in _check_and_enable_sdpa raise ValueError( ValueError: FalconForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,meta-llama/Llama-2-7b-hf,meta-llama/Llama-2-7b-hf,cuda,0,42,,,True,True,True,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ self.load_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 250, in load_model_with_no_weights self.pretrained_model = self.automodel_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3710, in from_pretrained model = cls(config, *model_args, **model_kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 1101, in __init__ self.model = LlamaModel(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 884, in __init__ [LlamaDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 884, in [LlamaDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 679, in __init__ self.mlp = LlamaMLP(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 192, in __init__ self.gate_proj = nn.Linear(self.hidden_size, self.intermediate_size, bias=config.mlp_bias) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/linear.py"", line 98, in __init__ self.weight = Parameter(torch.empty((out_features, in_features), **factory_kwargs)) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_device.py"", line 78, in __torch_function__ return func(*args, **kwargs) torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 172.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-rw-1b,tiiuae/falcon-rw-1b,cuda,0,42,,,True,True,True,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ self.load_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 250, in load_model_with_no_weights self.pretrained_model = self.automodel_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 559, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3704, in from_pretrained config = cls._autoset_attn_implementation( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1490, in _autoset_attn_implementation config = cls._check_and_enable_sdpa( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1656, in _check_and_enable_sdpa raise ValueError( ValueError: FalconForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,8,8,cuda,0,42,,,True,True,True,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/8/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66932432-3e9f3529644e2dfc603560cb;67b75eb5-fa4c-4933-a588-a7168fb5e666) Repository Not Found for url: https://huggingface.co/8/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: 8 is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-4B,,cuda,0,42,,,,,True,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,217063f5c507ed7cc255df7e1f64c4333a0b4dfe,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,,MB,1160.220672,18656.78848,0.0,18016.632832,17752.787968,s,10,1.174385383605957,0.1174385383605957,0.0003681859792099211,0.11736227035522462,0.11791879425048829,0.11805218200683594,0.11815889221191406,"[0.11818556976318359, 0.11754528045654297, 0.117067138671875, 0.11702988433837891, 0.11717536163330078, 0.11788915252685547, 0.11717926025390625, 0.11713085174560547, 0.11758019256591797, 0.11760269165039063]",tokens/s,2179.863642494856,kWh,1.4235753152105545e-06,7.800499649935103e-07,9.915941667686497e-06,1.2119566947890562e-05,tokens/kWh,21122866.939115956,MB,1160.220672,18656.78848,0.0,18016.632832,17960.87552,s,10,24.073228027343745,2.4073228027343747,0.0010902935442557537,2.4070374755859376,2.4080315429687498,2.40911611328125,2.4099837695312503,"[2.4076298828125, 2.4066884765625, 2.407550048828125, 2.40779052734375, 2.40717236328125, 2.406765380859375, 2.41020068359375, 2.406902587890625, 2.4064716796875, 2.406056396484375]",tokens/s,26.17015047937942,kWh,2.834813452429242e-05,1.553591605152485e-05,0.00011275951758371694,0.0001566435681595342,tokens/kWh,402186.95692527527,,s,629,25.08991483688357,0.039888576847191645,0.014112227762690815,0.03822895812988281,0.03832627105712891,0.038351910400390626,0.15672442810058596,"[0.03825356674194336, 0.038095870971679685, 0.038073345184326174, 0.03815628814697265, 0.03817881774902344, 0.038059009552001956, 0.0380579833984375, 0.03807436752319336, 0.03811532974243164, 0.03805286407470703, 0.03803033447265625, 0.038125568389892575, 0.038329345703125, 0.03822182464599609, 0.038084606170654296, 0.038046718597412106, 0.03806003189086914, 0.03806412887573242, 0.03802624130249024, 0.038084606170654296, 0.03811123275756836, 0.03807846450805664, 0.03805286407470703, 0.038042625427246096, 0.03805593490600586, 0.03806105422973633, 0.03805081558227539, 0.0380682258605957, 0.03806617736816406, 0.03809075164794922, 0.03812454223632813, 0.03824025726318359, 0.038196224212646485, 0.03824947357177735, 0.03820032119750977, 0.03824639892578125, 0.038234111785888675, 0.038346752166748044, 0.03829043197631836, 0.038242305755615234, 0.03829043197631836, 0.03829043197631836, 0.03828326416015625, 0.038284286499023434, 0.03829452896118164, 0.038284286499023434, 0.03826998519897461, 0.038293472290039064, 0.03825766372680664, 0.038312961578369144, 0.03827711868286133, 0.038307838439941407, 0.03829862213134766, 0.038307838439941407, 0.038312961578369144, 0.03830886459350586, 0.03834265518188477, 0.03830988693237305, 0.03830579376220703, 0.038317054748535154, 0.038307838439941407, 0.03834163284301758, 0.15708364868164063, 0.03829452896118164, 0.03809695816040039, 0.03807328033447266, 0.03805491256713867, 0.03807743835449219, 0.03809382247924804, 0.03808262252807617, 0.03807839965820312, 0.03808563232421875, 0.03803238296508789, 0.03804159927368164, 0.03804569625854492, 0.03802828979492188, 0.03806412887573242, 0.038027263641357424, 0.03810406494140625, 0.03809075164794922, 0.03806825637817383, 0.03803340911865234, 0.03804771041870117, 0.03803750228881836, 0.038063102722167966, 0.0380682258605957, 0.03812659072875976, 0.03807846450805664, 0.03806105422973633, 0.03810201644897461, 0.0381624641418457, 0.03806000137329101, 0.038095870971679685, 0.03806105422973633, 0.03827097702026367, 0.03820544052124023, 0.038250495910644534, 0.0382371826171875, 0.03827507019042969, 0.0382371826171875, 0.03825971221923828, 0.03826278305053711, 0.038250495910644534, 0.03827199935913086, 0.03827302551269531, 0.03827199935913086, 0.03830374526977539, 0.038258689880371094, 0.03833958435058594, 0.03824639892578125, 0.03829759979248047, 0.03827507019042969, 0.038277183532714844, 0.038287296295166015, 0.038338558197021484, 0.03830169677734375, 0.038284286499023434, 0.03829555130004883, 0.03832012939453125, 0.03829145431518555, 0.03829452896118164, 0.03829967880249024, 0.03833033752441406, 0.038286334991455076, 0.03835193634033203, 0.15671705627441407, 0.03831903839111328, 0.03808358383178711, 0.03812659072875976, 0.03808873748779297, 0.03809584045410156, 0.03806515121459961, 0.038076416015625, 0.03808153533935547, 0.03807743835449219, 0.038035457611083984, 0.03804467010498047, 0.0380579833984375, 0.038110271453857425, 0.03817465591430664, 0.03803852844238281, 0.03803852844238281, 0.03805286407470703, 0.03805699157714844, 0.038034400939941405, 0.03805286407470703, 0.03806617736816406, 0.03810508728027344, 0.03806825637817383, 0.03805283355712891, 0.038076416015625, 0.03813888168334961, 0.03805491256713867, 0.038079486846923825, 0.03807129669189453, 0.038150142669677735, 0.03806515121459961, 0.038247425079345705, 0.03820748901367187, 0.038234111785888675, 0.03820851135253906, 0.038242305755615234, 0.03822489547729492, 0.03827097702026367, 0.03826483154296875, 0.0382371826171875, 0.03825459289550781, 0.03825971221923828, 0.03823923110961914, 0.03825766372680664, 0.038286334991455076, 0.038362110137939456, 0.03827711868286133, 0.038296577453613284, 0.03828121566772461, 0.03828326416015625, 0.038266880035400394, 0.03830988693237305, 0.03841024017333984, 0.03853107070922852, 0.038386688232421876, 0.03832217788696289, 0.038319103240966795, 0.03834777450561523, 0.03829452896118164, 0.038351871490478515, 0.03831808090209961, 0.038391807556152346, 0.15717990112304686, 0.03860889434814453, 0.03807231903076172, 0.03805593490600586, 0.03806105422973633, 0.03809280014038086, 0.0380682258605957, 0.038040576934814455, 0.03807436752319336, 0.03805184173583984, 0.03806105422973633, 0.03803750228881836, 0.03808153533935547, 0.03803955078125, 0.03808870315551758, 0.03801599884033203, 0.03810201644897461, 0.03803852844238281, 0.038076416015625, 0.03801497650146484, 0.03806412887573242, 0.038046718597412106, 0.03808563232421875, 0.03803238296508789, 0.03809075164794922, 0.03805081558227539, 0.03808774566650391, 0.038186943054199216, 0.03825254440307617, 0.03810201644897461, 0.03808358383178711, 0.03809280014038086, 0.03823513412475586, 0.038300670623779294, 0.03825254440307617, 0.03829350280761719, 0.038284320831298825, 0.03825766372680664, 0.03825148773193359, 0.03825363159179687, 0.038252479553222654, 0.03829145431518555, 0.03831398391723633, 0.03825254440307617, 0.038346752166748044, 0.03823923110961914, 0.03827206420898437, 0.03823507308959961, 0.03825664138793945, 0.03827097702026367, 0.03826483154296875, 0.03828940963745117, 0.0383109130859375, 0.03836108779907227, 0.03828531265258789, 0.038286334991455076, 0.038312961578369144, 0.03830988693237305, 0.038346752166748044, 0.03828736114501953, 0.03830886459350586, 0.038340606689453126, 0.038329345703125, 0.156727294921875, 0.038284286499023434, 0.038255615234375, 0.03809382247924804, 0.038130687713623046, 0.03808051300048828, 0.0381214714050293, 0.03806208038330078, 0.038100990295410156, 0.038046718597412106, 0.03806515121459961, 0.038007808685302735, 0.03807743835449219, 0.03816550445556641, 0.03807027053833008, 0.03801804733276367, 0.03807436752319336, 0.0380211181640625, 0.03812966537475586, 0.03817062377929688, 0.03805388641357422, 0.03802624130249024, 0.03809894561767578, 0.038059009552001956, 0.038079486846923825, 0.03805388641357422, 0.03808768081665039, 0.03806924819946289, 0.038076416015625, 0.03809689712524414, 0.03807027053833008, 0.03808768081665039, 0.03823206329345703, 0.03823820877075195, 0.03822079849243164, 0.03823212814331055, 0.038252479553222654, 0.03820646286010742, 0.03825971221923828, 0.03823923110961914, 0.03828940963745117, 0.03821363067626953, 0.03829043197631836, 0.038330368041992184, 0.03829248046875, 0.038245376586914064, 0.03828531265258789, 0.038286334991455076, 0.03830681610107422, 0.03829043197631836, 0.03832115173339844, 0.03827916717529297, 0.03832115173339844, 0.03830988693237305, 0.03832729721069336, 0.03829452896118164, 0.03834368133544922, 0.03833651351928711, 0.03832115173339844, 0.03828531265258789, 0.03832524871826172, 0.03834163284301758, 0.038312961578369144, 0.15750245666503906, 0.03831398391723633, 0.038100990295410156, 0.03802624130249024, 0.038079486846923825, 0.0380682258605957, 0.03806719970703125, 0.03807027053833008, 0.03808256149291992, 0.038059009552001956, 0.0380313606262207, 0.037997566223144534, 0.038040576934814455, 0.03800883102416992, 0.0380948486328125, 0.038147071838378906, 0.03804774475097656, 0.038024192810058595, 0.03805388641357422, 0.03801804733276367, 0.03805081558227539, 0.03803955078125, 0.038056961059570314, 0.03804367828369141, 0.0380610237121582, 0.03808256149291992, 0.03809280014038086, 0.038133758544921875, 0.03806719970703125, 0.03807846450805664, 0.03805184173583984, 0.03806617736816406, 0.038250495910644534, 0.03826892852783203, 0.038228992462158204, 0.03824025726318359, 0.03821670532226563, 0.03825151824951172, 0.03832729721069336, 0.03828736114501953, 0.038422527313232424, 0.03830886459350586, 0.03825459289550781, 0.0382371826171875, 0.038258689880371094, 0.038284286499023434, 0.03827916717529297, 0.03828326416015625, 0.03827097702026367, 0.03828227233886719, 0.03829654312133789, 0.03830374526977539, 0.03834470367431641, 0.03829350280761719, 0.03834368133544922, 0.03828326416015625, 0.03829145431518555, 0.038286334991455076, 0.03831398391723633, 0.03832115173339844, 0.038351871490478515, 0.03833958435058594, 0.03830988693237305, 0.15692697143554687, 0.03840518569946289, 0.03817055892944336, 0.03811532974243164, 0.03806924819946289, 0.038100990295410156, 0.038063102722167966, 0.03803955078125, 0.03808665466308594, 0.03805081558227539, 0.03802422332763672, 0.03803644943237305, 0.038042625427246096, 0.03803238296508789, 0.038040576934814455, 0.03804064178466797, 0.03802310562133789, 0.038052894592285155, 0.03804771041870117, 0.03803955078125, 0.03803340911865234, 0.038089729309082034, 0.038040576934814455, 0.038121536254882814, 0.038064064025878905, 0.03809075164794922, 0.03803958511352539, 0.03811734390258789, 0.038075393676757815, 0.03809791946411133, 0.03806719970703125, 0.03811635208129883, 0.038245376586914064, 0.03824028778076172, 0.0382432975769043, 0.038245376586914064, 0.03825766372680664, 0.0382740478515625, 0.0399554557800293, 0.03853311920166016, 0.03849836730957031, 0.038320064544677734, 0.03834163284301758, 0.03833958435058594, 0.0384983024597168, 0.03837747192382813, 0.03829759979248047, 0.038242305755615234, 0.038265857696533206, 0.038302719116210936, 0.03826176071166992, 0.038296577453613284, 0.03830886459350586, 0.03836108779907227, 0.038317054748535154, 0.03831500625610352, 0.038319103240966795, 0.03827814483642578, 0.038317054748535154, 0.038406143188476564, 0.03852799987792969, 0.03851881790161133, 0.038350814819335936, 0.15666490173339845, 0.03823199844360352, 0.0380948486328125, 0.03806412887573242, 0.03809382247924804, 0.03804774475097656, 0.03808668899536133, 0.03805283355712891, 0.03811532974243164, 0.03804467010498047, 0.03805286407470703, 0.03803443145751953, 0.03805388641357422, 0.03803033447265625, 0.03808768081665039, 0.038005760192871094, 0.03807129669189453, 0.03801190567016602, 0.03808563232421875, 0.038079486846923825, 0.0380497932434082, 0.0380579833984375, 0.03805388641357422, 0.03804159927368164, 0.038266880035400394, 0.038593536376953126, 0.03810406494140625, 0.03806719970703125, 0.03809996795654297, 0.038079486846923825, 0.03806208038330078, 0.03807743835449219, 0.0382371826171875, 0.03822694396972656, 0.03821977615356445, 0.03820032119750977, 0.038212608337402344, 0.03823513412475586, 0.03828736114501953, 0.03824639892578125, 0.03827097702026367, 0.038250495910644534, 0.03825664138793945, 0.03824332809448242, 0.03827302551269531, 0.03825356674194336, 0.03827097702026367, 0.038266880035400394, 0.03826483154296875, 0.03832627105712891, 0.03830988693237305, 0.03827609634399414, 0.03829555130004883, 0.038267902374267575, 0.038302719116210936, 0.03826278305053711, 0.03830169677734375, 0.03830579376220703, 0.03832729721069336, 0.03839487838745117, 0.038302719116210936, 0.038299648284912106, 0.03832627105712891, 0.15713381958007813, 0.03827711868286133, 0.03812659072875976, 0.03803647994995117, 0.03812249755859375, 0.03806719970703125, 0.0381317138671875, 0.03806208038330078, 0.038084606170654296, 0.03806412887573242, 0.03803443145751953, 0.03803443145751953, 0.03803852844238281, 0.03802009582519531, 0.038059009552001956, 0.03802316665649414, 0.038040576934814455, 0.03805388641357422, 0.03805286407470703, 0.038024192810058595, 0.03805286407470703, 0.03802828979492188, 0.038059009552001956, 0.03802732849121094, 0.03805382537841797, 0.038059009552001956, 0.0380682258605957, 0.038084606170654296, 0.03806719970703125, 0.038089729309082034, 0.038073345184326174, 0.038089729309082034, 0.038232097625732424, 0.03822895812988281, 0.03825151824951172, 0.03820646286010742, 0.03822079849243164, 0.038217727661132815, 0.03825459289550781, 0.038247425079345705, 0.03828224182128906, 0.03823923110961914, 0.03826176071166992, 0.03822489547729492, 0.03827199935913086, 0.038247425079345705, 0.038432769775390625, 0.038484992980957033, 0.038307838439941407, 0.03825766372680664, 0.03828531265258789, 0.038263809204101565, 0.03833651351928711, 0.03827507019042969, 0.03831808090209961, 0.038296577453613284, 0.03830476760864258, 0.03829043197631836, 0.03831808090209961, 0.03832524871826172, 0.03830579376220703, 0.03830681610107422, 0.03834163284301758, 0.15718194580078124, 0.03828531265258789, 0.03810201644897461, 0.0380497932434082, 0.038073345184326174, 0.038075454711914064, 0.03810297775268555, 0.03806105422973633, 0.03807231903076172, 0.03804774475097656, 0.03802624130249024, 0.038012928009033206, 0.03802521514892578, 0.038019073486328124, 0.03803955078125, 0.038010879516601564, 0.03803340911865234, 0.0380211181640625, 0.03804159927368164, 0.038013950347900394, 0.038043647766113284, 0.03803340911865234, 0.03804569625854492, 0.0380211181640625, 0.03805388641357422, 0.03805081558227539, 0.03807436752319336, 0.038095870971679685, 0.03805593490600586, 0.03809075164794922, 0.03807846450805664, 0.03815935897827148, 0.03825356674194336, 0.038324222564697266, 0.03824639892578125, 0.03824540710449219, 0.038219745635986326, 0.03826483154296875, 0.03830169677734375, 0.03826176071166992, 0.03828121566772461, 0.038255615234375, 0.038250495910644534, 0.03823308944702149, 0.038263809204101565, 0.03823820877075195, 0.03825667190551758, 0.03825968170166016, 0.03824947357177735, 0.0382740478515625, 0.03827916717529297, 0.03829555130004883, 0.03827199935913086, 0.038300670623779294, 0.038299648284912106, 0.03827302551269531, 0.03831398391723633, 0.03832627105712891, 0.03829759979248047, 0.038267902374267575, 0.038300670623779294, 0.03834982299804687, 0.03831193542480469]",tokens/s,25.069833998612687,,,main,False,False,,,,,,,,,,,,,,,,,,,,,,, float32-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,internlm/internlm2-20b,internlm/internlm2-20b,cuda,0,42,,,True,,,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",internlm2,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,i,i,cuda,0,42,,,True,True,True,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/i/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669474ba-41ace83e737fa42b3701f939;8a02723a-b3e0-4fbe-935e-7171972b67d6) Repository Not Found for url: https://huggingface.co/i/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: i is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,stabilityai/stablelm-3b-4e1t,stabilityai/stablelm-3b-4e1t,cuda,0,42,,,True,True,True,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1138.663424,12505.841664,0.0,11865.686016,11620.236288,s,10,0.8953843536376954,0.08953843536376953,0.0005007823305073723,0.0896009292602539,0.0900410270690918,0.09027113914489747,0.09045522880554199,"[0.09050125122070313, 0.08883433532714843, 0.08888851165771484, 0.0890060806274414, 0.0893661117553711, 0.08958678436279296, 0.08961507415771484, 0.0899898910522461, 0.08976191711425781, 0.08983439636230468]",tokens/s,2859.1073650097173,kWh,1.0913446682308792e-06,5.980059409871459e-07,7.663048943889849e-06,9.352399553107876e-06,tokens/kWh,27372654.3168196,MB,1138.663424,12505.841664,0.0,11865.686016,11816.868352,s,10,18.09325134277344,1.8093251342773438,0.0034958578847460696,1.8082963256835938,1.8097242553710937,1.814743963623047,1.8187597302246095,"[1.819763671875, 1.8083280029296875, 1.8080596923828125, 1.8074124755859375, 1.8083851318359374, 1.80781689453125, 1.8080611572265626, 1.8085509033203124, 1.8086087646484375, 1.8082646484375]",tokens/s,34.81961246570678,kWh,2.12664901080117e-05,1.1654451651398178e-05,8.844563905411209e-05,0.00012136658081352198,tokens/kWh,519088.529788926,,s,629,18.86789333915711,0.02999665077767425,0.010752799533975921,0.02872217559814453,0.028827853012084962,0.028974860382080077,0.11874865234375,"[0.02944819259643555, 0.02935603141784668, 0.029410303115844725, 0.02919628715515137, 0.02934988784790039, 0.02930892753601074, 0.02930790328979492, 0.02921062469482422, 0.029287424087524414, 0.029328384399414063, 0.02928748893737793, 0.029151168823242188, 0.02902739143371582, 0.028851167678833008, 0.029152223587036133, 0.029071359634399413, 0.02876518440246582, 0.02855833625793457, 0.02858700752258301, 0.028556287765502928, 0.028579839706420897, 0.02857676887512207, 0.028602367401123048, 0.02857676887512207, 0.028753919601440428, 0.028618751525878908, 0.02857676887512207, 0.028588064193725587, 0.02862179183959961, 0.028621824264526367, 0.028589056015014647, 0.028692480087280273, 0.028708864212036132, 0.028697599411010744, 0.028770303726196288, 0.028721151351928712, 0.02872831916809082, 0.02871500778198242, 0.028711936950683595, 0.028735488891601563, 0.02879078483581543, 0.028815359115600587, 0.028875776290893555, 0.028743711471557618, 0.028741600036621094, 0.028795904159545898, 0.0288143367767334, 0.028826623916625976, 0.02877644729614258, 0.02876825523376465, 0.028804096221923828, 0.02875187110900879, 0.028759040832519532, 0.028787712097167968, 0.02892902374267578, 0.028932096481323243, 0.028799999237060548, 0.028837888717651368, 0.028778495788574218, 0.028812288284301758, 0.028901376724243165, 0.028793855667114256, 0.11871129608154297, 0.028834815979003905, 0.028648448944091798, 0.02857779121398926, 0.028556287765502928, 0.0285665283203125, 0.028646400451660156, 0.028598272323608398, 0.02855219268798828, 0.028589056015014647, 0.028651519775390624, 0.028669952392578125, 0.02858700752258301, 0.028612607955932616, 0.02854604721069336, 0.02854297637939453, 0.02853376007080078, 0.02856345558166504, 0.028531711578369142, 0.02855526351928711, 0.028572671890258788, 0.028579839706420897, 0.02857472038269043, 0.028614656448364258, 0.02855936050415039, 0.02857062339782715, 0.028589056015014647, 0.028606464385986328, 0.028641279220581056, 0.02865974426269531, 0.02859107208251953, 0.028593151092529297, 0.02879897689819336, 0.028870655059814454, 0.028721151351928712, 0.02873958396911621, 0.028710912704467774, 0.02872319984436035, 0.028725248336791992, 0.02878156852722168, 0.028705791473388673, 0.02873139190673828, 0.02874982452392578, 0.02889727973937988, 0.02899046325683594, 0.028831743240356447, 0.028803071975708007, 0.028750848770141602, 0.02875801658630371, 0.028794879913330077, 0.02874367904663086, 0.0287774715423584, 0.02876108741760254, 0.0287774715423584, 0.02879283142089844, 0.028753919601440428, 0.0287324161529541, 0.02875699234008789, 0.02873244857788086, 0.028798944473266603, 0.02875596809387207, 0.02877337646484375, 0.02874060821533203, 0.11872358703613281, 0.02882252883911133, 0.028638208389282226, 0.028596223831176756, 0.02859110450744629, 0.02858188819885254, 0.02855731201171875, 0.02855526351928711, 0.028556287765502928, 0.028564479827880858, 0.02856959915161133, 0.0286167049407959, 0.028693504333496093, 0.028589056015014647, 0.028590080261230468, 0.028610559463500978, 0.02854400062561035, 0.028564479827880858, 0.02857779121398926, 0.028572671890258788, 0.028553216934204102, 0.02856755256652832, 0.028564479827880858, 0.028590112686157226, 0.028569568634033204, 0.02860032081604004, 0.028625919342041017, 0.028631040573120117, 0.02856755256652832, 0.02856345558166504, 0.02857574462890625, 0.028627967834472655, 0.028795904159545898, 0.028864511489868162, 0.028886016845703126, 0.028795936584472655, 0.028867551803588867, 0.02873958396911621, 0.02871500778198242, 0.028724288940429686, 0.028708799362182617, 0.02874060821533203, 0.028759040832519532, 0.028712959289550782, 0.0287324161529541, 0.02874675178527832, 0.028753919601440428, 0.028806175231933594, 0.028761056900024413, 0.028750848770141602, 0.028753919601440428, 0.02876620864868164, 0.02876825523376465, 0.028787712097167968, 0.028763200759887694, 0.028756927490234375, 0.02875395202636719, 0.02877641677856445, 0.02879283142089844, 0.02875187110900879, 0.028759040832519532, 0.02876518440246582, 0.02876108741760254, 0.1187583999633789, 0.028806175231933594, 0.028635103225708006, 0.02860748863220215, 0.028536832809448243, 0.028542015075683595, 0.028544960021972657, 0.02854617691040039, 0.02855615997314453, 0.02855731201171875, 0.02858393669128418, 0.02855014419555664, 0.02856959915161133, 0.0285296630859375, 0.028522592544555664, 0.028571552276611328, 0.028807167053222657, 0.02877132797241211, 0.028538976669311523, 0.02853264045715332, 0.02852454376220703, 0.028550239562988283, 0.02854390335083008, 0.02857676887512207, 0.028564672470092773, 0.02853356742858887, 0.02854400062561035, 0.02855423927307129, 0.02856550407409668, 0.02859212875366211, 0.028603456497192384, 0.028627904891967773, 0.02869862365722656, 0.028708864212036132, 0.028705791473388673, 0.02874470329284668, 0.02882252883911133, 0.0287457275390625, 0.028719104766845704, 0.0287324161529541, 0.028718080520629883, 0.02875289535522461, 0.028762111663818358, 0.02873139190673828, 0.02876006317138672, 0.028733440399169922, 0.02874982452392578, 0.02880102348327637, 0.02875289535522461, 0.02876825523376465, 0.02874880027770996, 0.0287774715423584, 0.02879795265197754, 0.028794879913330077, 0.02877644729614258, 0.028804096221923828, 0.02874777603149414, 0.028831743240356447, 0.028770303726196288, 0.02876620864868164, 0.028762111663818358, 0.028799999237060548, 0.028801055908203126, 0.11917820739746093, 0.029079551696777343, 0.028626943588256838, 0.02856755256652832, 0.0285347843170166, 0.02855219268798828, 0.02858291244506836, 0.0285665283203125, 0.02859519958496094, 0.0285614070892334, 0.02858598327636719, 0.028630016326904296, 0.02857369613647461, 0.02858291244506836, 0.02858188819885254, 0.02858700752258301, 0.02854297637939453, 0.0285665283203125, 0.02853785514831543, 0.02855219268798828, 0.02857062339782715, 0.028642303466796876, 0.028625919342041017, 0.028627967834472655, 0.0285296630859375, 0.028553216934204102, 0.02855116844177246, 0.02857881546020508, 0.028611583709716795, 0.0286167049407959, 0.02857574462890625, 0.028564479827880858, 0.02873855972290039, 0.02875596809387207, 0.028969024658203123, 0.028978111267089844, 0.02875494384765625, 0.02876620864868164, 0.028711936950683595, 0.02876108741760254, 0.02878156852722168, 0.02873958396911621, 0.02876313591003418, 0.02872422409057617, 0.02876620864868164, 0.028787712097167968, 0.028753919601440428, 0.028750848770141602, 0.028750848770141602, 0.02878463935852051, 0.028838911056518556, 0.02875494384765625, 0.028762111663818358, 0.02875596809387207, 0.02877132797241211, 0.028815359115600587, 0.02877440071105957, 0.02879283142089844, 0.02876620864868164, 0.02878156852722168, 0.028820480346679687, 0.02875494384765625, 0.02875801658630371, 0.1193359375, 0.028793855667114256, 0.028620800018310546, 0.028588031768798827, 0.02855014419555664, 0.028596223831176756, 0.02854297637939453, 0.02854604721069336, 0.028543071746826174, 0.02854185676574707, 0.02857369613647461, 0.02858598327636719, 0.02876006317138672, 0.02878361511230469, 0.0285665283203125, 0.028621824264526367, 0.02859929656982422, 0.02862387275695801, 0.028651519775390624, 0.028609535217285157, 0.02855936050415039, 0.02855833625793457, 0.02875494384765625, 0.02857881546020508, 0.02860748863220215, 0.028638208389282226, 0.028553216934204102, 0.0285849609375, 0.02857676887512207, 0.028580863952636718, 0.02858700752258301, 0.028612607955932616, 0.02872217559814453, 0.02871500778198242, 0.028701696395874023, 0.028708864212036132, 0.0287324161529541, 0.02876416015625, 0.028726303100585937, 0.028717023849487305, 0.028711936950683595, 0.028734464645385743, 0.02875699234008789, 0.02874470329284668, 0.028729343414306642, 0.02874163246154785, 0.02873855972290039, 0.02877235221862793, 0.028778495788574218, 0.02875699234008789, 0.02874060821533203, 0.02874880027770996, 0.02876736068725586, 0.02878553581237793, 0.028727296829223634, 0.02874777603149414, 0.028759040832519532, 0.02877337646484375, 0.02877542304992676, 0.02880102348327637, 0.02878976058959961, 0.02877145576477051, 0.02875276756286621, 0.1193493423461914, 0.028752799987792968, 0.028579839706420897, 0.02858393669128418, 0.02853785514831543, 0.028536832809448243, 0.02856038475036621, 0.028636159896850585, 0.02857676887512207, 0.028593151092529297, 0.02857164764404297, 0.02854707145690918, 0.028536832809448243, 0.028553216934204102, 0.0285347843170166, 0.02853273582458496, 0.02856857681274414, 0.028590080261230468, 0.02855219268798828, 0.02855833625793457, 0.02856038475036621, 0.028577823638916016, 0.028570592880249022, 0.028627967834472655, 0.02860748863220215, 0.02857574462890625, 0.02855423927307129, 0.028593151092529297, 0.028628992080688476, 0.028646400451660156, 0.028612607955932616, 0.02861568069458008, 0.028702720642089844, 0.02872012710571289, 0.028703744888305665, 0.02872319984436035, 0.02879692840576172, 0.028725248336791992, 0.02872319984436035, 0.02872319984436035, 0.028735488891601563, 0.02877337646484375, 0.028734464645385743, 0.02876620864868164, 0.02873651123046875, 0.028729343414306642, 0.028785663604736327, 0.028838911056518556, 0.028767232894897462, 0.028770303726196288, 0.028803071975708007, 0.028835840225219726, 0.028932096481323243, 0.02875494384765625, 0.02878054428100586, 0.02878156852722168, 0.028793920516967775, 0.028791744232177733, 0.02877542304992676, 0.02876416015625, 0.02878156852722168, 0.02895564842224121, 0.02905708885192871, 0.11985504150390625, 0.0287825927734375, 0.028598272323608398, 0.02856755256652832, 0.02854911994934082, 0.028598272323608398, 0.028599327087402343, 0.028550111770629882, 0.02860032081604004, 0.02855833625793457, 0.028564479827880858, 0.02857676887512207, 0.028588031768798827, 0.02857676887512207, 0.028539903640747072, 0.02855219268798828, 0.02857369613647461, 0.028545024871826172, 0.02856857681274414, 0.02859110450744629, 0.028588031768798827, 0.02857472038269043, 0.02854707145690918, 0.02857472038269043, 0.0285665283203125, 0.02858198356628418, 0.028593055725097655, 0.028618751525878908, 0.028589056015014647, 0.02859212875366211, 0.02857676887512207, 0.028632064819335938, 0.02875596809387207, 0.02876518440246582, 0.028704767227172853, 0.028725248336791992, 0.028713983535766603, 0.02872217559814453, 0.02874675178527832, 0.028735488891601563, 0.02872217559814453, 0.02873139190673828, 0.028729343414306642, 0.02874163246154785, 0.028815359115600587, 0.028821504592895508, 0.028778560638427736, 0.02882348823547363, 0.028827648162841796, 0.02880512046813965, 0.02905606460571289, 0.029043647766113283, 0.02896895980834961, 0.028851200103759765, 0.02877235221862793, 0.028799999237060548, 0.02877440071105957, 0.028785663604736327, 0.028821504592895508, 0.02876313591003418, 0.028750848770141602, 0.02877952003479004, 0.02879897689819336, 0.11958271789550781, 0.028827648162841796, 0.0286167049407959, 0.02855219268798828, 0.02855014419555664, 0.028553216934204102, 0.02855116844177246, 0.02856755256652832, 0.02859929656982422, 0.02859110450744629, 0.02855731201171875, 0.02856755256652832, 0.02858291244506836, 0.02857062339782715, 0.0285614070892334, 0.02858598327636719, 0.02856038475036621, 0.028589056015014647, 0.02855014419555664, 0.0285614070892334, 0.02855833625793457, 0.028588064193725587, 0.02859823989868164, 0.02860032081604004, 0.02856038475036621, 0.02856755256652832, 0.028621824264526367, 0.028589056015014647, 0.028602367401123048, 0.028618751525878908, 0.02860851287841797, 0.02859929656982422, 0.028711936950683595, 0.028708864212036132, 0.02894540786743164, 0.02932326316833496, 0.02896998405456543, 0.02873651123046875, 0.02873139190673828, 0.028759040832519532, 0.028721151351928712, 0.02872831916809082, 0.028726272583007813, 0.028729343414306642, 0.02873958396911621, 0.028762111663818358, 0.02878156852722168, 0.028762111663818358, 0.02876108741760254, 0.0287825927734375, 0.028804096221923828, 0.02877132797241211, 0.0287825927734375, 0.028802047729492186, 0.028839935302734376, 0.028853248596191407, 0.02876927947998047, 0.02876006317138672, 0.028759040832519532, 0.028785663604736327, 0.02879692840576172, 0.02876108741760254, 0.02875699234008789, 0.11962470245361329, 0.028828672409057617, 0.028604415893554686, 0.028593151092529297, 0.028589056015014647, 0.02855014419555664, 0.02855731201171875, 0.02855731201171875, 0.028606464385986328, 0.02855423927307129, 0.0285665283203125, 0.028609535217285157, 0.02855731201171875, 0.02852556800842285, 0.02853887939453125, 0.02855014419555664, 0.028575775146484374, 0.0285644474029541, 0.02856755256652832, 0.02857779121398926, 0.02855014419555664, 0.02857779121398926, 0.02857779121398926, 0.02857676887512207, 0.028635135650634767, 0.028580863952636718, 0.02857676887512207, 0.0285614070892334, 0.028598272323608398, 0.028642303466796876, 0.028596223831176756, 0.02860748863220215, 0.02874163246154785, 0.028705791473388673, 0.028701696395874023, 0.02876927947998047, 0.028721151351928712, 0.02875494384765625, 0.02876825523376465, 0.02877644729614258, 0.02873859214782715, 0.02878460884094238, 0.028818431854248046, 0.02876108741760254, 0.0287324161529541, 0.02876416015625, 0.0287457275390625, 0.02877952003479004, 0.02877132797241211, 0.02882252883911133, 0.028762111663818358, 0.028865535736083983, 0.029043712615966798, 0.028887039184570314, 0.02877337646484375, 0.02875289535522461, 0.028753919601440428, 0.028835840225219726, 0.028838911056518556, 0.02879078483581543, 0.02876416015625, 0.028767232894897462, 0.028824640274047852]",tokens/s,33.33705510697463,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,huggyllama/llama-13b,huggyllama/llama-13b,cuda,0,42,,,True,,,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",llama,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-2.7b,facebook/opt-2.7b,cuda,0,42,,,True,,,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",opt,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,microsoft/phi-1_5,microsoft/phi-1_5,cuda,0,42,,,True,True,True,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1111.207936,6635.913216,0.0,5995.757568,5850.446848,s,10,0.5192879981994629,0.05192879981994629,0.0003536356024242777,0.05182969665527344,0.05213985481262207,0.0525259916305542,0.0528349010848999,"[0.05291212844848633, 0.05195967864990234, 0.051644351959228514, 0.051617534637451175, 0.05178771209716797, 0.05175680160522461, 0.05187168121337891, 0.051945472717285154, 0.05205404663085938, 0.05173859024047851]",tokens/s,4929.827010977216,kWh,6.312132956147476e-07,3.458736419187071e-07,4.418233440021314e-06,5.39532037755477e-06,tokens/kWh,47448526.145915836,MB,1111.53152,6635.913216,0.0,5995.757568,5949.83168,s,10,11.423619506835937,1.1423619506835938,0.015494070722493689,1.1383552246093749,1.1664203979492187,1.1705165954589845,1.173793553466797,"[1.1655101318359375, 1.1348154296875, 1.1288631591796876, 1.13693115234375, 1.139779296875, 1.17461279296875, 1.147938232421875, 1.1311065673828125, 1.12259326171875, 1.141469482421875]",tokens/s,55.14889563881269,kWh,1.329687149178052e-05,7.286300949080351e-06,4.823730868437829e-05,6.882048112523917e-05,tokens/kWh,915425.1607941089,,s,629,11.870597097396844,0.01887217344578196,0.006216391706346474,0.017894399642944335,0.018617132568359375,0.018752960205078124,0.07001046844482423,"[0.018082815170288084, 0.017992704391479493, 0.01887948799133301, 0.01847500801086426, 0.018266111373901366, 0.01820876884460449, 0.017976320266723633, 0.017993728637695314, 0.01784524726867676, 0.017861631393432616, 0.017821695327758787, 0.017802240371704102, 0.017838079452514647, 0.018083839416503905, 0.018381824493408205, 0.018371583938598633, 0.018317312240600587, 0.018326528549194337, 0.018345983505249023, 0.018522111892700196, 0.018334720611572267, 0.01824051284790039, 0.018287616729736327, 0.01840742492675781, 0.018335744857788085, 0.018382848739624022, 0.01779916763305664, 0.01781452751159668, 0.019116031646728517, 0.02211737632751465, 0.02030182456970215, 0.019090431213378906, 0.01861222457885742, 0.018685951232910156, 0.018564096450805666, 0.01864908790588379, 0.018509824752807616, 0.018588672637939452, 0.018510847091674804, 0.018516992568969725, 0.01863680076599121, 0.01860710334777832, 0.01864908790588379, 0.01863987159729004, 0.018464767456054687, 0.018551807403564453, 0.018465791702270508, 0.018541568756103514, 0.01880166435241699, 0.01882521629333496, 0.01868083190917969, 0.018568191528320312, 0.017853439331054686, 0.01790771293640137, 0.018508800506591795, 0.01870342445373535, 0.018460607528686522, 0.01861529541015625, 0.01843609619140625, 0.018541568756103514, 0.018530303955078126, 0.018533376693725585, 0.07158169555664062, 0.018902015686035157, 0.0186562557220459, 0.018553888320922852, 0.018570207595825197, 0.01860812759399414, 0.018573312759399413, 0.017854463577270507, 0.017822719573974608, 0.017902591705322265, 0.017847295761108398, 0.01788211250305176, 0.017863679885864257, 0.017903615951538086, 0.01785241508483887, 0.017821695327758787, 0.017839168548583983, 0.017867712020874022, 0.01783193588256836, 0.017854463577270507, 0.017822719573974608, 0.017763328552246094, 0.018010112762451173, 0.017786880493164063, 0.017821695327758787, 0.017821760177612306, 0.017844160079956053, 0.017993728637695314, 0.018647039413452148, 0.018570240020751954, 0.01856716728210449, 0.018337791442871093, 0.017892351150512697, 0.017821695327758787, 0.01796403121948242, 0.01781452751159668, 0.017821695327758787, 0.017878015518188475, 0.017847295761108398, 0.017855487823486327, 0.01784524726867676, 0.017822719573974608, 0.01798860740661621, 0.017923072814941408, 0.01783705520629883, 0.017864704132080078, 0.017902591705322265, 0.01788211250305176, 0.017894399642944335, 0.017829887390136717, 0.017889280319213868, 0.017950719833374023, 0.017836032867431642, 0.01782067108154297, 0.017913856506347657, 0.017926143646240233, 0.01787494468688965, 0.017861631393432616, 0.017960960388183594, 0.017908735275268553, 0.01785139274597168, 0.01786777687072754, 0.017928192138671875, 0.07001190185546875, 0.01774284744262695, 0.01768550491333008, 0.017718271255493166, 0.017681407928466796, 0.017720319747924804, 0.017722368240356445, 0.017687551498413084, 0.017707008361816406, 0.017657855987548828, 0.01770086479187012, 0.01770806312561035, 0.01769059181213379, 0.017833984375, 0.01782067108154297, 0.017884159088134767, 0.017803264617919923, 0.017819648742675782, 0.017830911636352538, 0.01782476806640625, 0.01775103950500488, 0.017812480926513673, 0.017847295761108398, 0.018341888427734376, 0.01855897521972656, 0.018712575912475587, 0.018577407836914063, 0.018534400939941405, 0.0184586238861084, 0.01861631965637207, 0.018463743209838866, 0.018548736572265623, 0.018345983505249023, 0.018150400161743165, 0.017689599990844726, 0.017901567459106444, 0.017794048309326172, 0.01782681655883789, 0.01780121612548828, 0.017903615951538086, 0.017838079452514647, 0.01782681655883789, 0.017811456680297853, 0.01783705520629883, 0.017803264617919923, 0.01783296012878418, 0.017812480926513673, 0.01784320068359375, 0.01782681655883789, 0.01780633544921875, 0.01779097557067871, 0.017829887390136717, 0.01787392044067383, 0.01785958480834961, 0.01783705520629883, 0.017830911636352538, 0.01782067108154297, 0.01764659118652344, 0.01764659118652344, 0.017884159088134767, 0.0177838077545166, 0.017787904739379884, 0.01780121612548828, 0.06979993438720702, 0.017745920181274414, 0.01784217643737793, 0.017827840805053712, 0.017869823455810546, 0.01780940818786621, 0.01787596893310547, 0.017908735275268553, 0.017709056854248048, 0.017641471862792968, 0.017673215866088866, 0.01759539222717285, 0.017680383682250975, 0.017691648483276368, 0.017726463317871095, 0.01764761543273926, 0.017868799209594728, 0.01782067108154297, 0.01787494468688965, 0.017785856246948242, 0.01785753631591797, 0.01779199981689453, 0.01782476806640625, 0.017707008361816406, 0.017718271255493166, 0.017771520614624024, 0.017984512329101563, 0.01778483200073242, 0.01782476806640625, 0.01778892707824707, 0.01780735969543457, 0.01764556884765625, 0.017746944427490235, 0.017689599990844726, 0.017688575744628905, 0.01761587142944336, 0.01770086479187012, 0.017625087738037108, 0.017811456680297853, 0.017885183334350584, 0.017758207321166994, 0.018339839935302735, 0.018453504562377928, 0.01849241638183594, 0.018697216033935548, 0.018546688079833985, 0.01856716728210449, 0.018554880142211915, 0.0185231990814209, 0.018141120910644532, 0.018380800247192384, 0.01857535934448242, 0.018514944076538087, 0.018505727767944336, 0.018512895584106445, 0.01858355140686035, 0.018557952880859374, 0.018577407836914063, 0.018654207229614257, 0.018545663833618165, 0.018485248565673826, 0.01846272087097168, 0.019483648300170898, 0.07017881774902343, 0.01780735969543457, 0.018017280578613282, 0.017886207580566405, 0.017898496627807618, 0.017941503524780272, 0.01787596893310547, 0.01785036849975586, 0.017947647094726564, 0.017789951324462892, 0.01765273666381836, 0.017737728118896484, 0.017738752365112305, 0.017683456420898438, 0.01782374382019043, 0.01785856056213379, 0.017961984634399415, 0.017861631393432616, 0.017861631393432616, 0.017905664443969727, 0.017861631393432616, 0.01787596893310547, 0.01787494468688965, 0.017951744079589844, 0.01781657600402832, 0.01777459144592285, 0.01765478324890137, 0.017725439071655275, 0.017675264358520508, 0.017674240112304687, 0.017757183074951173, 0.017878015518188475, 0.018140159606933593, 0.018112512588500978, 0.017912832260131836, 0.017862655639648437, 0.01789132881164551, 0.017963008880615236, 0.017903615951538086, 0.018293792724609376, 0.018533344268798827, 0.018288639068603514, 0.017744895935058593, 0.01785139274597168, 0.017827840805053712, 0.018695167541503906, 0.018556928634643553, 0.019104768753051758, 0.01856716728210449, 0.018353151321411132, 0.018685951232910156, 0.018540544509887694, 0.01820364761352539, 0.018541568756103514, 0.018522111892700196, 0.018556928634643553, 0.018512895584106445, 0.01864192008972168, 0.01862860870361328, 0.018545696258544922, 0.018581472396850585, 0.018684928894042968, 0.018549760818481444, 0.07054847717285156, 0.018756671905517577, 0.018483135223388673, 0.018587648391723634, 0.018496511459350586, 0.018559999465942383, 0.01862144088745117, 0.01860812759399414, 0.01860710334777832, 0.018576383590698242, 0.018482175827026368, 0.01861529541015625, 0.018535423278808593, 0.01863270378112793, 0.018482175827026368, 0.0184268798828125, 0.018115583419799804, 0.018374656677246092, 0.01829478454589844, 0.018511871337890624, 0.01861631965637207, 0.018511871337890624, 0.018824192047119142, 0.01860508728027344, 0.018563039779663087, 0.018538496017456055, 0.018579456329345705, 0.018472959518432617, 0.018531328201293946, 0.018590784072875975, 0.018366399765014647, 0.018298879623413086, 0.01844326400756836, 0.018594816207885743, 0.018577407836914063, 0.018375680923461913, 0.018981887817382814, 0.018912256240844725, 0.01969254493713379, 0.0189040641784668, 0.01921331214904785, 0.01865216064453125, 0.018589696884155273, 0.018573312759399413, 0.018546688079833985, 0.019531776428222656, 0.018783231735229493, 0.018747392654418944, 0.018571296691894532, 0.0185118408203125, 0.018551807403564453, 0.01821900749206543, 0.01864806365966797, 0.01865727996826172, 0.018528255462646484, 0.018543615341186523, 0.01857535934448242, 0.01862348747253418, 0.018707456588745116, 0.020135936737060548, 0.018868223190307617, 0.018717695236206054, 0.01861427116394043, 0.07025357055664062, 0.017932287216186525, 0.01786675262451172, 0.017854463577270507, 0.017895423889160156, 0.01782476806640625, 0.017894399642944335, 0.017943552017211914, 0.017740800857543947, 0.017864704132080078, 0.017731584548950196, 0.017959936141967774, 0.017884159088134767, 0.01784217643737793, 0.017942527770996093, 0.017819648742675782, 0.017939456939697264, 0.01783193588256836, 0.017944576263427735, 0.01783296012878418, 0.01781657600402832, 0.017884159088134767, 0.017769472122192383, 0.017846271514892577, 0.017872896194458008, 0.01780838394165039, 0.01788313674926758, 0.018570240020751954, 0.018496511459350586, 0.018489376068115234, 0.01852207946777344, 0.018497535705566406, 0.018521087646484375, 0.018487295150756835, 0.018530303955078126, 0.018499584197998048, 0.018545663833618165, 0.018431999206542968, 0.018553855895996094, 0.018532352447509767, 0.01864089584350586, 0.018465791702270508, 0.018588672637939452, 0.018570240020751954, 0.018535423278808593, 0.018516992568969725, 0.018548736572265623, 0.018518016815185546, 0.018510847091674804, 0.018506752014160157, 0.01857535934448242, 0.018502656936645507, 0.01861222457885742, 0.018506752014160157, 0.018541568756103514, 0.01857535934448242, 0.018565120697021483, 0.018480127334594726, 0.0182108154296875, 0.01823846435546875, 0.01804902458190918, 0.01804083251953125, 0.018198528289794923, 0.07034880065917969, 0.017811456680297853, 0.017839103698730468, 0.017847295761108398, 0.01788313674926758, 0.01783705520629883, 0.017872896194458008, 0.017912832260131836, 0.01785139274597168, 0.0178657283782959, 0.017900543212890627, 0.017797119140625, 0.01785036849975586, 0.017884159088134767, 0.017886207580566405, 0.017898496627807618, 0.017900543212890627, 0.0178606071472168, 0.0178288631439209, 0.01787494468688965, 0.01779814338684082, 0.0178155517578125, 0.01780940818786621, 0.01785036849975586, 0.017888256072998047, 0.017876991271972655, 0.017904640197753906, 0.01783296012878418, 0.017824800491333007, 0.01782268714904785, 0.01786675262451172, 0.017821695327758787, 0.017886207580566405, 0.017830911636352538, 0.017864704132080078, 0.017838079452514647, 0.017892351150512697, 0.017725439071655275, 0.017757183074951173, 0.017925119400024413, 0.01776639938354492, 0.017696767807006835, 0.01764454460144043, 0.017690624237060547, 0.017667072296142578, 0.01762611198425293, 0.017758207321166994, 0.01781452751159668, 0.0178288631439209, 0.017869823455810546, 0.017804288864135744, 0.017797119140625, 0.018274303436279296, 0.018562047958374024, 0.0184586238861084, 0.01866547203063965, 0.018550783157348632, 0.018511871337890624, 0.018530303955078126, 0.018536447525024414, 0.01848736000061035, 0.018489280700683595, 0.01846272087097168, 0.07037235260009765, 0.017750015258789064, 0.017758207321166994, 0.01780121612548828, 0.017804288864135744, 0.01781657600402832, 0.017763328552246094, 0.017935359954833984, 0.017894399642944335, 0.017876991271972655, 0.017770496368408203, 0.01775923156738281, 0.01789132881164551, 0.01779916763305664, 0.017770496368408203, 0.017755136489868165, 0.017760255813598632, 0.017911808013916015, 0.01776742362976074, 0.01776742362976074, 0.017796096801757814, 0.017770496368408203, 0.017761280059814453, 0.017822719573974608, 0.017762304306030274, 0.01785036849975586, 0.017855487823486327, 0.017765375137329103, 0.017756160736083985, 0.01783193588256836, 0.01779097557067871, 0.017819648742675782, 0.01780735969543457, 0.017836032867431642, 0.01785036849975586, 0.01780121612548828, 0.017871871948242187, 0.017785856246948242, 0.017773567199707033, 0.017777664184570312, 0.01783705520629883, 0.01780944061279297, 0.01785545539855957, 0.01780633544921875, 0.017777664184570312, 0.018118656158447266, 0.017861631393432616, 0.01785139274597168, 0.017779712677001954, 0.017894399642944335, 0.017694719314575197, 0.017705984115600586, 0.0177838077545166, 0.01782374382019043, 0.01782374382019043, 0.017881120681762695, 0.017771488189697267, 0.017787967681884766, 0.017796031951904295, 0.017853439331054686, 0.017771520614624024, 0.01777561569213867, 0.01781350326538086, 0.07000678253173828, 0.017943552017211914, 0.017840160369873046, 0.018070528030395508, 0.017927135467529297, 0.017753087997436523, 0.017846271514892577, 0.017780736923217775, 0.017846271514892577, 0.01838591957092285, 0.01859993553161621, 0.018589696884155273, 0.018420736312866212, 0.018473983764648438, 0.018524160385131837, 0.017780736923217775, 0.01781657600402832, 0.017786880493164063, 0.01785753631591797, 0.01778892707824707, 0.01780735969543457, 0.01780019187927246, 0.01780121612548828, 0.017910783767700195, 0.01778278350830078, 0.01780838394165039, 0.01882009506225586, 0.01904128074645996, 0.018693119049072265, 0.018549760818481444, 0.018542591094970702, 0.018481151580810547, 0.017927167892456054, 0.01784524726867676, 0.017878015518188475, 0.018503679275512695, 0.017802240371704102, 0.018053119659423827, 0.017861631393432616, 0.017861631393432616, 0.01786675262451172, 0.017869823455810546, 0.017863679885864257, 0.017986560821533205, 0.017853439331054686, 0.017880064010620117, 0.017953792572021485, 0.017811456680297853, 0.017969152450561524, 0.017846271514892577, 0.017887231826782226, 0.017868799209594728, 0.017936384201049805, 0.017855487823486327, 0.01860095977783203, 0.018553855895996094, 0.018559999465942383, 0.01846067237854004, 0.018660383224487306, 0.018620384216308593, 0.018518016815185546, 0.018489343643188477, 0.018508800506591795]",tokens/s,52.98806747791446,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,google/recurrentgemma-2b,google/recurrentgemma-2b,cuda,0,42,,,True,,,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",recurrent_gemma,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,stabilityai/stablelm-2-1_6b,stabilityai/stablelm-2-1_6b,cuda,0,42,,,True,True,True,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1140.084736,7590.117376,0.0,6949.961728,6804.719616,s,10,0.5006790390014648,0.050067903900146496,0.00021801737872100758,0.050056127548217776,0.050273246383666995,0.05040283184051514,0.050506500205993655,"[0.05053241729736328, 0.04974985504150391, 0.05010054397583008, 0.050112926483154296, 0.04992403030395508, 0.05001171112060547, 0.05024444961547852, 0.04996432113647461, 0.05022534561157226, 0.04981343841552734]",tokens/s,5113.056071022197,kWh,6.257487843136168e-07,3.4287514085701243e-07,4.363943820370283e-06,5.332567745540912e-06,tokens/kWh,48006891.27935918,MB,1140.084736,7590.117376,0.0,6949.961728,6904.4864,s,10,11.425099365234374,1.1425099365234375,0.010515235760784074,1.139412780761719,1.1487585693359375,1.1606250122070312,1.1701181665039062,"[1.172491455078125, 1.1392100830078125, 1.14612158203125, 1.1408779296875, 1.1381038818359375, 1.13300341796875, 1.1394222412109376, 1.140939453125, 1.1394033203125, 1.1355260009765624]",tokens/s,55.14175236996516,kWh,1.3590661998288481e-05,7.447281955385169e-06,5.087762815062976e-05,7.191557210430343e-05,tokens/kWh,876027.2380038548,,s,629,11.85613618469239,0.0188491831235173,0.0059958730071987595,0.01808076858520508,0.01849671630859375,0.018830361557006837,0.0684296826171875,"[0.019128320693969726, 0.018716672897338867, 0.018827327728271483, 0.019075008392333986, 0.018792448043823243, 0.019017791748046874, 0.01915078353881836, 0.018762752532958983, 0.019168256759643554, 0.018784255981445314, 0.018940927505493164, 0.01922870445251465, 0.01889686393737793, 0.019079168319702147, 0.01863372802734375, 0.01882624053955078, 0.018778240203857422, 0.018760576248168945, 0.01909760093688965, 0.019019775390625, 0.018817024230957033, 0.018926591873168946, 0.018637823104858398, 0.018710527420043945, 0.018832384109497072, 0.019763200759887696, 0.019982336044311523, 0.020032512664794923, 0.019074047088623047, 0.018647039413452148, 0.018619392395019533, 0.018579456329345705, 0.018534400939941405, 0.018496511459350586, 0.01822412872314453, 0.01807360076904297, 0.01818726348876953, 0.018151487350463867, 0.018004928588867188, 0.018173952102661133, 0.018177024841308592, 0.018158592224121094, 0.01805414390563965, 0.018174976348876954, 0.018119680404663087, 0.01805721664428711, 0.018755584716796874, 0.01863577651977539, 0.018514944076538087, 0.01878835105895996, 0.018344959259033202, 0.017921024322509766, 0.017903615951538086, 0.017950719833374023, 0.017906688690185548, 0.017898496627807618, 0.018148351669311523, 0.018111488342285157, 0.01798259162902832, 0.018066303253173828, 0.018177024841308592, 0.018106367111206053, 0.06830079650878906, 0.018106367111206053, 0.018067455291748045, 0.017963008880615236, 0.018077695846557617, 0.018069503784179687, 0.01807360076904297, 0.01803264045715332, 0.01860095977783203, 0.018140159606933593, 0.017985536575317384, 0.01814630317687988, 0.01815449523925781, 0.018115583419799804, 0.01819340705871582, 0.018121728897094725, 0.018110464096069336, 0.017941503524780272, 0.01788211250305176, 0.017958911895751953, 0.017770496368408203, 0.017876991271972655, 0.018058240890502928, 0.0180992317199707, 0.01808483123779297, 0.018107391357421874, 0.01794867134094238, 0.018137088775634767, 0.01806540870666504, 0.018068479537963866, 0.01798041534423828, 0.018128896713256838, 0.018043903350830077, 0.01803264045715332, 0.018027519226074217, 0.01807257652282715, 0.018084863662719726, 0.018000896453857423, 0.018060287475585936, 0.018131967544555663, 0.0180316162109375, 0.018078720092773438, 0.018043935775756834, 0.018157535552978516, 0.018124799728393554, 0.01805516815185547, 0.018113536834716795, 0.018082815170288084, 0.0179814395904541, 0.0180633602142334, 0.018091007232666014, 0.01824460792541504, 0.018130943298339842, 0.018135040283203126, 0.01816582489013672, 0.018078655242919923, 0.018099199295043944, 0.018081792831420897, 0.018086912155151368, 0.01812588882446289, 0.018087871551513673, 0.018076671600341796, 0.01803264045715332, 0.06877184295654297, 0.01818009567260742, 0.01815449523925781, 0.018140159606933593, 0.01817087936401367, 0.018125823974609375, 0.018109439849853515, 0.018101247787475586, 0.01813811111450195, 0.01815449523925781, 0.018084863662719726, 0.018122751235961913, 0.018136064529418947, 0.018203680038452148, 0.018161632537841796, 0.018106367111206053, 0.01817087936401367, 0.018110464096069336, 0.018158592224121094, 0.019107839584350587, 0.01982975959777832, 0.018743295669555664, 0.018694143295288086, 0.01846067237854004, 0.018070528030395508, 0.01810950469970703, 0.018100160598754883, 0.01823846435546875, 0.018116607666015624, 0.017953792572021485, 0.017936384201049805, 0.01789030456542969, 0.018586624145507814, 0.018497535705566406, 0.018386943817138672, 0.018117631912231445, 0.018117631912231445, 0.018074623107910158, 0.018149375915527344, 0.018030624389648437, 0.018158559799194337, 0.018096128463745118, 0.018109439849853515, 0.018089984893798827, 0.018109439849853515, 0.018053119659423827, 0.018092031478881835, 0.018106367111206053, 0.01804083251953125, 0.01808896064758301, 0.018084863662719726, 0.018103296279907227, 0.018119680404663087, 0.018060287475585936, 0.0179814395904541, 0.018111488342285157, 0.018111488342285157, 0.018145280838012694, 0.018089984893798827, 0.01802444839477539, 0.017958911895751953, 0.01807974433898926, 0.017923072814941408, 0.06848921966552735, 0.017969152450561524, 0.018156543731689453, 0.018137088775634767, 0.018114559173583983, 0.018157567977905274, 0.018149375915527344, 0.018142240524291992, 0.01813603210449219, 0.01801523208618164, 0.018109439849853515, 0.018190336227416993, 0.018131967544555663, 0.018111488342285157, 0.018132991790771484, 0.018110464096069336, 0.018100223541259765, 0.018130943298339842, 0.018118656158447266, 0.018069503784179687, 0.018155519485473632, 0.01804902458190918, 0.018118719100952148, 0.018103231430053712, 0.018107391357421874, 0.018086912155151368, 0.018100223541259765, 0.01800499153137207, 0.018089984893798827, 0.018149375915527344, 0.018139135360717772, 0.018145280838012694, 0.01820876884460449, 0.018140159606933593, 0.018197504043579102, 0.018129919052124025, 0.018044927597045898, 0.018062335968017578, 0.018112512588500978, 0.01804083251953125, 0.018036735534667968, 0.018121728897094725, 0.01802444839477539, 0.018093055725097656, 0.018089984893798827, 0.01804800033569336, 0.01816985511779785, 0.018128896713256838, 0.018111488342285157, 0.018183168411254884, 0.018137088775634767, 0.01802444839477539, 0.01807257652282715, 0.018106367111206053, 0.018100223541259765, 0.01807974433898926, 0.018100223541259765, 0.018089984893798827, 0.018069503784179687, 0.018041856765747072, 0.018118656158447266, 0.018075647354125975, 0.018131967544555663, 0.06849740600585938, 0.018106367111206053, 0.018107391357421874, 0.01798963165283203, 0.018081792831420897, 0.01807974433898926, 0.018174976348876954, 0.0180316162109375, 0.01810534477233887, 0.017978368759155275, 0.017912832260131836, 0.017935359954833984, 0.01790771293640137, 0.01802649688720703, 0.018109439849853515, 0.018089984893798827, 0.018075647354125975, 0.01807974433898926, 0.018053119659423827, 0.018087968826293946, 0.01807459259033203, 0.018111488342285157, 0.018136064529418947, 0.018098175048828127, 0.018067455291748045, 0.01820979118347168, 0.018068479537963866, 0.018091007232666014, 0.018094079971313477, 0.018070528030395508, 0.01805516815185547, 0.018127872467041017, 0.018123775482177733, 0.018084863662719726, 0.018061311721801757, 0.018157567977905274, 0.018159616470336915, 0.01818009567260742, 0.01823232078552246, 0.018027519226074217, 0.01813811111450195, 0.018119680404663087, 0.01807974433898926, 0.01807257652282715, 0.01803980827331543, 0.018052095413208007, 0.018075647354125975, 0.01807360076904297, 0.018108415603637695, 0.018044927597045898, 0.018100223541259765, 0.018062335968017578, 0.018059263229370116, 0.018107391357421874, 0.017929216384887696, 0.018058240890502928, 0.018093055725097656, 0.018063455581665038, 0.01787910461425781, 0.017816415786743166, 0.01783296012878418, 0.017787967681884766, 0.01780851173400879, 0.06847980499267578, 0.017846336364746095, 0.017819583892822264, 0.017937408447265626, 0.018001920700073244, 0.0178606071472168, 0.017881088256835938, 0.01784217643737793, 0.017906688690185548, 0.017862655639648437, 0.017905759811401366, 0.017723295211791994, 0.017756160736083985, 0.017683456420898438, 0.01784934425354004, 0.017827840805053712, 0.01780838394165039, 0.017789951324462892, 0.017821695327758787, 0.017896448135375977, 0.017829887390136717, 0.017893375396728514, 0.017925119400024413, 0.017863679885864257, 0.017886207580566405, 0.017743871688842772, 0.017869823455810546, 0.01786777687072754, 0.018125823974609375, 0.018114559173583983, 0.018094079971313477, 0.017966079711914062, 0.01798860740661621, 0.018176000595092775, 0.01807360076904297, 0.018068479537963866, 0.018081792831420897, 0.018074623107910158, 0.018139135360717772, 0.01803264045715332, 0.01813811111450195, 0.01817190361022949, 0.018089984893798827, 0.018070528030395508, 0.0180316162109375, 0.01803468894958496, 0.01805721664428711, 0.018130943298339842, 0.01807974433898926, 0.018098175048828127, 0.018096128463745118, 0.01805721664428711, 0.018028543472290038, 0.01799884796142578, 0.01802448081970215, 0.01806230354309082, 0.01804800033569336, 0.018084863662719726, 0.018067455291748045, 0.018095104217529297, 0.018132991790771484, 0.01805619239807129, 0.01804595184326172, 0.06893055725097656, 0.018086912155151368, 0.018109439849853515, 0.018062400817871093, 0.01811244773864746, 0.018067455291748045, 0.018107391357421874, 0.0180633602142334, 0.018149375915527344, 0.01965363121032715, 0.01887846374511719, 0.018144256591796876, 0.018123775482177733, 0.01803980827331543, 0.018140159606933593, 0.018091007232666014, 0.018139135360717772, 0.018156543731689453, 0.01805414390563965, 0.018132991790771484, 0.018099199295043944, 0.01807360076904297, 0.018393119812011718, 0.01804182434082031, 0.018082815170288084, 0.018066432952880858, 0.018116607666015624, 0.01801318359375, 0.018077695846557617, 0.018155519485473632, 0.018036735534667968, 0.018053184509277342, 0.018037696838378907, 0.017902591705322265, 0.01803980827331543, 0.017898496627807618, 0.01797427177429199, 0.017843231201171875, 0.017934303283691406, 0.01778892707824707, 0.017796096801757814, 0.017864704132080078, 0.018084863662719726, 0.018060287475585936, 0.018150400161743165, 0.018035711288452147, 0.01799679946899414, 0.018076671600341796, 0.018068479537963866, 0.01801625633239746, 0.018051071166992186, 0.01803059196472168, 0.017979391098022462, 0.01799884796142578, 0.018145280838012694, 0.01788313674926758, 0.017934335708618163, 0.01786777687072754, 0.018001920700073244, 0.017827840805053712, 0.017932287216186525, 0.017960960388183594, 0.018084863662719726, 0.0684933090209961, 0.0180633602142334, 0.018077695846557617, 0.018027519226074217, 0.018093055725097656, 0.01817190361022949, 0.018116607666015624, 0.018085887908935547, 0.018104320526123048, 0.01807257652282715, 0.01805516815185547, 0.018206720352172853, 0.018151424407958985, 0.017871904373168945, 0.017973215103149413, 0.017872896194458008, 0.017968128204345703, 0.01804697608947754, 0.017892351150512697, 0.017953792572021485, 0.01781452751159668, 0.01788211250305176, 0.018206720352172853, 0.017905664443969727, 0.018542591094970702, 0.018733055114746093, 0.018140159606933593, 0.018044927597045898, 0.01822105598449707, 0.018095104217529297, 0.01805516815185547, 0.018124799728393554, 0.01921331214904785, 0.018689023971557618, 0.01821388816833496, 0.01807974433898926, 0.01796505546569824, 0.017906688690185548, 0.017932287216186525, 0.018082815170288084, 0.018044927597045898, 0.01804083251953125, 0.01802137565612793, 0.018103296279907227, 0.018066432952880858, 0.018043903350830077, 0.018009088516235353, 0.018207744598388673, 0.018070528030395508, 0.01823027229309082, 0.018050048828125, 0.018051071166992186, 0.018050048828125, 0.01805414390563965, 0.018117631912231445, 0.01810534477233887, 0.018100223541259765, 0.018028543472290038, 0.018060287475585936, 0.01803775978088379, 0.018159616470336915, 0.018053119659423827, 0.018083839416503905, 0.06882508850097656, 0.01818726348876953, 0.01808896064758301, 0.01810534477233887, 0.018069503784179687, 0.018166784286499024, 0.018832384109497072, 0.01859993553161621, 0.018597888946533202, 0.018544639587402344, 0.018471935272216796, 0.018231296539306642, 0.01838591957092285, 0.01805414390563965, 0.01806540870666504, 0.018061311721801757, 0.018069503784179687, 0.018025472640991212, 0.01804607963562012, 0.01806732749938965, 0.018081792831420897, 0.018008064270019532, 0.01819647979736328, 0.018127872467041017, 0.01803980827331543, 0.018136064529418947, 0.01803878402709961, 0.018035711288452147, 0.018133056640625, 0.018062271118164063, 0.017993728637695314, 0.01803059196472168, 0.018121728897094725, 0.01807257652282715, 0.01804697608947754, 0.018069503784179687, 0.018126848220825196, 0.01802956771850586, 0.01801420783996582, 0.018051071166992186, 0.018060287475585936, 0.018081792831420897, 0.018119680404663087, 0.01801625633239746, 0.018084863662719726, 0.017904640197753906, 0.017957887649536132, 0.017886207580566405, 0.01803059196472168, 0.018050048828125, 0.01805731201171875, 0.017820575714111327, 0.017976320266723633, 0.017796096801757814, 0.0178288631439209, 0.01787494468688965, 0.017947647094726564, 0.017859647750854492, 0.017836992263793944, 0.01785139274597168, 0.017911808013916015, 0.017961984634399415, 0.01783193588256836, 0.0682425308227539, 0.017862560272216797, 0.017928192138671875, 0.01787392044067383, 0.017937408447265626, 0.01786684799194336, 0.017915807723999023, 0.01788313674926758, 0.01785139274597168, 0.017839103698730468, 0.017932287216186525, 0.017868864059448243, 0.017847328186035155, 0.017670047760009765, 0.01782067108154297, 0.017663999557495116, 0.017683456420898438, 0.01779097557067871, 0.017905664443969727, 0.01786777687072754, 0.018234560012817383, 0.017982271194458006, 0.0180316162109375, 0.01802956771850586, 0.01803878402709961, 0.018051071166992186, 0.018207744598388673, 0.018096128463745118, 0.018066432952880858, 0.018156543731689453, 0.018084863662719726, 0.018106399536132814, 0.01809814453125, 0.018070528030395508, 0.01820364761352539, 0.01842585563659668, 0.018111488342285157, 0.018124799728393554, 0.018110464096069336, 0.018114559173583983, 0.018081792831420897, 0.018051071166992186, 0.01806550407409668, 0.01803664016723633, 0.018094079971313477, 0.018122751235961913, 0.01808076858520508, 0.018070528030395508, 0.01802649688720703, 0.01805721664428711, 0.018126848220825196, 0.018035711288452147, 0.01803468894958496, 0.018144256591796876, 0.018362367630004883, 0.01802239990234375, 0.018076671600341796, 0.01804083251953125, 0.01805619239807129, 0.018027519226074217, 0.01806438446044922, 0.018020383834838866, 0.01806844711303711]",tokens/s,53.052696949627695,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,togethercomputer/RedPajama-INCITE-Base-3B-v1,togethercomputer/RedPajama-INCITE-Base-3B-v1,cuda,0,42,,,True,True,True,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.42.3,,0.31.0,,,,1.20.0,,,,0.11.1,,,,MB,1116.07808,12541.493248,0.0,11901.3376,11746.415104,s,10,0.9637533798217773,0.09637533798217773,0.0005159312604963635,0.09631588745117187,0.09708050689697266,0.0972128303527832,0.09731868911743165,"[0.09540128326416016, 0.09643523406982422, 0.09611929321289063, 0.09600508880615234, 0.09617833709716797, 0.09621756744384766, 0.09641420745849609, 0.09658611297607422, 0.09705110168457032, 0.09734515380859375]",tokens/s,2656.2812163350436,kWh,1.1690498502166184e-06,6.405878167283199e-07,7.919181171941307e-06,9.728818838886244e-06,tokens/kWh,26313574.570508387,MB,1116.40576,12541.493248,0.0,11901.3376,11861.465088,s,10,17.246585205078123,1.7246585205078122,0.0007374892436067599,1.7244789428710936,1.725922216796875,1.726025927734375,1.726108896484375,"[1.726129638671875, 1.7248255615234376, 1.72419873046875, 1.72460986328125, 1.72410498046875, 1.72378271484375, 1.725899169921875, 1.72407666015625, 1.7243858642578125, 1.724572021484375]",tokens/s,36.52897037348016,kWh,2.0342288304258277e-05,1.114723796447517e-05,8.273219690465921e-05,0.00011422172317339262,tokens/kWh,551558.8300516511,,s,629,18.086841373443615,0.028754914743153583,0.011550466294339848,0.027404287338256835,0.027480064392089845,0.02751119384765625,0.1242020248413086,"[0.027450368881225585, 0.02730700874328613, 0.027236352920532225, 0.027248640060424805, 0.027223039627075195, 0.027235328674316408, 0.027214847564697265, 0.027237375259399413, 0.027228160858154295, 0.027253759384155272, 0.027243520736694334, 0.027243520736694334, 0.027245567321777343, 0.027222015380859374, 0.027255807876586914, 0.027244543075561522, 0.027453439712524414, 0.027456512451171877, 0.02733568000793457, 0.02738380813598633, 0.02725273513793945, 0.027249664306640626, 0.027239423751831054, 0.027249727249145508, 0.02724448013305664, 0.02727628707885742, 0.027274240493774415, 0.027295743942260742, 0.027266048431396486, 0.027281408309936524, 0.02730188751220703, 0.027425792694091795, 0.027676671981811524, 0.02746675109863281, 0.027412479400634765, 0.027425823211669923, 0.02739401626586914, 0.02741974449157715, 0.027418527603149414, 0.027460607528686523, 0.027422719955444336, 0.027427839279174804, 0.027428863525390625, 0.027429887771606445, 0.027440128326416017, 0.027448320388793947, 0.027459583282470702, 0.027486207962036133, 0.02746675109863281, 0.027481088638305663, 0.027470848083496095, 0.027476991653442383, 0.027860992431640624, 0.027535360336303712, 0.027513856887817382, 0.027461631774902344, 0.027488256454467775, 0.027615232467651366, 0.02754560089111328, 0.02750873565673828, 0.027495424270629884, 0.02751283264160156, 0.12462387084960938, 0.0275230712890625, 0.02733875274658203, 0.027226112365722657, 0.027247615814208984, 0.027226112365722657, 0.027237375259399413, 0.027240447998046875, 0.027263999938964844, 0.027232255935668945, 0.027234304428100587, 0.027229183197021483, 0.027254783630371093, 0.027243520736694334, 0.027225088119506836, 0.027222015380859374, 0.027212799072265623, 0.027257856369018556, 0.027251712799072264, 0.027251712799072264, 0.027218944549560548, 0.027224063873291016, 0.027380735397338866, 0.027257856369018556, 0.02733670425415039, 0.027281408309936524, 0.02731724739074707, 0.02725273513793945, 0.02739200019836426, 0.027312128067016602, 0.027270143508911132, 0.027279359817504883, 0.027389951705932617, 0.027418624877929686, 0.027406335830688477, 0.027413503646850586, 0.027433984756469725, 0.027411455154418944, 0.027425792694091795, 0.027407360076904298, 0.027404287338256835, 0.027417600631713866, 0.027437055587768554, 0.027458560943603515, 0.027439104080200196, 0.027449344635009764, 0.027459583282470702, 0.027458560943603515, 0.027459583282470702, 0.027476991653442383, 0.02750464057922363, 0.027463680267333986, 0.027494400024414063, 0.027459583282470702, 0.027472896575927733, 0.027457536697387694, 0.027453439712524414, 0.027475967407226562, 0.027464704513549806, 0.027486207962036133, 0.02776371192932129, 0.02750464057922363, 0.027462656021118165, 0.12419686126708984, 0.027471872329711915, 0.02731007957458496, 0.027242496490478517, 0.027242496490478517, 0.027220991134643553, 0.027225088119506836, 0.027258880615234377, 0.02726092720031738, 0.027249664306640626, 0.027275264739990233, 0.027278335571289062, 0.027253759384155272, 0.027237375259399413, 0.027254783630371093, 0.027251712799072264, 0.027216896057128907, 0.027237375259399413, 0.027216896057128907, 0.027261951446533202, 0.02726092720031738, 0.027231231689453125, 0.027240447998046875, 0.027241472244262696, 0.02731110382080078, 0.027270143508911132, 0.027263999938964844, 0.027236352920532225, 0.027262975692749023, 0.027259904861450194, 0.02730700874328613, 0.027281408309936524, 0.027381759643554687, 0.02740940856933594, 0.027398143768310547, 0.027410432815551757, 0.02740121650695801, 0.027423744201660157, 0.027442176818847655, 0.02740121650695801, 0.02740121650695801, 0.027428863525390625, 0.027437055587768554, 0.027459583282470702, 0.027435007095336913, 0.027438079833984375, 0.027444223403930663, 0.027588607788085938, 0.027463680267333986, 0.027456512451171877, 0.027464704513549806, 0.02747494316101074, 0.027488256454467775, 0.02747494316101074, 0.027467775344848632, 0.027459583282470702, 0.027446271896362305, 0.027478015899658204, 0.027469823837280274, 0.027497472763061522, 0.027472896575927733, 0.027468799591064453, 0.027464704513549806, 0.1241190414428711, 0.02751283264160156, 0.02740940856933594, 0.027231231689453125, 0.027248640060424805, 0.027232255935668945, 0.027225088119506836, 0.027248640060424805, 0.027233280181884766, 0.027253759384155272, 0.027244543075561522, 0.027249664306640626, 0.027250688552856447, 0.027227136611938478, 0.027216896057128907, 0.027207679748535156, 0.027224063873291016, 0.027226112365722657, 0.027218944549560548, 0.027230207443237304, 0.027228160858154295, 0.027242496490478517, 0.027223072052001952, 0.027247583389282227, 0.027245567321777343, 0.02727731132507324, 0.027258880615234377, 0.02750464057922363, 0.02730291175842285, 0.027271232604980468, 0.02729363250732422, 0.027289600372314454, 0.027417600631713866, 0.02739302444458008, 0.027404287338256835, 0.027411455154418944, 0.027414527893066407, 0.027439104080200196, 0.027461631774902344, 0.027778047561645508, 0.027418624877929686, 0.027416576385498048, 0.02754150390625, 0.027427839279174804, 0.027418624877929686, 0.027471872329711915, 0.027446271896362305, 0.027465728759765624, 0.027463680267333986, 0.027463680267333986, 0.027464704513549806, 0.027460607528686523, 0.027490304946899413, 0.027496448516845705, 0.027468799591064453, 0.027462656021118165, 0.027464704513549806, 0.027456512451171877, 0.027451391220092772, 0.027460607528686523, 0.027476991653442383, 0.027459583282470702, 0.027463680267333986, 0.12420403289794922, 0.027428863525390625, 0.02730188751220703, 0.027290624618530275, 0.027246591567993163, 0.027253759384155272, 0.027251712799072264, 0.027234304428100587, 0.027237375259399413, 0.027217920303344727, 0.027278335571289062, 0.027246591567993163, 0.027247615814208984, 0.027215871810913086, 0.027218944549560548, 0.0273305606842041, 0.027312128067016602, 0.027240447998046875, 0.027244543075561522, 0.027231231689453125, 0.027230207443237304, 0.02728550338745117, 0.027250688552856447, 0.027233280181884766, 0.027246591567993163, 0.027239423751831054, 0.027279359817504883, 0.027259904861450194, 0.027267072677612306, 0.027268096923828124, 0.027251712799072264, 0.027251712799072264, 0.027367488861083984, 0.027397056579589844, 0.027412576675415037, 0.02754243278503418, 0.027396095275878905, 0.027427839279174804, 0.027433984756469725, 0.027470848083496095, 0.027418624877929686, 0.02740940856933594, 0.027436031341552734, 0.027422719955444336, 0.027440128326416017, 0.027458560943603515, 0.02751692771911621, 0.027472896575927733, 0.027456512451171877, 0.027446271896362305, 0.027472896575927733, 0.027471872329711915, 0.027480064392089845, 0.02749849510192871, 0.027463680267333986, 0.027465728759765624, 0.027450368881225585, 0.027464704513549806, 0.027443199157714843, 0.027450368881225585, 0.027461631774902344, 0.027490304946899413, 0.02747494316101074, 0.12426649475097656, 0.027451391220092772, 0.027283456802368163, 0.027217920303344727, 0.027218944549560548, 0.027250688552856447, 0.027244543075561522, 0.027229183197021483, 0.027220991134643553, 0.027230207443237304, 0.027241472244262696, 0.027231231689453125, 0.027246591567993163, 0.027218944549560548, 0.027259904861450194, 0.027219968795776366, 0.027253759384155272, 0.027242496490478517, 0.027233280181884766, 0.027223039627075195, 0.027219968795776366, 0.027250688552856447, 0.027241472244262696, 0.027263999938964844, 0.027237375259399413, 0.027242496490478517, 0.027279359817504883, 0.027267072677612306, 0.027262975692749023, 0.027267072677612306, 0.027255807876586914, 0.027251712799072264, 0.027410432815551757, 0.027395072937011718, 0.02740121650695801, 0.027404287338256835, 0.02740838432312012, 0.02740940856933594, 0.027428863525390625, 0.027403263092041014, 0.027423744201660157, 0.027415552139282227, 0.027416576385498048, 0.02740838432312012, 0.027444223403930663, 0.027464704513549806, 0.027446271896362305, 0.027448320388793947, 0.027470848083496095, 0.027472896575927733, 0.027468799591064453, 0.027452415466308593, 0.027502592086791993, 0.027486207962036133, 0.027490304946899413, 0.027475967407226562, 0.027459583282470702, 0.027470848083496095, 0.027462656021118165, 0.027460607528686523, 0.027505664825439452, 0.027475967407226562, 0.027625471115112304, 0.1250324478149414, 0.02794803237915039, 0.027428863525390625, 0.027257856369018556, 0.027233280181884766, 0.027232255935668945, 0.027242496490478517, 0.02725273513793945, 0.027226112365722657, 0.027245567321777343, 0.027230207443237304, 0.027233280181884766, 0.027258880615234377, 0.027243520736694334, 0.027219968795776366, 0.027225088119506836, 0.027212799072265623, 0.027228160858154295, 0.027230207443237304, 0.027422719955444336, 0.027309087753295897, 0.027276256561279296, 0.027241472244262696, 0.027406335830688477, 0.027432960510253908, 0.027389951705932617, 0.027280384063720704, 0.027251712799072264, 0.027259904861450194, 0.027273216247558595, 0.027274240493774415, 0.027267072677612306, 0.027381759643554687, 0.027402240753173827, 0.02739302444458008, 0.027429887771606445, 0.02740838432312012, 0.027424768447875978, 0.027450368881225585, 0.027407360076904298, 0.027420671463012695, 0.027417600631713866, 0.027467775344848632, 0.027445247650146484, 0.027441152572631834, 0.027460607528686523, 0.027450368881225585, 0.027456512451171877, 0.027476991653442383, 0.027471872329711915, 0.027563007354736328, 0.02749849510192871, 0.027476991653442383, 0.027453439712524414, 0.027468799591064453, 0.027472896575927733, 0.027462656021118165, 0.02754457664489746, 0.027478015899658204, 0.027468799591064453, 0.027469823837280274, 0.027458560943603515, 0.027456512451171877, 0.12457574462890625, 0.027446271896362305, 0.027287551879882813, 0.027225088119506836, 0.027217920303344727, 0.027226112365722657, 0.027217920303344727, 0.027224063873291016, 0.027236352920532225, 0.027255807876586914, 0.027243520736694334, 0.027378688812255858, 0.027241472244262696, 0.027212799072265623, 0.027228160858154295, 0.027278335571289062, 0.027219968795776366, 0.027227136611938478, 0.02733363151550293, 0.027238399505615234, 0.027456512451171877, 0.027254783630371093, 0.027244543075561522, 0.027232255935668945, 0.027258880615234377, 0.027291648864746092, 0.027246591567993163, 0.027279359817504883, 0.027254783630371093, 0.02728447914123535, 0.027251712799072264, 0.027253759384155272, 0.027378688812255858, 0.027412479400634765, 0.027417600631713866, 0.02740019226074219, 0.027413503646850586, 0.027427839279174804, 0.02740838432312012, 0.02740838432312012, 0.027416576385498048, 0.027419647216796874, 0.027438079833984375, 0.027425792694091795, 0.027449344635009764, 0.027457536697387694, 0.027449440002441407, 0.027466655731201172, 0.027459583282470702, 0.027464704513549806, 0.027463680267333986, 0.027475967407226562, 0.027478015899658204, 0.027476991653442383, 0.027462656021118165, 0.027472896575927733, 0.027458560943603515, 0.02751283264160156, 0.027465728759765624, 0.027479040145874024, 0.027471872329711915, 0.027480064392089845, 0.027460607528686523, 0.12509081268310546, 0.027484159469604492, 0.027325439453125, 0.027287551879882813, 0.027243520736694334, 0.027251712799072264, 0.027241472244262696, 0.027236352920532225, 0.027281408309936524, 0.027241472244262696, 0.027272192001342774, 0.027278335571289062, 0.027244543075561522, 0.027205631256103514, 0.027218944549560548, 0.027215871810913086, 0.027249664306640626, 0.027226112365722657, 0.027226112365722657, 0.027246591567993163, 0.027227136611938478, 0.027256832122802735, 0.027245567321777343, 0.027249664306640626, 0.027255807876586914, 0.027241472244262696, 0.027262975692749023, 0.027230207443237304, 0.027245567321777343, 0.02726092720031738, 0.02726911926269531, 0.027253759384155272, 0.02747494316101074, 0.027418624877929686, 0.027440128326416017, 0.02740838432312012, 0.027407360076904298, 0.027424768447875978, 0.02740940856933594, 0.027441152572631834, 0.027421695709228516, 0.027403263092041014, 0.027424768447875978, 0.027440128326416017, 0.027430912017822266, 0.027442176818847655, 0.02747494316101074, 0.02747494316101074, 0.027464704513549806, 0.027479040145874024, 0.027461631774902344, 0.027457536697387694, 0.02747494316101074, 0.027505664825439452, 0.02747494316101074, 0.027478015899658204, 0.027453439712524414, 0.02746675109863281, 0.027448320388793947, 0.027469823837280274, 0.027462656021118165, 0.02771865653991699, 0.027618303298950195, 0.1254471664428711, 0.02750054359436035, 0.027325439453125, 0.027242496490478517, 0.027273216247558595, 0.027268096923828124, 0.027241472244262696, 0.027229183197021483, 0.02730188751220703, 0.027244543075561522, 0.027236352920532225, 0.027248640060424805, 0.027261951446533202, 0.027265024185180665, 0.027219968795776366, 0.027239423751831054, 0.027262975692749023, 0.027250688552856447, 0.027243520736694334, 0.027261951446533202, 0.027232255935668945, 0.027233280181884766, 0.027242496490478517, 0.027258880615234377, 0.027259904861450194, 0.027275264739990233, 0.027245567321777343, 0.027268096923828124, 0.027245567321777343, 0.027266048431396486, 0.027287551879882813, 0.027297792434692384, 0.027386880874633788, 0.027387903213500975, 0.027433984756469725, 0.027463680267333986, 0.027402240753173827, 0.027413503646850586, 0.027486207962036133, 0.027429887771606445, 0.027433984756469725, 0.02756096076965332, 0.027442176818847655, 0.027453439712524414, 0.027429887771606445, 0.027458560943603515, 0.027438079833984375, 0.027471872329711915, 0.027463680267333986, 0.027493375778198242, 0.027494400024414063, 0.027479040145874024, 0.02749951934814453, 0.027480064392089845, 0.027470848083496095, 0.027472896575927733, 0.027475967407226562, 0.027476991653442383, 0.02746675109863281, 0.027488256454467775, 0.027502592086791993, 0.027471872329711915, 0.027473920822143554]",tokens/s,34.77666370887417,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,meta-llama/Meta-Llama-3-8B,meta-llama/Meta-Llama-3-8B,cuda,0,42,,,True,True,True,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ self.load_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 250, in load_model_with_no_weights self.pretrained_model = self.automodel_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3710, in from_pretrained model = cls(config, *model_args, **model_kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 1101, in __init__ self.model = LlamaModel(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 884, in __init__ [LlamaDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 884, in [LlamaDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 679, in __init__ self.mlp = LlamaMLP(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 193, in __init__ self.up_proj = nn.Linear(self.hidden_size, self.intermediate_size, bias=config.mlp_bias) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/linear.py"", line 98, in __init__ self.weight = Parameter(torch.empty((out_features, in_features), **factory_kwargs)) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_device.py"", line 78, in __torch_function__ return func(*args, **kwargs) torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 224.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen2-beta-14B,Qwen/Qwen2-beta-14B,cuda,0,42,,,True,,,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",qwen2,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-32B,Qwen/Qwen1.5-32B,cuda,0,42,,,True,,,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",qwen2,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,0,0,cuda,0,42,,,True,True,True,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/0/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6693259b-4960ce01513c379e778c36dc;6d0a06ba-4962-480c-bda2-464657794947) Repository Not Found for url: https://huggingface.co/0/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: 0 is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,2,2,cuda,0,42,,,True,True,True,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/2/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669324ce-06d5609a427f5e8f20f5dee5;45ce0664-56e0-4bc8-aa24-2e54f5768e7d) Repository Not Found for url: https://huggingface.co/2/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: 2 is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/xglm-564M,facebook/xglm-564M,cuda,0,42,,,True,,,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",xglm,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-14B,Qwen/Qwen-14B,cuda,0,42,,,True,,,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",qwen,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,huggyllama/llama-7b,huggyllama/llama-7b,cuda,0,42,,,True,,,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",llama,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-180B,tiiuae/falcon-180B,cuda,0,42,,,True,True,True,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 403 Client Error: Forbidden for url: https://huggingface.co/tiiuae/falcon-180B/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-66932706-7d36dbcb1c9622976d161a54;ba8bf22e-4627-4ed0-935b-9f152280218c) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/tiiuae/falcon-180B/resolve/main/config.json. If you are trying to create or update content,make sure you have a token with the `write` role. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1826, in _raise_on_head_call_error raise LocalEntryNotFoundError( huggingface_hub.utils._errors.LocalEntryNotFoundError: An error happened while trying to locate the file on the Hub and we cannot find the requested files in the local cache. Please check your connection and try again or make sure your Internet connection is on. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 445, in cached_file raise EnvironmentError( OSError: We couldn't connect to 'https://huggingface.co' to load this file, couldn't find it in the cached files and it looks like tiiuae/falcon-180B is not the path to a directory containing a file named config.json. Checkout your internet connection or see how to run the library in offline mode at 'https://huggingface.co/docs/transformers/installation#offline-mode'. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,mistralai/Mistral-7B-v0.1,mistralai/Mistral-7B-v0.1,cuda,0,42,,,True,True,True,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ self.load_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 250, in load_model_with_no_weights self.pretrained_model = self.automodel_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3710, in from_pretrained model = cls(config, *model_args, **model_kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/mistral/modeling_mistral.py"", line 1126, in __init__ self.model = MistralModel(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/mistral/modeling_mistral.py"", line 879, in __init__ [MistralDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/mistral/modeling_mistral.py"", line 879, in [MistralDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/mistral/modeling_mistral.py"", line 679, in __init__ self.mlp = MistralMLP(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/mistral/modeling_mistral.py"", line 165, in __init__ self.gate_proj = nn.Linear(self.hidden_size, self.intermediate_size, bias=False) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/linear.py"", line 98, in __init__ self.weight = Parameter(torch.empty((out_features, in_features), **factory_kwargs)) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_device.py"", line 78, in __torch_function__ return func(*args, **kwargs) torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 224.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,google/recurrentgemma-7b,google/recurrentgemma-7b,cuda,0,42,,,True,True,True,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/google/recurrentgemma-7b/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6694703b-12263a711dc1333c03b211f2;90ddfe47-8e92-4ecc-8500-1ad11442ab70) Repository Not Found for url: https://huggingface.co/google/recurrentgemma-7b/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: google/recurrentgemma-7b is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,mistralai/Mixtral-8x7B-v0.1,mistralai/Mixtral-8x7B-v0.1,cuda,0,42,,,True,True,True,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ self.load_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 250, in load_model_with_no_weights self.pretrained_model = self.automodel_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3710, in from_pretrained model = cls(config, *model_args, **model_kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/mixtral/modeling_mixtral.py"", line 1323, in __init__ self.model = MixtralModel(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/mixtral/modeling_mixtral.py"", line 1087, in __init__ [MixtralDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/mixtral/modeling_mixtral.py"", line 1087, in [MixtralDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/mixtral/modeling_mixtral.py"", line 877, in __init__ self.block_sparse_moe = MixtralSparseMoeBlock(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/mixtral/modeling_mixtral.py"", line 824, in __init__ self.experts = nn.ModuleList([MixtralBlockSparseTop2MLP(config) for _ in range(self.num_experts)]) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/mixtral/modeling_mixtral.py"", line 824, in self.experts = nn.ModuleList([MixtralBlockSparseTop2MLP(config) for _ in range(self.num_experts)]) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/mixtral/modeling_mixtral.py"", line 791, in __init__ self.w2 = nn.Linear(self.ffn_dim, self.hidden_dim, bias=False) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/linear.py"", line 98, in __init__ self.weight = Parameter(torch.empty((out_features, in_features), **factory_kwargs)) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_device.py"", line 78, in __torch_function__ return func(*args, **kwargs) torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 224.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,t,t,cuda,0,42,,,True,True,True,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/t/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6694739f-7626b30c1b22fa1d2e4e8109;f9a19fca-4468-40ac-91d9-363e5d71246c) Repository Not Found for url: https://huggingface.co/t/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: t is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-72B,Qwen/Qwen-72B,cuda,0,42,,,True,,,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",qwen,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-30b,facebook/opt-30b,cuda,0,42,,,True,,,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",opt,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-MoE-A2.7B,Qwen/Qwen1.5-MoE-A2.7B,cuda,0,42,,,True,,,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",qwen2_moe,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,x,x,cuda,0,42,,,True,True,True,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/x/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66932467-07e936f40825445c571e09f2;7cfe6e2d-7bb8-4669-b2eb-00ace7ca3078) Repository Not Found for url: https://huggingface.co/x/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: x is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,01-ai/Yi-34B,01-ai/Yi-34B,cuda,0,42,,,True,,,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",llama,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/xglm-4.5B,facebook/xglm-4.5B,cuda,0,42,,,True,,,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",xglm,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/xglm-7.5B,facebook/xglm-7.5B,cuda,0,42,,,True,,,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",xglm,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-14B,Qwen/Qwen1.5-14B,cuda,0,42,,,True,,,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",qwen2,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,meta-llama/Llama-2-70b-hf,meta-llama/Llama-2-70b-hf,cuda,0,42,,,True,True,True,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ self.load_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 250, in load_model_with_no_weights self.pretrained_model = self.automodel_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3710, in from_pretrained model = cls(config, *model_args, **model_kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 1101, in __init__ self.model = LlamaModel(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 884, in __init__ [LlamaDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 884, in [LlamaDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 679, in __init__ self.mlp = LlamaMLP(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 193, in __init__ self.up_proj = nn.Linear(self.hidden_size, self.intermediate_size, bias=config.mlp_bias) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/linear.py"", line 98, in __init__ self.weight = Parameter(torch.empty((out_features, in_features), **factory_kwargs)) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_device.py"", line 78, in __torch_function__ return func(*args, **kwargs) torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 896.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Deci/DeciCoder-1b,Deci/DeciCoder-1b,cuda,0,42,,,True,,,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",llama,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,meta-llama/Meta-Llama-3-70B,meta-llama/Meta-Llama-3-70B,cuda,0,42,,,True,True,True,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ self.load_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 250, in load_model_with_no_weights self.pretrained_model = self.automodel_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3710, in from_pretrained model = cls(config, *model_args, **model_kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 1101, in __init__ self.model = LlamaModel(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 884, in __init__ [LlamaDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 884, in [LlamaDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 679, in __init__ self.mlp = LlamaMLP(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 193, in __init__ self.up_proj = nn.Linear(self.hidden_size, self.intermediate_size, bias=config.mlp_bias) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/linear.py"", line 98, in __init__ self.weight = Parameter(torch.empty((out_features, in_features), **factory_kwargs)) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_device.py"", line 78, in __torch_function__ return func(*args, **kwargs) torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 896.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,r,r,cuda,0,42,,,True,True,True,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/r/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669473d9-6dd19c917c34caca28c99768;a6028978-f4ab-42fe-ac38-5e0457847fff) Repository Not Found for url: https://huggingface.co/r/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: r is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,togethercomputer/RedPajama-INCITE-Base-7B-v0.1,togethercomputer/RedPajama-INCITE-Base-7B-v0.1,cuda,0,42,,,True,True,True,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ self.load_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 250, in load_model_with_no_weights self.pretrained_model = self.automodel_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3710, in from_pretrained model = cls(config, *model_args, **model_kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 1028, in __init__ self.gpt_neox = GPTNeoXModel(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 853, in __init__ self.layers = nn.ModuleList([GPTNeoXLayer(config) for _ in range(config.num_hidden_layers)]) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 853, in self.layers = nn.ModuleList([GPTNeoXLayer(config) for _ in range(config.num_hidden_layers)]) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 741, in __init__ self.mlp = GPTNeoXMLP(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 715, in __init__ self.dense_4h_to_h = nn.Linear(config.intermediate_size, config.hidden_size) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/linear.py"", line 98, in __init__ self.weight = Parameter(torch.empty((out_features, in_features), **factory_kwargs)) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_device.py"", line 78, in __torch_function__ return func(*args, **kwargs) torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 256.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-72B,Qwen/Qwen1.5-72B,cuda,0,42,,,True,,,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",qwen2,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,internlm/internlm-20b,internlm/internlm-20b,cuda,0,42,,,True,,,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",internlm,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-110B,Qwen/Qwen1.5-110B,cuda,0,42,,,True,,,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",qwen2,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Salesforce/codegen-16B-nl,Salesforce/codegen-16B-nl,cuda,0,42,,,True,,,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",codegen,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,m,m,cuda,0,42,,,True,True,True,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/m/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669472da-18aa21bc6b2b69bc663ff33e;4d074193-0d32-4dd1-998d-2c6235f4b10d) Repository Not Found for url: https://huggingface.co/m/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: m is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,openai-community/gpt2-large,openai-community/gpt2-large,cuda,0,42,,,True,True,True,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.42.3,,0.31.0,,,,1.20.0,,,,0.11.1,,,,MB,1056.44032,4190.633984,0.0,3550.478336,3412.88704,s,10,0.32883602905273435,0.03288360290527344,0.00016511338811359813,0.032827455520629885,0.032955459594726566,0.0331556339263916,0.03331577339172363,"[0.03335580825805664, 0.03274934387207031, 0.03283055877685547, 0.03280831909179688, 0.03276803207397461, 0.03291097640991211, 0.03281785583496094, 0.0328243522644043, 0.03290982437133789, 0.032860958099365234]",tokens/s,7785.035013877573,kWh,3.934829833103573e-07,2.1560940236699826e-07,2.677206829840655e-06,3.2862992155180106e-06,tokens/kWh,77899175.70230968,MB,1056.44032,4190.633984,0.0,3550.478336,3457.678336,s,10,10.957459960937499,1.09574599609375,0.007644036751543853,1.0941619873046875,1.0990013916015624,1.1082768188476562,1.1156971606445312,"[1.11755224609375, 1.096940185546875, 1.094740478515625, 1.094123291015625, 1.09420068359375, 1.092636962890625, 1.0892796630859376, 1.0907955322265626, 1.0907576904296874, 1.0964332275390625]",tokens/s,57.49507661865994,kWh,1.267496970938378e-05,6.945489095361142e-06,3.816407584315526e-05,5.778453464790017e-05,tokens/kWh,1090257.1143625078,,s,629,11.234593795776364,0.01786103942094812,0.0038813500039302264,0.017342464447021484,0.017566105651855467,0.01811233367919922,0.04977354644775391,"[0.018313215255737304, 0.01824563217163086, 0.01811052894592285, 0.018373567581176756, 0.01820876884460449, 0.018190336227416993, 0.018493440628051756, 0.018106367111206053, 0.018035711288452147, 0.018156543731689453, 0.01829478454589844, 0.018332672119140626, 0.018149375915527344, 0.018375680923461913, 0.018339839935302735, 0.018165760040283203, 0.018061311721801757, 0.018113536834716795, 0.018157567977905274, 0.018107391357421874, 0.01824051284790039, 0.01796505546569824, 0.017992704391479493, 0.017696767807006835, 0.01801113510131836, 0.01820159912109375, 0.01796505546569824, 0.017967103958129883, 0.017827840805053712, 0.01740595245361328, 0.01747148895263672, 0.017286144256591796, 0.017366016387939453, 0.018159616470336915, 0.017942527770996093, 0.017292287826538084, 0.017343551635742187, 0.01733216094970703, 0.01738751983642578, 0.017359872817993165, 0.017368064880371094, 0.017361919403076173, 0.017362943649291994, 0.017385471343994142, 0.01720627212524414, 0.017297407150268555, 0.017361919403076173, 0.017319936752319336, 0.017352703094482422, 0.017323007583618166, 0.017270784378051757, 0.017228799819946287, 0.017352703094482422, 0.017467391967773437, 0.01740492820739746, 0.017308671951293944, 0.017331199645996095, 0.017359872817993165, 0.01744588851928711, 0.017532928466796875, 0.016861183166503906, 0.016488447189331054, 0.04970502471923828, 0.017845184326171874, 0.017747968673706056, 0.017338367462158204, 0.01742336082458496, 0.01745715141296387, 0.017278976440429687, 0.017390592575073242, 0.0176363525390625, 0.018264064788818358, 0.018183168411254884, 0.01760051155090332, 0.01740595245361328, 0.017315839767456053, 0.01724723243713379, 0.017452032089233398, 0.017694719314575197, 0.017291263580322267, 0.0172728328704834, 0.017305599212646485, 0.01725132751464844, 0.017434623718261717, 0.017558528900146485, 0.017491968154907226, 0.01725951957702637, 0.017288192749023438, 0.01741107177734375, 0.017335296630859375, 0.017366016387939453, 0.01741209602355957, 0.017492992401123047, 0.017501184463500977, 0.017325056076049804, 0.017459199905395507, 0.017270784378051757, 0.017261568069458007, 0.017269760131835937, 0.017300479888916014, 0.017356800079345702, 0.017475584030151366, 0.01744691276550293, 0.017361919403076173, 0.01725951957702637, 0.017308671951293944, 0.017275903701782228, 0.017332223892211913, 0.01722163200378418, 0.017325056076049804, 0.017306623458862306, 0.018251775741577148, 0.017694719314575197, 0.017269760131835937, 0.017310720443725586, 0.017201152801513672, 0.01717350387573242, 0.017357824325561523, 0.017306623458862306, 0.017258495330810548, 0.017302528381347656, 0.017300479888916014, 0.0174653434753418, 0.017333248138427734, 0.0172728328704834, 0.04997017669677734, 0.017357824325561523, 0.017276927947998046, 0.01739263916015625, 0.0172677116394043, 0.017549312591552735, 0.017373184204101562, 0.01740083122253418, 0.017361919403076173, 0.017271808624267578, 0.017342464447021484, 0.01723289680480957, 0.017356800079345702, 0.017327104568481445, 0.01723391914367676, 0.017282047271728516, 0.017360895156860352, 0.0172677116394043, 0.01728102493286133, 0.017350656509399414, 0.01724415969848633, 0.017285120010375975, 0.0172677116394043, 0.017298431396484376, 0.01722777557373047, 0.01782476806640625, 0.017460224151611328, 0.017275903701782228, 0.01741823959350586, 0.01722163200378418, 0.017364992141723632, 0.017345535278320313, 0.017335296630859375, 0.017254400253295898, 0.017356800079345702, 0.017318912506103516, 0.017245183944702147, 0.017378303527832033, 0.01724006462097168, 0.017287168502807617, 0.017435647964477538, 0.017257471084594727, 0.01721036720275879, 0.01740083122253418, 0.017332223892211913, 0.017435647964477538, 0.017507328033447265, 0.018059263229370116, 0.01843814468383789, 0.017721343994140625, 0.017499135971069335, 0.017383424758911133, 0.017315839767456053, 0.017314815521240236, 0.01741926383972168, 0.017336320877075196, 0.01739263916015625, 0.01741209602355957, 0.017368064880371094, 0.017339391708374022, 0.017332223892211913, 0.01739571189880371, 0.01738444709777832, 0.05029478454589844, 0.017252351760864256, 0.017362943649291994, 0.017518592834472657, 0.017327104568481445, 0.017427455902099608, 0.017338367462158204, 0.017317888259887695, 0.01754521560668945, 0.017311744689941407, 0.017309696197509765, 0.017275903701782228, 0.01734758377075195, 0.017308671951293944, 0.017378303527832033, 0.017367040634155274, 0.017383424758911133, 0.017334272384643554, 0.017535999298095704, 0.017304576873779298, 0.017480703353881837, 0.017304576873779298, 0.017382400512695313, 0.0172728328704834, 0.017297407150268555, 0.01745510482788086, 0.017373184204101562, 0.017398784637451172, 0.0172728328704834, 0.01741926383972168, 0.01739468765258789, 0.017289215087890625, 0.01740492820739746, 0.017287168502807617, 0.017505279541015627, 0.01738137626647949, 0.017329151153564454, 0.017253376007080077, 0.017481727600097655, 0.017341440200805663, 0.017252351760864256, 0.017459199905395507, 0.017304576873779298, 0.017277952194213866, 0.017523712158203125, 0.017356800079345702, 0.017299455642700197, 0.017369087219238282, 0.017503231048583985, 0.017562623977661132, 0.01745408058166504, 0.017284095764160155, 0.01740492820739746, 0.017304576873779298, 0.017238016128540038, 0.01740185546875, 0.017368064880371094, 0.017299455642700197, 0.017329151153564454, 0.01721855926513672, 0.01736396789550781, 0.017435680389404296, 0.017368032455444337, 0.050078720092773435, 0.017507328033447265, 0.017292287826538084, 0.017295360565185547, 0.017293312072753905, 0.017292287826538084, 0.01722470474243164, 0.017460224151611328, 0.017360895156860352, 0.017319936752319336, 0.017427455902099608, 0.017360895156860352, 0.0172728328704834, 0.017535999298095704, 0.01737113571166992, 0.01741004753112793, 0.017305599212646485, 0.01746329689025879, 0.017443840026855468, 0.017361919403076173, 0.017316864013671874, 0.017252351760864256, 0.017300479888916014, 0.017376256942749024, 0.017330175399780275, 0.017303552627563477, 0.017426431655883787, 0.01737932777404785, 0.017320959091186524, 0.017399808883666993, 0.017484800338745117, 0.017327104568481445, 0.017406976699829102, 0.01740185546875, 0.017313791275024415, 0.017370111465454103, 0.017483776092529296, 0.01759846305847168, 0.017317888259887695, 0.017374208450317383, 0.017377279281616212, 0.017369087219238282, 0.017451007843017577, 0.01740595245361328, 0.017262592315673828, 0.017326080322265625, 0.01746124839782715, 0.01744179153442383, 0.017346559524536134, 0.01741209602355957, 0.017275903701782228, 0.017307647705078123, 0.01739571189880371, 0.01741926383972168, 0.017354751586914064, 0.017257471084594727, 0.017302528381347656, 0.017503231048583985, 0.017266687393188478, 0.017311744689941407, 0.017382400512695313, 0.017282047271728516, 0.017490943908691405, 0.05009519958496094, 0.017307552337646484, 0.017317888259887695, 0.017306623458862306, 0.01721343994140625, 0.017283071517944337, 0.017362943649291994, 0.01724928092956543, 0.01737932777404785, 0.017458175659179686, 0.017283071517944337, 0.017309696197509765, 0.017328128814697266, 0.017252351760864256, 0.01738751983642578, 0.017305599212646485, 0.017382400512695313, 0.01741209602355957, 0.01740390396118164, 0.017278976440429687, 0.017750015258789064, 0.017399808883666993, 0.01723494338989258, 0.017334272384643554, 0.017260543823242186, 0.01740595245361328, 0.017289215087890625, 0.017314815521240236, 0.017321983337402345, 0.017324031829833983, 0.017324031829833983, 0.017310720443725586, 0.017285120010375975, 0.01743974494934082, 0.01742131233215332, 0.017354751586914064, 0.01738035202026367, 0.017308671951293944, 0.01744691276550293, 0.01741107177734375, 0.017242111206054688, 0.017328128814697266, 0.017333248138427734, 0.017265663146972657, 0.017340415954589843, 0.017500160217285156, 0.017311744689941407, 0.01738956832885742, 0.017390592575073242, 0.01722572708129883, 0.017378303527832033, 0.017319936752319336, 0.01740390396118164, 0.017304576873779298, 0.017290239334106446, 0.017344512939453126, 0.017368064880371094, 0.017286144256591796, 0.0172677116394043, 0.017487871170043946, 0.017339391708374022, 0.01749504089355469, 0.01739263916015625, 0.049800193786621094, 0.017291263580322267, 0.017349632263183593, 0.017160192489624023, 0.01720012855529785, 0.017328191757202148, 0.01731884765625, 0.0172410888671875, 0.017360895156860352, 0.017345535278320313, 0.01722675132751465, 0.017340415954589843, 0.017229824066162108, 0.017456127166748048, 0.01724006462097168, 0.01721855926513672, 0.017364992141723632, 0.017293312072753905, 0.01721446418762207, 0.01725951957702637, 0.01739263916015625, 0.01726361656188965, 0.017357824325561523, 0.017304576873779298, 0.017359872817993165, 0.017223712921142578, 0.017224672317504883, 0.017339391708374022, 0.01723084831237793, 0.01740287971496582, 0.01740083122253418, 0.01723289680480957, 0.016495616912841796, 0.017150976181030272, 0.01742848014831543, 0.01728102493286133, 0.017315839767456053, 0.01721139144897461, 0.017309696197509765, 0.017236064910888672, 0.01741302490234375, 0.017356800079345702, 0.01725132751464844, 0.017314815521240236, 0.01739776039123535, 0.01744076728820801, 0.017324031829833983, 0.017345535278320313, 0.01721548843383789, 0.017309696197509765, 0.01742336082458496, 0.017286144256591796, 0.017334272384643554, 0.017340415954589843, 0.017313791275024415, 0.017291263580322267, 0.01739263916015625, 0.017278976440429687, 0.017398784637451172, 0.017282047271728516, 0.01726464080810547, 0.017330175399780275, 0.01740390396118164, 0.049688575744628906, 0.016590848922729492, 0.01722572708129883, 0.017236991882324217, 0.01741823959350586, 0.01741107177734375, 0.017364992141723632, 0.017415168762207032, 0.01724006462097168, 0.017332223892211913, 0.017362943649291994, 0.01729539108276367, 0.017276895523071288, 0.017362943649291994, 0.017353727340698243, 0.017330175399780275, 0.01740287971496582, 0.017180671691894533, 0.017316864013671874, 0.017313791275024415, 0.017283071517944337, 0.017378303527832033, 0.01740287971496582, 0.01738956832885742, 0.017540096282958984, 0.017309696197509765, 0.01745408058166504, 0.017292287826538084, 0.017330175399780275, 0.017312768936157227, 0.017242111206054688, 0.017335296630859375, 0.017377279281616212, 0.017426431655883787, 0.017300479888916014, 0.01737932777404785, 0.01735580825805664, 0.01728303909301758, 0.017316864013671874, 0.017308671951293944, 0.017345535278320313, 0.017269760131835937, 0.017368064880371094, 0.017373184204101562, 0.017305599212646485, 0.017360895156860352, 0.017366016387939453, 0.01727801513671875, 0.017268672943115234, 0.017360895156860352, 0.017284095764160155, 0.017504255294799806, 0.017335296630859375, 0.01724928092956543, 0.017426431655883787, 0.01725644874572754, 0.017287168502807617, 0.017338367462158204, 0.01721855926513672, 0.017353727340698243, 0.017315839767456053, 0.017340415954589843, 0.017313791275024415, 0.04982476806640625, 0.017253376007080077, 0.01723187255859375, 0.017376256942749024, 0.017280000686645508, 0.017299455642700197, 0.017318912506103516, 0.01741312026977539, 0.017300479888916014, 0.017312768936157227, 0.017286144256591796, 0.01728102493286133, 0.017292287826538084, 0.01724723243713379, 0.01740185546875, 0.01743667221069336, 0.01724723243713379, 0.017472511291503907, 0.017317888259887695, 0.01719193649291992, 0.017346559524536134, 0.017357824325561523, 0.017310720443725586, 0.017435647964477538, 0.017246208190917968, 0.017301504135131835, 0.01725132751464844, 0.017538047790527343, 0.01743667221069336, 0.017377279281616212, 0.01723494338989258, 0.017459199905395507, 0.017302591323852538, 0.017220544815063476, 0.017300479888916014, 0.017236991882324217, 0.0172728328704834, 0.017292287826538084, 0.017315839767456053, 0.017334272384643554, 0.017321983337402345, 0.017209344863891602, 0.01736396789550781, 0.017334272384643554, 0.017261568069458007, 0.01743769645690918, 0.01723904037475586, 0.017294336318969726, 0.017303552627563477, 0.017354751586914064, 0.017312768936157227, 0.017362943649291994, 0.01735577583312988, 0.017294336318969726, 0.017217536926269532, 0.01721446418762207, 0.017484800338745117, 0.01740902328491211, 0.017338367462158204, 0.01737932777404785, 0.01725951957702637, 0.017342464447021484, 0.01724723243713379, 0.050726913452148435, 0.017326080322265625, 0.01739366340637207, 0.017289215087890625, 0.01723289680480957, 0.017580032348632812, 0.017476608276367187, 0.017326080322265625, 0.017456127166748048, 0.01722777557373047, 0.017289215087890625, 0.017306655883789063, 0.017250272750854494, 0.017369087219238282, 0.01734758377075195, 0.017303552627563477, 0.017310720443725586, 0.01722470474243164, 0.017286144256591796, 0.01737215995788574, 0.017291263580322267, 0.017307647705078123, 0.017496063232421876, 0.017382400512695313, 0.017442815780639647, 0.017447935104370118, 0.01765990447998047, 0.017970176696777345, 0.017509376525878906, 0.01743155288696289, 0.01743974494934082, 0.017375232696533204, 0.017375232696533204, 0.01741926383972168, 0.017337343215942384, 0.017693695068359376, 0.017286144256591796, 0.017342464447021484, 0.01725542449951172, 0.017366016387939453, 0.017265663146972657, 0.017361919403076173, 0.01719398307800293, 0.017270784378051757, 0.01741004753112793, 0.01759846305847168, 0.017374208450317383, 0.017313791275024415, 0.017311744689941407, 0.017258495330810548, 0.017296384811401368, 0.017277952194213866, 0.017386495590209963, 0.017228799819946287, 0.017278976440429687, 0.017581056594848633, 0.017290239334106446, 0.017383424758911133, 0.01817804718017578, 0.017864704132080078, 0.017509376525878906, 0.01741004753112793, 0.017345535278320313]",tokens/s,55.98778304173951,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,huggyllama/llama-65b,huggyllama/llama-65b,cuda,0,42,,,True,True,True,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ self.load_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 250, in load_model_with_no_weights self.pretrained_model = self.automodel_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3710, in from_pretrained model = cls(config, *model_args, **model_kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 1101, in __init__ self.model = LlamaModel(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 884, in __init__ [LlamaDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 884, in [LlamaDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 679, in __init__ self.mlp = LlamaMLP(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 194, in __init__ self.down_proj = nn.Linear(self.intermediate_size, self.hidden_size, bias=config.mlp_bias) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/linear.py"", line 98, in __init__ self.weight = Parameter(torch.empty((out_features, in_features), **factory_kwargs)) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_device.py"", line 78, in __torch_function__ return func(*args, **kwargs) torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 688.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,l,l,cuda,0,42,,,True,True,True,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/l/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6694744e-712a93c4783b51d04ae3344d;d5c632eb-8c05-40f0-b2f5-49e14531a52f) Repository Not Found for url: https://huggingface.co/l/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: l is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,B,B,cuda,0,42,,,True,True,True,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/B/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66932502-5f65831e67da07b462d69ecc;a3fe9487-4f56-4732-8745-8f7ddf82dddf) Repository Not Found for url: https://huggingface.co/B/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: B is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,huggyllama/llama-30b,huggyllama/llama-30b,cuda,0,42,,,True,,,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",llama,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,meta-llama/Llama-2-13b-hf,meta-llama/Llama-2-13b-hf,cuda,0,42,,,True,True,True,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ self.load_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 250, in load_model_with_no_weights self.pretrained_model = self.automodel_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3710, in from_pretrained model = cls(config, *model_args, **model_kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 1101, in __init__ self.model = LlamaModel(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 884, in __init__ [LlamaDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 884, in [LlamaDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 677, in __init__ self.self_attn = LLAMA_ATTENTION_CLASSES[config._attn_implementation](config=config, layer_idx=layer_idx) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 262, in __init__ self.q_proj = nn.Linear(self.hidden_size, self.num_heads * self.head_dim, bias=config.attention_bias) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/linear.py"", line 98, in __init__ self.weight = Parameter(torch.empty((out_features, in_features), **factory_kwargs)) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_device.py"", line 78, in __torch_function__ return func(*args, **kwargs) torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 100.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-7B,Qwen/Qwen1.5-7B,cuda,0,42,,,True,,,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",qwen2,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,M,M,cuda,0,42,,,True,True,True,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/M/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66932295-0d83d2190c6f7e4b0758d99a;8848d028-6c6c-4202-9b2a-0eb772e581ea) Repository Not Found for url: https://huggingface.co/M/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: M is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,stabilityai/stablelm-base-alpha-3b,stabilityai/stablelm-base-alpha-3b,cuda,0,42,,,True,True,True,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.42.3,,0.31.0,,,,1.20.0,,,,0.11.1,,,,MB,1116.28288,15947.268096,0.0,15307.112448,15168.111104,s,10,1.0988862762451173,0.10988862762451172,0.0005380568377233971,0.10978670501708984,0.11060799026489258,0.11075855445861817,0.11087900581359864,"[0.11057453155517578, 0.10984390258789062, 0.10947325134277344, 0.10972950744628907, 0.11031839752197266, 0.10949616241455078, 0.10999590301513672, 0.11090911865234375, 0.10946572875976562, 0.10907977294921875]",tokens/s,2329.631423505891,kWh,1.3403258780563965e-06,7.344417003514514e-07,9.00521631777531e-06,1.1079983896183158e-05,tokens/kWh,23104726.721506074,MB,1116.606464,15947.268096,0.0,15307.112448,15247.910912,s,10,19.436102783203125,1.9436102783203126,0.0019726647912775373,1.9427960205078125,1.9448706298828125,1.9470916259765625,1.9488684228515625,"[1.9493126220703125, 1.9425208740234374, 1.9426795654296876, 1.9443770751953124, 1.9428212890625, 1.9428489990234374, 1.942770751953125, 1.9426640625, 1.943462890625, 1.9426446533203126]",tokens/s,32.41390555643965,kWh,2.2905737917298238e-05,1.2552945304920392e-05,0.00010370754051902718,0.00013916622374124577,tokens/kWh,452696.0515730959,,s,629,20.392660993576072,0.032420764695669395,0.01312721985273252,0.03084492874145508,0.030908621215820314,0.031032934951782228,0.14104846069335936,"[0.031355903625488284, 0.03121151924133301, 0.031152128219604492, 0.031153152465820313, 0.031160320281982422, 0.03120128059387207, 0.03115724754333496, 0.031169536590576172, 0.03115007972717285, 0.03118387222290039, 0.031138816833496095, 0.03113369560241699, 0.03118387222290039, 0.030852096557617188, 0.03079475212097168, 0.03080601692199707, 0.03078041648864746, 0.03078144073486328, 0.030757888793945313, 0.03100262451171875, 0.03079884719848633, 0.030766080856323243, 0.03078758430480957, 0.030765056610107422, 0.03078041648864746, 0.03076198387145996, 0.03079475212097168, 0.03080499267578125, 0.03077734375, 0.03097907257080078, 0.03079782485961914, 0.030859264373779297, 0.030851072311401367, 0.030857215881347655, 0.030855167388916017, 0.030845951080322266, 0.030859264373779297, 0.030856191635131838, 0.030858240127563476, 0.03083776092529297, 0.030854143142700196, 0.030859264373779297, 0.030864383697509764, 0.030861312866210938, 0.030872575759887694, 0.030860288619995117, 0.030879743576049806, 0.030861312866210938, 0.030892032623291016, 0.030870527267456056, 0.030905344009399413, 0.030874624252319335, 0.031014911651611327, 0.030882816314697265, 0.030870527267456056, 0.03102003288269043, 0.030896127700805662, 0.030872575759887694, 0.030885887145996094, 0.030897151947021483, 0.030895103454589845, 0.030939136505126953, 0.14132223510742187, 0.030917631149291993, 0.03079475212097168, 0.03076710319519043, 0.03076300811767578, 0.03075379180908203, 0.03078041648864746, 0.0307640323638916, 0.0307589111328125, 0.03077529525756836, 0.030769151687622072, 0.030876672744750977, 0.030740480422973632, 0.030757888793945313, 0.03075379180908203, 0.030744575500488282, 0.030760959625244142, 0.03076300811767578, 0.03075379180908203, 0.03078144073486328, 0.030769151687622072, 0.03078656005859375, 0.030757888793945313, 0.030757888793945313, 0.03077631950378418, 0.030757888793945313, 0.03078451156616211, 0.03076198387145996, 0.030814207077026368, 0.030819328308105468, 0.03078656005859375, 0.03077529525756836, 0.030870527267456056, 0.030834688186645507, 0.030849023818969725, 0.030849023818969725, 0.030854143142700196, 0.030834688186645507, 0.030852096557617188, 0.030886911392211915, 0.030841856002807616, 0.030849023818969725, 0.03085312080383301, 0.030864383697509764, 0.030860288619995117, 0.030851072311401367, 0.030871551513671876, 0.030864383697509764, 0.030866432189941406, 0.030864383697509764, 0.030874624252319335, 0.030862335205078126, 0.030889984130859374, 0.030886911392211915, 0.030872575759887694, 0.030877695083618165, 0.030878719329833985, 0.030871551513671876, 0.030911487579345705, 0.030888959884643553, 0.030900224685668946, 0.030922752380371094, 0.030889984130859374, 0.1410662384033203, 0.030946304321289062, 0.030826496124267577, 0.030785535812377928, 0.03074355125427246, 0.03075993537902832, 0.030757888793945313, 0.030748672485351562, 0.03075584030151367, 0.030757888793945313, 0.03077120018005371, 0.03076198387145996, 0.03076300811767578, 0.03077324867248535, 0.030757888793945313, 0.030766080856323243, 0.030748672485351562, 0.03075584030151367, 0.03077017593383789, 0.03077734375, 0.03075584030151367, 0.030760959625244142, 0.03077120018005371, 0.030766080856323243, 0.03077222442626953, 0.03077836799621582, 0.03078348731994629, 0.030847999572753908, 0.03078860855102539, 0.03082035255432129, 0.0307906551361084, 0.030785535812377928, 0.030845951080322266, 0.030849023818969725, 0.030834688186645507, 0.030861312866210938, 0.030850048065185546, 0.030846975326538087, 0.030843904495239258, 0.030855167388916017, 0.030897151947021483, 0.030849023818969725, 0.030863359451293947, 0.030852096557617188, 0.030857215881347655, 0.030875648498535156, 0.03084492874145508, 0.030868480682373047, 0.030854143142700196, 0.030868480682373047, 0.030876672744750977, 0.030887935638427736, 0.030864383697509764, 0.030887935638427736, 0.030883840560913086, 0.030869504928588868, 0.030876672744750977, 0.030879743576049806, 0.030888959884643553, 0.030938112258911132, 0.030930944442749023, 0.030886911392211915, 0.030888959884643553, 0.14128128051757813, 0.03095961570739746, 0.03079987144470215, 0.03075481605529785, 0.030748672485351562, 0.03074764823913574, 0.03075071907043457, 0.03076300811767578, 0.03099443244934082, 0.0307640323638916, 0.03077017593383789, 0.03077939224243164, 0.030741504669189453, 0.03077222442626953, 0.03075481605529785, 0.0307640323638916, 0.03076198387145996, 0.03078246307373047, 0.03075993537902832, 0.03075481605529785, 0.030748672485351562, 0.03076300811767578, 0.030774272918701173, 0.03075481605529785, 0.03075584030151367, 0.030858240127563476, 0.03080703926086426, 0.03078246307373047, 0.03077324867248535, 0.030785535812377928, 0.03078758430480957, 0.03076812744140625, 0.030850048065185546, 0.030852096557617188, 0.030843904495239258, 0.030841856002807616, 0.030867456436157226, 0.031185920715332032, 0.030908416748046875, 0.030852096557617188, 0.030883840560913086, 0.030845951080322266, 0.030840831756591795, 0.030840831756591795, 0.030845951080322266, 0.030869504928588868, 0.030871551513671876, 0.030876672744750977, 0.030877695083618165, 0.030873600006103515, 0.030870527267456056, 0.030900224685668946, 0.030869504928588868, 0.030864383697509764, 0.030879743576049806, 0.031297536849975584, 0.030906368255615234, 0.03134668731689453, 0.030916608810424805, 0.031080448150634765, 0.03099545669555664, 0.03091967964172363, 0.030886911392211915, 0.14183424377441406, 0.030938112258911132, 0.03079680061340332, 0.030757888793945313, 0.03076198387145996, 0.03079475212097168, 0.030760959625244142, 0.0307589111328125, 0.030752767562866212, 0.0307640323638916, 0.03075993537902832, 0.03075993537902832, 0.030757888793945313, 0.03075174331665039, 0.030744575500488282, 0.03076198387145996, 0.03077734375, 0.0307640323638916, 0.030748672485351562, 0.03076812744140625, 0.030757888793945313, 0.03076812744140625, 0.03077324867248535, 0.03078348731994629, 0.03076300811767578, 0.030769151687622072, 0.03078758430480957, 0.03078041648864746, 0.03078860855102539, 0.03077734375, 0.03078144073486328, 0.03078758430480957, 0.030839807510375978, 0.030855167388916017, 0.030841856002807616, 0.030855167388916017, 0.030840831756591795, 0.030864383697509764, 0.030858240127563476, 0.030865407943725585, 0.030843904495239258, 0.030899200439453125, 0.030873600006103515, 0.030918655395507814, 0.030862335205078126, 0.03118284797668457, 0.030903295516967775, 0.030862335205078126, 0.030873600006103515, 0.030863359451293947, 0.030869504928588868, 0.030878719329833985, 0.030866432189941406, 0.030875648498535156, 0.030872575759887694, 0.030865407943725585, 0.030896127700805662, 0.030901248931884766, 0.030869504928588868, 0.030882816314697265, 0.030898176193237304, 0.030885887145996094, 0.030897151947021483, 0.14108262634277344, 0.03095961570739746, 0.03081113624572754, 0.030745599746704103, 0.03073843193054199, 0.03074662399291992, 0.030774272918701173, 0.03078758430480957, 0.030757888793945313, 0.03077324867248535, 0.03075071907043457, 0.030760959625244142, 0.031094783782958983, 0.03079987144470215, 0.03080396842956543, 0.03073945617675781, 0.03075379180908203, 0.03075993537902832, 0.030760959625244142, 0.03077734375, 0.0307640323638916, 0.03075993537902832, 0.03077631950378418, 0.03074355125427246, 0.03075993537902832, 0.03076198387145996, 0.03077222442626953, 0.03075993537902832, 0.0307906551361084, 0.03076812744140625, 0.03078758430480957, 0.030902271270751954, 0.030849023818969725, 0.030839807510375978, 0.03096575927734375, 0.030839807510375978, 0.030835712432861328, 0.030838783264160157, 0.03083776092529297, 0.030869504928588868, 0.030866432189941406, 0.030857215881347655, 0.030849023818969725, 0.03083673667907715, 0.03084492874145508, 0.030855167388916017, 0.030889984130859374, 0.030846975326538087, 0.030862335205078126, 0.030871551513671876, 0.030886911392211915, 0.030858240127563476, 0.030866432189941406, 0.030863359451293947, 0.030899200439453125, 0.030876672744750977, 0.030870527267456056, 0.030871551513671876, 0.030870527267456056, 0.030871551513671876, 0.030875648498535156, 0.030879743576049806, 0.030906368255615234, 0.14153523254394532, 0.030948352813720704, 0.030810111999511718, 0.03075584030151367, 0.03074355125427246, 0.030749696731567383, 0.03076710319519043, 0.030774272918701173, 0.030757888793945313, 0.03076198387145996, 0.03076710319519043, 0.03075174331665039, 0.030801919937133788, 0.03077120018005371, 0.03074764823913574, 0.03077017593383789, 0.030752767562866212, 0.030757888793945313, 0.03076710319519043, 0.0307640323638916, 0.03080396842956543, 0.03082444763183594, 0.03076300811767578, 0.0307589111328125, 0.0307640323638916, 0.030765056610107422, 0.03079475212097168, 0.03077017593383789, 0.03078348731994629, 0.03077120018005371, 0.03080089569091797, 0.03078656005859375, 0.030840831756591795, 0.030841856002807616, 0.030835712432861328, 0.03084492874145508, 0.030862335205078126, 0.030856191635131838, 0.030875648498535156, 0.030868480682373047, 0.030846975326538087, 0.030850048065185546, 0.03083673667907715, 0.030869504928588868, 0.030852096557617188, 0.030852096557617188, 0.030862335205078126, 0.030862335205078126, 0.030874624252319335, 0.030880767822265624, 0.030899200439453125, 0.030891008377075195, 0.030865407943725585, 0.030881792068481444, 0.030871551513671876, 0.030889984130859374, 0.030865407943725585, 0.030868480682373047, 0.030873600006103515, 0.030882816314697265, 0.030874624252319335, 0.03116646385192871, 0.030908416748046875, 0.14258688354492188, 0.031156223297119142, 0.03077120018005371, 0.030745599746704103, 0.03075993537902832, 0.03079782485961914, 0.03074355125427246, 0.030757888793945313, 0.03077017593383789, 0.030860288619995117, 0.03075993537902832, 0.030756864547729492, 0.030756864547729492, 0.03074355125427246, 0.03075584030151367, 0.03073945617675781, 0.03075993537902832, 0.030749696731567383, 0.0307589111328125, 0.030766080856323243, 0.03074764823913574, 0.03075584030151367, 0.030760959625244142, 0.0307589111328125, 0.03075481605529785, 0.030760959625244142, 0.03077017593383789, 0.03079475212097168, 0.0307957763671875, 0.0307906551361084, 0.03079884719848633, 0.03077324867248535, 0.030842880249023437, 0.03082444763183594, 0.03083263969421387, 0.030841856002807616, 0.030851072311401367, 0.030855167388916017, 0.030854143142700196, 0.030842880249023437, 0.03083776092529297, 0.03085312080383301, 0.030859264373779297, 0.030846975326538087, 0.030842880249023437, 0.030859264373779297, 0.030920703887939452, 0.030854143142700196, 0.030863359451293947, 0.030859264373779297, 0.030883840560913086, 0.030882816314697265, 0.030865407943725585, 0.030869504928588868, 0.030868480682373047, 0.030866432189941406, 0.030883840560913086, 0.030858240127563476, 0.030874624252319335, 0.030901248931884766, 0.03097292709350586, 0.030891008377075195, 0.030882816314697265, 0.14100274658203124, 0.030921728134155273, 0.03080499267578125, 0.03076812744140625, 0.03075379180908203, 0.030748672485351562, 0.030748672485351562, 0.030736383438110353, 0.030765056610107422, 0.03078246307373047, 0.030766080856323243, 0.030765056610107422, 0.03077017593383789, 0.030756864547729492, 0.03074764823913574, 0.03074662399291992, 0.03075993537902832, 0.03080703926086426, 0.03075379180908203, 0.03077017593383789, 0.03077324867248535, 0.03077017593383789, 0.0307589111328125, 0.03076300811767578, 0.03076710319519043, 0.03080089569091797, 0.03078963279724121, 0.03077324867248535, 0.030889984130859374, 0.030929920196533203, 0.03081113624572754, 0.03081625556945801, 0.03083263969421387, 0.030841856002807616, 0.03084492874145508, 0.030888959884643553, 0.030851072311401367, 0.03085312080383301, 0.030845951080322266, 0.030847999572753908, 0.030842880249023437, 0.030859264373779297, 0.030865407943725585, 0.030864383697509764, 0.030868480682373047, 0.030861312866210938, 0.031109119415283205, 0.0309616641998291, 0.030882816314697265, 0.030916608810424805, 0.030877695083618165, 0.031041536331176758, 0.030920703887939452, 0.030897151947021483, 0.030864383697509764, 0.030895103454589845, 0.030904319763183592, 0.030904319763183592, 0.030905344009399413, 0.030874624252319335, 0.030889984130859374, 0.030905344009399413, 0.030904319763183592, 0.14063717651367189, 0.03097395133972168, 0.030825471878051756, 0.03077120018005371, 0.030891008377075195, 0.030749696731567383, 0.030765056610107422, 0.03080396842956543, 0.03076300811767578, 0.0307589111328125, 0.03077324867248535, 0.03075481605529785, 0.030765056610107422, 0.03074355125427246, 0.030752767562866212, 0.030830591201782227, 0.03077529525756836, 0.03074662399291992, 0.030748672485351562, 0.030760959625244142, 0.030752767562866212, 0.030745599746704103, 0.03075379180908203, 0.03075993537902832, 0.0307640323638916, 0.030769151687622072, 0.03077120018005371, 0.03077836799621582, 0.0307957763671875, 0.03078656005859375, 0.03077120018005371, 0.03077324867248535, 0.030846975326538087, 0.03085312080383301, 0.030859264373779297, 0.030924800872802735, 0.030840831756591795, 0.030854143142700196, 0.030840831756591795, 0.030862335205078126, 0.030852096557617188, 0.03084492874145508, 0.030861312866210938, 0.030862335205078126, 0.030868480682373047, 0.03085312080383301, 0.030867456436157226, 0.030870527267456056, 0.030888959884643553, 0.030866432189941406, 0.030872575759887694, 0.030870527267456056, 0.030868480682373047, 0.030888959884643553, 0.030872575759887694, 0.030862335205078126, 0.030909439086914063, 0.030885887145996094, 0.030881792068481444, 0.030882816314697265, 0.030891008377075195, 0.030886911392211915, 0.030880767822265624]",tokens/s,30.844429777857,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,google/recurrentgemma-9b,google/recurrentgemma-9b,cuda,0,42,,,True,,,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",recurrent_gemma,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,mistralai/Mixtral-8x22B-v0.1,mistralai/Mixtral-8x22B-v0.1,cuda,0,42,,,True,True,True,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/base.py"", line 43, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 54, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1021, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 403 Client Error: Forbidden for url: https://huggingface.co/mistralai/Mixtral-8x22B-v0.1/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 398, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 321, in hf_raise_for_status raise GatedRepoError(message, response) from e huggingface_hub.utils._errors.GatedRepoError: 403 Client Error. (Request ID: Root=1-6647507c-615a64826fa2ef4b08c75083;e1b8c738-8df0-4850-8bdd-9aea9042c908) Cannot access gated repo for url https://huggingface.co/mistralai/Mixtral-8x22B-v0.1/resolve/main/config.json. Access to model mistralai/Mixtral-8x22B-v0.1 is restricted and you are not in the authorized list. Visit https://huggingface.co/mistralai/Mixtral-8x22B-v0.1 to ask for access. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 87, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/base.py"", line 56, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 37, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 62, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 928, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 631, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 686, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 416, in cached_file raise EnvironmentError( OSError: You are trying to access a gated repo. Make sure to have access to it at https://huggingface.co/mistralai/Mixtral-8x22B-v0.1. 403 Client Error. (Request ID: Root=1-6647507c-615a64826fa2ef4b08c75083;e1b8c738-8df0-4850-8bdd-9aea9042c908) Cannot access gated repo for url https://huggingface.co/mistralai/Mixtral-8x22B-v0.1/resolve/main/config.json. Access to model mistralai/Mixtral-8x22B-v0.1 is restricted and you are not in the authorized list. Visit https://huggingface.co/mistralai/Mixtral-8x22B-v0.1 to ask for access. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,01-ai/Yi-6B,01-ai/Yi-6B,cuda,0,42,,,True,,,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",llama,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,/,/,cuda,0,42,,,True,True,True,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 373, in cached_file raise EnvironmentError( OSError: / does not appear to have a file named config.json. Checkout 'https://huggingface.co///tree/None' for available files. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,TencentARC/Mistral_Pro_8B_v0.1,TencentARC/Mistral_Pro_8B_v0.1,cuda,0,42,,,True,,,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",mistral,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-6.7b,EleutherAI/pythia-6.7b,cuda,0,42,,,True,,,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",gpt_neox,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-13b,facebook/opt-13b,cuda,0,42,,,True,,,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",opt,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-6.7b,facebook/opt-6.7b,cuda,0,42,,,True,,,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",opt,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,microsoft/rho-math-1b-v0.1,microsoft/rho-math-1b-v0.1,cuda,0,42,,,True,True,True,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1135.742976,5172.10112,0.0,4531.945472,4478.990336,s,10,0.3975123519897461,0.03975123519897461,0.0002546493339729439,0.039652048110961916,0.03992862548828125,0.040197560501098635,0.04041270851135254,"[0.04046649551391602, 0.039794273376464843, 0.03960784149169922, 0.039637569427490235, 0.039667007446289065, 0.0395384635925293, 0.03963785552978515, 0.03962774276733398, 0.03966624069213867, 0.03986886215209961]",tokens/s,6440.051452957205,kWh,4.958956027407804e-07,2.717270876475679e-07,3.4272040434811473e-06,4.194826733869495e-06,tokens/kWh,61027550.418954775,MB,1135.742976,5172.10112,0.0,4531.945472,4492.86912,s,10,13.15693322753906,1.3156933227539063,0.012732326368063573,1.3180799560546874,1.3296379516601562,1.331017889404297,1.3321218395996095,"[1.3323978271484376, 1.3132144775390624, 1.3219439697265625, 1.31023486328125, 1.2851285400390624, 1.316025634765625, 1.329331298828125, 1.322101318359375, 1.32013427734375, 1.3064210205078124]",tokens/s,47.88349907266637,kWh,1.5338403623861138e-05,8.405267576681391e-06,4.508622365611892e-05,6.882989485666146e-05,tokens/kWh,915299.9598676965,,s,629,13.492060171127314,0.021450016170313703,0.004714803071535541,0.021008384704589843,0.021200077438354493,0.02137784385681152,0.0603588200378418,"[0.02043187141418457, 0.020913152694702147, 0.020732927322387695, 0.02106675148010254, 0.021152767181396484, 0.021196800231933592, 0.021191680908203125, 0.021094400405883788, 0.021115903854370118, 0.021109760284423826, 0.021183488845825195, 0.021596160888671875, 0.021082111358642578, 0.02109542465209961, 0.021053440093994142, 0.021124095916748048, 0.02110771179199219, 0.021151744842529296, 0.021168127059936523, 0.02110771179199219, 0.021214208602905273, 0.021329919815063478, 0.0212490234375, 0.021102655410766603, 0.021130176544189454, 0.02142207908630371, 0.021587968826293946, 0.021522432327270507, 0.021393407821655275, 0.021209087371826172, 0.02125312042236328, 0.021196800231933592, 0.02124595260620117, 0.02107596778869629, 0.021147647857666017, 0.02110771179199219, 0.021046272277832033, 0.02110771179199219, 0.021166080474853514, 0.021045248031616212, 0.020993024826049804, 0.02105855941772461, 0.021174272537231444, 0.021167104721069335, 0.02122137641906738, 0.021111808776855468, 0.021230592727661132, 0.021004287719726563, 0.021086208343505858, 0.02126950454711914, 0.021152767181396484, 0.021131263732910157, 0.020728832244873048, 0.021046272277832033, 0.02105958366394043, 0.02100223922729492, 0.021170175552368165, 0.02105958366394043, 0.02147225570678711, 0.02107084846496582, 0.02106470489501953, 0.020653055191040038, 0.06047948837280273, 0.020197376251220703, 0.020176895141601564, 0.02024345588684082, 0.020132863998413086, 0.02026905632019043, 0.020198400497436524, 0.020126720428466797, 0.020121599197387697, 0.020187135696411132, 0.020164608001708984, 0.020231168746948244, 0.02021990394592285, 0.020208639144897463, 0.02167296028137207, 0.021166112899780272, 0.02107491111755371, 0.021005311965942384, 0.02107596778869629, 0.021152767181396484, 0.020932607650756836, 0.02092748832702637, 0.020957183837890626, 0.020913152694702147, 0.021032960891723632, 0.0208353271484375, 0.02068172836303711, 0.020720640182495118, 0.02074015998840332, 0.020693952560424805, 0.021008384704589843, 0.02110873603820801, 0.020988927841186524, 0.021021696090698243, 0.02105958366394043, 0.021012479782104493, 0.020993024826049804, 0.02104012870788574, 0.020988927841186524, 0.020956159591674805, 0.02101862335205078, 0.020943872451782225, 0.02142310333251953, 0.02090598487854004, 0.02106265640258789, 0.02105855941772461, 0.02109542465209961, 0.020950016021728517, 0.020999168395996092, 0.02102579116821289, 0.021195775985717775, 0.021112831115722656, 0.021000192642211913, 0.02104115104675293, 0.02122444725036621, 0.021137407302856445, 0.021137439727783203, 0.021055456161499023, 0.020993024826049804, 0.021014528274536134, 0.02091929626464844, 0.021000192642211913, 0.020961280822753905, 0.0604139518737793, 0.020188159942626953, 0.020197376251220703, 0.020792320251464845, 0.02129715156555176, 0.021787647247314454, 0.021170175552368165, 0.02109132766723633, 0.020982784271240236, 0.020940799713134766, 0.020962303161621093, 0.02089779281616211, 0.02066534423828125, 0.020855808258056642, 0.020972543716430665, 0.020757503509521484, 0.020767744064331056, 0.02086502456665039, 0.020926528930664063, 0.021031871795654297, 0.02103193664550781, 0.02104319953918457, 0.020995071411132812, 0.020978687286376953, 0.021174272537231444, 0.02106675148010254, 0.020946943283081054, 0.021098495483398438, 0.020993024826049804, 0.02104217529296875, 0.021120000839233398, 0.020996095657348633, 0.020942848205566408, 0.021012544631958008, 0.02105132865905762, 0.02106265640258789, 0.02101759910583496, 0.020983808517456053, 0.02099404716491699, 0.020993024826049804, 0.020925439834594727, 0.02108006477355957, 0.02100223922729492, 0.021003263473510742, 0.021005311965942384, 0.021004287719726563, 0.020978687286376953, 0.02101043128967285, 0.020972543716430665, 0.02110873603820801, 0.02109542465209961, 0.021115903854370118, 0.020942848205566408, 0.02105036735534668, 0.020945920944213867, 0.021009408950805664, 0.02104115104675293, 0.02105241584777832, 0.02106470489501953, 0.02104012870788574, 0.021008384704589843, 0.02102272033691406, 0.02104422378540039, 0.06033817672729492, 0.02025164794921875, 0.020212736129760742, 0.020356096267700196, 0.020153343200683595, 0.02023628807067871, 0.02053222465515137, 0.020296703338623046, 0.020189184188842774, 0.02021683120727539, 0.020215808868408205, 0.020184064865112306, 0.020147199630737304, 0.02023423957824707, 0.02024345588684082, 0.02023526382446289, 0.02022809600830078, 0.02024038314819336, 0.019929088592529298, 0.020762624740600585, 0.02106982421875, 0.020896768569946288, 0.021038080215454103, 0.02110361671447754, 0.02103910446166992, 0.020974592208862306, 0.020915199279785156, 0.020970495223999023, 0.020702207565307617, 0.020757503509521484, 0.02088140869140625, 0.02269491195678711, 0.02131865692138672, 0.02108518409729004, 0.02098182487487793, 0.021071807861328125, 0.020999200820922853, 0.02106262397766113, 0.02103910446166992, 0.02099404716491699, 0.020907007217407226, 0.021007360458374022, 0.02103398323059082, 0.02089779281616211, 0.02101043128967285, 0.021015552520751952, 0.02103606414794922, 0.020960224151611327, 0.02100223922729492, 0.021001216888427734, 0.020976640701293944, 0.020923391342163086, 0.02105753517150879, 0.02090598487854004, 0.02093062400817871, 0.021104576110839844, 0.021372928619384765, 0.021001216888427734, 0.020979711532592774, 0.02089472007751465, 0.02106060791015625, 0.020970495223999023, 0.021053440093994142, 0.060386302947998044, 0.020133888244628906, 0.020171775817871093, 0.02023321533203125, 0.020175872802734376, 0.020188159942626953, 0.02018611145019531, 0.02026393508911133, 0.020199424743652345, 0.020222976684570314, 0.020264959335327147, 0.020221952438354493, 0.020205568313598633, 0.020134912490844727, 0.020239360809326173, 0.020256767272949217, 0.020212736129760742, 0.020358144760131838, 0.020221952438354493, 0.02023219108581543, 0.020176895141601564, 0.0202106876373291, 0.02025164794921875, 0.020184064865112306, 0.02022707176208496, 0.02025267219543457, 0.020304895401000975, 0.020404224395751954, 0.02028339195251465, 0.020185087203979494, 0.01999564743041992, 0.020125696182250977, 0.020174848556518556, 0.020197376251220703, 0.02023321533203125, 0.020246528625488282, 0.020230144500732423, 0.020192256927490236, 0.020249599456787108, 0.020625408172607423, 0.020185087203979494, 0.02027212715148926, 0.020214784622192384, 0.020213760375976563, 0.020321279525756835, 0.020288511276245116, 0.020180992126464844, 0.02021683120727539, 0.020166656494140626, 0.020561920166015626, 0.02112928009033203, 0.021022655487060546, 0.021016576766967773, 0.020997119903564454, 0.021011455535888672, 0.021097471237182617, 0.020982784271240236, 0.02103910446166992, 0.020960256576538085, 0.021081087112426757, 0.021004287719726563, 0.020907007217407226, 0.020695039749145508, 0.0602470703125, 0.02016764831542969, 0.020189184188842774, 0.020199424743652345, 0.020153343200683595, 0.020168703079223634, 0.020146175384521483, 0.020158464431762696, 0.020213760375976563, 0.020247552871704103, 0.020166656494140626, 0.020140031814575195, 0.020988927841186524, 0.020925439834594727, 0.02106675148010254, 0.021545984268188476, 0.021132287979125978, 0.02087833595275879, 0.020917247772216797, 0.020935680389404295, 0.020982784271240236, 0.020983808517456053, 0.020930559158325195, 0.02142207908630371, 0.021004287719726563, 0.020965375900268556, 0.020980735778808594, 0.020921344757080077, 0.02108723258972168, 0.021218303680419923, 0.022168575286865236, 0.021304319381713867, 0.021300224304199217, 0.02105446434020996, 0.021155839920043946, 0.021098495483398438, 0.02104217529296875, 0.021008384704589843, 0.020966400146484376, 0.020970495223999023, 0.020999168395996092, 0.0206059513092041, 0.020985919952392577, 0.02094483184814453, 0.02107494354248047, 0.02102579116821289, 0.020948991775512696, 0.02067865562438965, 0.02106982421875, 0.02106368064880371, 0.020971519470214844, 0.021008384704589843, 0.02123263931274414, 0.02105958366394043, 0.021045248031616212, 0.021134336471557616, 0.021150720596313476, 0.020955135345458984, 0.021021696090698243, 0.020816896438598635, 0.02068172836303711, 0.020958208084106447, 0.021007360458374022, 0.06107648086547852, 0.020995071411132812, 0.020976640701293944, 0.020980735778808594, 0.021007360458374022, 0.021505023956298826, 0.021173248291015623, 0.02103398323059082, 0.020961280822753905, 0.020959232330322267, 0.020993024826049804, 0.02129100799560547, 0.021020671844482423, 0.020995136260986327, 0.02105952072143555, 0.021021696090698243, 0.020983808517456053, 0.021109760284423826, 0.020961280822753905, 0.021016576766967773, 0.021021696090698243, 0.02110873603820801, 0.021132287979125978, 0.020989952087402345, 0.02106368064880371, 0.02104319953918457, 0.020996095657348633, 0.02106982421875, 0.020965375900268556, 0.02103910446166992, 0.02150809669494629, 0.02129817581176758, 0.021082111358642578, 0.021120000839233398, 0.021171199798583985, 0.02104832077026367, 0.021081087112426757, 0.020967424392700194, 0.021164031982421876, 0.021169151306152344, 0.02168217658996582, 0.02204979133605957, 0.021548032760620117, 0.021113855361938477, 0.02118147277832031, 0.021089248657226563, 0.02107187271118164, 0.02100223922729492, 0.021008384704589843, 0.020983808517456053, 0.02104012870788574, 0.020989952087402345, 0.02103603172302246, 0.021381120681762695, 0.020976640701293944, 0.020907007217407226, 0.020690944671630858, 0.02091929626464844, 0.021014528274536134, 0.020990976333618162, 0.0210882568359375, 0.020990976333618162, 0.021098495483398438, 0.06039449691772461, 0.02021990394592285, 0.020229120254516602, 0.0202608642578125, 0.020198400497436524, 0.02087424087524414, 0.02103091239929199, 0.021014528274536134, 0.021106687545776368, 0.021053440093994142, 0.021008384704589843, 0.02102681541442871, 0.02100223922729492, 0.021198848724365234, 0.02107187271118164, 0.021194751739501954, 0.021172224044799806, 0.020951072692871095, 0.02091209602355957, 0.02109651184082031, 0.020949951171875, 0.02088755226135254, 0.020808704376220705, 0.020801536560058592, 0.02102579116821289, 0.021156864166259767, 0.021114879608154297, 0.02103398323059082, 0.021098495483398438, 0.021029888153076173, 0.021140480041503908, 0.02107494354248047, 0.02103500747680664, 0.02121625518798828, 0.021126144409179686, 0.021073919296264648, 0.02147532844543457, 0.021143552780151367, 0.02105548858642578, 0.021012479782104493, 0.02104012870788574, 0.021038080215454103, 0.021163007736206055, 0.021124095916748048, 0.021169151306152344, 0.021133312225341795, 0.02124595260620117, 0.021098495483398438, 0.021089279174804687, 0.021173248291015623, 0.020959232330322267, 0.020962303161621093, 0.02131865692138672, 0.0209039363861084, 0.02082611274719238, 0.020962303161621093, 0.020997119903564454, 0.021165056228637694, 0.02126848030090332, 0.02105548858642578, 0.021114879608154297, 0.02043391990661621, 0.02026700782775879, 0.06036684799194336, 0.020290592193603515, 0.020193248748779296, 0.02024448013305664, 0.02021785545349121, 0.020340736389160157, 0.02042572784423828, 0.02036735916137695, 0.02110771179199219, 0.020945920944213867, 0.021104639053344726, 0.02100223922729492, 0.02104729652404785, 0.021204992294311522, 0.020979711532592774, 0.02106572723388672, 0.02101862335205078, 0.020963327407836914, 0.021321727752685548, 0.02103603172302246, 0.021004287719726563, 0.02101862335205078, 0.02106675148010254, 0.020986879348754883, 0.020948991775512696, 0.02109337615966797, 0.020977664947509765, 0.021009408950805664, 0.02108518409729004, 0.02109951972961426, 0.021004287719726563, 0.021184511184692383, 0.021045248031616212, 0.02104934310913086, 0.021029888153076173, 0.02109337615966797, 0.02104422378540039, 0.021106687545776368, 0.021090303421020508, 0.02110771179199219, 0.020957183837890626, 0.021020671844482423, 0.02105036735534668, 0.021009408950805664, 0.021111808776855468, 0.02106368064880371, 0.021078016281127928, 0.021146623611450196, 0.02126233673095703, 0.021008384704589843, 0.02109443283081055, 0.021007328033447265, 0.021094400405883788, 0.02101862335205078, 0.020711423873901368, 0.020786176681518553, 0.020919328689575196, 0.020999168395996092, 0.0210861759185791, 0.021045248031616212, 0.021036096572875976, 0.02097657585144043, 0.0210513916015625, 0.06038323211669922, 0.020132863998413086, 0.02003865623474121, 0.020231168746948244, 0.020247552871704103, 0.02026700782775879, 0.02028339195251465, 0.02024345588684082, 0.020335615158081053, 0.02025881576538086, 0.02020966339111328, 0.020247552871704103, 0.020304895401000975, 0.020338687896728515, 0.02026393508911133, 0.02025372886657715, 0.020227039337158203, 0.020325439453125, 0.02019014358520508, 0.020358144760131838, 0.020377599716186523, 0.020360191345214843, 0.020274175643920898, 0.020246528625488282, 0.020170751571655272, 0.020137983322143553, 0.020025344848632814, 0.02031001663208008, 0.020520959854125977, 0.021547008514404296, 0.02227302360534668, 0.021370880126953123, 0.021106687545776368, 0.021145599365234375, 0.02130534362792969, 0.021159936904907226, 0.021429248809814453, 0.0211015682220459, 0.02109132766723633, 0.020986879348754883, 0.02105753517150879, 0.020958208084106447, 0.02103603172302246, 0.020995071411132812, 0.021061632156372072, 0.02103603172302246, 0.020990976333618162, 0.02101043128967285, 0.020993024826049804, 0.020731903076171874, 0.02128486442565918, 0.020934656143188478, 0.02103193664550781, 0.020953088760375976, 0.021168127059936523, 0.02122956848144531, 0.02109542465209961, 0.02106572723388672, 0.021311487197875977, 0.02108518409729004, 0.021142559051513674, 0.021019647598266602, 0.021042144775390625]",tokens/s,46.620011475048464,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen2-beta-72B,Qwen/Qwen2-beta-72B,cuda,0,42,,,True,,,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",qwen2,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-0.5B,,cuda,0,42,,,,,True,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,217063f5c507ed7cc255df7e1f64c4333a0b4dfe,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,,MB,1173.282816,3657.957376,0.0,3017.801728,2552.274944,s,10,0.23174764823913574,0.023174764823913573,0.0002751802978943441,0.023092767715454102,0.02324979190826416,0.023619583988189696,0.02391541765213013,"[0.023989376068115235, 0.023047456741333006, 0.022992639541625975, 0.023107263565063478, 0.023067232131958007, 0.02316761589050293, 0.023083744049072267, 0.023101791381835938, 0.023125696182250976, 0.02306483268737793]",tokens/s,11046.498290064146,kWh,2.7612111456730903e-07,1.5129925178099195e-07,1.846879073556017e-06,2.274299439904318e-06,tokens/kWh,112562134.74280685,MB,1173.610496,3657.957376,0.0,3017.801728,2600.838656,s,10,12.068020141601561,1.206802014160156,0.011724387981231543,1.20719384765625,1.2187086059570311,1.2204795471191405,1.221896300048828,"[1.20141796875, 1.2009669189453125, 1.1865916748046874, 1.203585205078125, 1.1897933349609375, 1.2183150634765625, 1.22225048828125, 1.2181893310546874, 1.216107666015625, 1.210802490234375]",tokens/s,52.204089205007904,kWh,1.4145191280502028e-05,7.751208501315879e-06,3.11326195322475e-05,5.302901931406539e-05,tokens/kWh,1188028.7588741041,,s,629,12.254944248199465,0.019483218200635075,0.002764060016410226,0.019300352096557616,0.01945415725708008,0.01960693778991699,0.04181360717773437,"[0.02071353530883789, 0.018939840316772462, 0.018711551666259766, 0.018320383071899413, 0.018288639068603514, 0.018331647872924805, 0.018336767196655272, 0.018711551666259766, 0.01845964813232422, 0.01845248031616211, 0.018556928634643553, 0.018464767456054687, 0.018449407577514648, 0.018413568496704103, 0.018456575393676757, 0.018527231216430663, 0.018422847747802736, 0.018424768447875977, 0.01841152000427246, 0.01842790412902832, 0.018413568496704103, 0.01841971206665039, 0.01846784019470215, 0.01839308738708496, 0.0183767032623291, 0.01841152000427246, 0.01842278480529785, 0.01840230369567871, 0.01838387107849121, 0.020634624481201173, 0.021939199447631837, 0.019973119735717772, 0.01960140800476074, 0.01960041618347168, 0.019356639862060546, 0.019349504470825195, 0.019326976776123047, 0.019281919479370118, 0.01902284812927246, 0.018982912063598634, 0.01923379135131836, 0.0193832950592041, 0.019417087554931642, 0.019288063049316406, 0.019396608352661132, 0.019351551055908203, 0.019374080657958984, 0.019313663482666017, 0.019297279357910157, 0.01927577590942383, 0.019361791610717775, 0.019019775390625, 0.01929523277282715, 0.019342336654663086, 0.019668991088867188, 0.019302400588989257, 0.019300352096557616, 0.01942937660217285, 0.019324928283691405, 0.01943244743347168, 0.01945497512817383, 0.01942323112487793, 0.04182732772827148, 0.018144287109375, 0.018191328048706056, 0.018397184371948243, 0.01835215950012207, 0.01851593589782715, 0.018397184371948243, 0.018374656677246092, 0.018366464614868162, 0.018358272552490236, 0.018395135879516602, 0.018530303955078126, 0.01944063949584961, 0.01938739204406738, 0.019301376342773437, 0.019369983673095705, 0.019289087295532227, 0.019358783721923827, 0.01931769561767578, 0.019334144592285156, 0.019337215423583985, 0.01902694320678711, 0.01901670455932617, 0.019325952529907226, 0.019357696533203125, 0.01944166374206543, 0.019322879791259767, 0.018921472549438476, 0.018982912063598634, 0.019021823883056642, 0.019099647521972657, 0.018968576431274413, 0.018952192306518553, 0.01905766487121582, 0.019116031646728517, 0.019351551055908203, 0.019296255111694336, 0.0193832950592041, 0.019353599548339845, 0.019334144592285156, 0.019391551971435547, 0.01934329605102539, 0.019326976776123047, 0.019393535614013673, 0.019315744400024416, 0.019332063674926757, 0.01926553535461426, 0.019336191177368164, 0.019273727416992188, 0.019298303604125978, 0.019315711975097655, 0.01908940887451172, 0.018931711196899414, 0.018990079879760743, 0.019216384887695313, 0.019264511108398438, 0.019317760467529296, 0.01924505615234375, 0.019240991592407226, 0.019043296813964845, 0.018979839324951172, 0.018913280487060546, 0.018921472549438476, 0.041724929809570314, 0.018271263122558595, 0.01833366394042969, 0.018336767196655272, 0.018334720611572267, 0.018320383071899413, 0.01834809684753418, 0.01832134437561035, 0.01840540885925293, 0.018377695083618164, 0.018342912673950194, 0.018239488601684572, 0.018125823974609375, 0.018081792831420897, 0.018300928115844727, 0.018345983505249023, 0.01830297660827637, 0.01830297660827637, 0.01829478454589844, 0.018309120178222657, 0.018336767196655272, 0.018332672119140626, 0.01825484848022461, 0.018326528549194337, 0.018325504302978517, 0.018341888427734376, 0.018341888427734376, 0.018367488861083983, 0.01838591957092285, 0.018301952362060548, 0.01827840042114258, 0.018307071685791015, 0.019517440795898438, 0.02072985649108887, 0.019341312408447265, 0.01903615951538086, 0.019357696533203125, 0.019347455978393553, 0.01938643264770508, 0.019445695877075196, 0.019332096099853514, 0.019321855545043946, 0.019330047607421876, 0.019397632598876953, 0.019042303085327148, 0.018965503692626954, 0.01922764778137207, 0.019310592651367187, 0.019344383239746094, 0.01957683181762695, 0.019343360900878907, 0.019353599548339845, 0.019294208526611328, 0.019369983673095705, 0.019293184280395507, 0.019310592651367187, 0.019318784713745117, 0.019336191177368164, 0.019333120346069335, 0.01927065658569336, 0.019385343551635743, 0.019298303604125978, 0.019364864349365234, 0.04183244705200195, 0.018319360733032225, 0.018343936920166014, 0.018296831130981444, 0.01838591957092285, 0.018382848739624022, 0.018375680923461913, 0.018347007751464844, 0.018283519744873047, 0.018347007751464844, 0.018334720611572267, 0.018386943817138672, 0.018387968063354493, 0.018362367630004883, 0.0183767032623291, 0.018971647262573242, 0.019322879791259767, 0.019343360900878907, 0.01926348876953125, 0.019289087295532227, 0.019365888595581054, 0.019861503601074217, 0.019386367797851564, 0.01945907211303711, 0.019352575302124024, 0.019417087554931642, 0.019331071853637697, 0.019171327590942384, 0.018885631561279297, 0.01919385528564453, 0.019308544158935546, 0.019360767364501954, 0.01927884864807129, 0.0192993278503418, 0.019351551055908203, 0.019353599548339845, 0.01944576072692871, 0.019340288162231444, 0.019357759475708006, 0.01931564712524414, 0.019315711975097655, 0.019341312408447265, 0.019348480224609374, 0.019322879791259767, 0.01941094398498535, 0.019350528717041016, 0.01943654441833496, 0.019381248474121093, 0.019330047607421876, 0.019385343551635743, 0.019297279357910157, 0.01929113578796387, 0.019203071594238282, 0.019316736221313476, 0.019310592651367187, 0.019369983673095705, 0.019331104278564454, 0.01930441665649414, 0.019298303604125978, 0.01925836753845215, 0.01939967918395996, 0.01942527961730957, 0.019507200241088866, 0.041815040588378906, 0.018103296279907227, 0.018100223541259765, 0.018309120178222657, 0.018489343643188477, 0.018325504302978517, 0.018365472793579102, 0.018322399139404297, 0.01843302345275879, 0.01842278480529785, 0.018367488861083983, 0.01838591957092285, 0.018344959259033202, 0.018298879623413086, 0.01829478454589844, 0.01827846336364746, 0.018314176559448243, 0.018298879623413086, 0.018311168670654295, 0.018297855377197265, 0.018280511856079103, 0.018277311325073244, 0.018266111373901366, 0.018351104736328124, 0.01841459274291992, 0.018293760299682618, 0.01827840042114258, 0.01835212707519531, 0.0182794246673584, 0.018366464614868162, 0.01922559928894043, 0.0192542724609375, 0.019371007919311522, 0.019348480224609374, 0.019318784713745117, 0.019368959426879884, 0.019286016464233398, 0.019310592651367187, 0.019292160034179686, 0.019350528717041016, 0.019319807052612305, 0.01926246452331543, 0.019341375350952147, 0.019480512619018554, 0.019310592651367187, 0.01942937660217285, 0.019377151489257814, 0.019343360900878907, 0.019309568405151366, 0.019318784713745117, 0.01940275192260742, 0.019413055419921874, 0.01936684799194336, 0.019350591659545897, 0.019314624786376952, 0.019364864349365234, 0.019412992477416992, 0.019362815856933592, 0.019286016464233398, 0.019309568405151366, 0.019581951141357423, 0.02045030403137207, 0.019610624313354492, 0.04297011184692383, 0.01923891258239746, 0.01920921516418457, 0.019215360641479492, 0.01920921516418457, 0.01925836753845215, 0.01927577590942383, 0.01927168083190918, 0.019482624053955077, 0.01924710464477539, 0.019173376083374022, 0.0192491512298584, 0.019224576950073242, 0.01925939178466797, 0.01920614433288574, 0.01918976020812988, 0.018976768493652343, 0.018905088424682616, 0.019168256759643554, 0.019224576950073242, 0.019179519653320314, 0.019297279357910157, 0.01920102310180664, 0.01923788833618164, 0.019152896881103516, 0.019204095840454103, 0.019298303604125978, 0.019559423446655275, 0.019380224227905272, 0.01986457633972168, 0.019556352615356445, 0.02034079933166504, 0.019587007522583008, 0.019335168838500977, 0.019269632339477538, 0.019361791610717775, 0.019335168838500977, 0.019768320083618163, 0.019483648300170898, 0.019281919479370118, 0.019347455978393553, 0.019194879531860352, 0.019277887344360353, 0.01954911994934082, 0.019361791610717775, 0.01927065658569336, 0.01920102310180664, 0.0192675838470459, 0.019388416290283202, 0.019392511367797852, 0.01963417625427246, 0.0192491512298584, 0.019361791610717775, 0.019302400588989257, 0.01926144027709961, 0.01923174476623535, 0.019516416549682617, 0.01928704071044922, 0.019319807052612305, 0.01944883155822754, 0.019327999114990235, 0.019368000030517578, 0.019319744110107423, 0.04284108734130859, 0.019286016464233398, 0.019317760467529296, 0.019334144592285156, 0.01938739204406738, 0.018997247695922852, 0.019506175994873046, 0.019342336654663086, 0.01928294372558594, 0.019284992218017577, 0.019375104904174805, 0.01927577590942383, 0.019333120346069335, 0.01929113578796387, 0.019361791610717775, 0.0194201602935791, 0.019412992477416992, 0.019301376342773437, 0.019374080657958984, 0.019276800155639647, 0.019232799530029297, 0.01934435272216797, 0.019346431732177736, 0.019744768142700195, 0.01966182327270508, 0.019677183151245118, 0.021542911529541017, 0.019766271591186522, 0.01946735954284668, 0.01937603187561035, 0.01924505615234375, 0.019345407485961915, 0.019418111801147463, 0.019422208786010742, 0.019377151489257814, 0.01923686408996582, 0.01977446365356445, 0.019528703689575197, 0.019296255111694336, 0.019409919738769533, 0.019361791610717775, 0.019332096099853514, 0.019377151489257814, 0.01927168083190918, 0.01925017547607422, 0.019355648040771483, 0.01939148712158203, 0.019326976776123047, 0.019323904037475585, 0.01926246452331543, 0.019354623794555666, 0.019350528717041016, 0.01926144027709961, 0.019357696533203125, 0.019288063049316406, 0.019336191177368164, 0.019297279357910157, 0.019293184280395507, 0.019311616897583008, 0.019302400588989257, 0.019348480224609374, 0.019311616897583008, 0.01907302474975586, 0.04186316680908203, 0.01827840042114258, 0.01919385528564453, 0.01945599937438965, 0.01939148712158203, 0.019258399963378907, 0.019320831298828126, 0.01925321578979492, 0.01939455986022949, 0.019334144592285156, 0.019340288162231444, 0.019361791610717775, 0.01942732810974121, 0.019332096099853514, 0.019323936462402345, 0.019355615615844726, 0.01943244743347168, 0.019165184020996092, 0.01903206443786621, 0.018937856674194335, 0.01941094398498535, 0.019347455978393553, 0.019371007919311522, 0.019346431732177736, 0.019722272872924804, 0.019405792236328125, 0.019382272720336914, 0.019389440536499023, 0.019397632598876953, 0.019340288162231444, 0.019418111801147463, 0.019316736221313476, 0.019363840103149413, 0.019317792892456054, 0.01946006393432617, 0.019377151489257814, 0.019382335662841796, 0.0193822078704834, 0.019408895492553712, 0.019462207794189453, 0.019428287506103516, 0.0194150390625, 0.019385343551635743, 0.019409919738769533, 0.019365888595581054, 0.019337215423583985, 0.01997209548950195, 0.019385343551635743, 0.01926655960083008, 0.019374080657958984, 0.019345407485961915, 0.01924300765991211, 0.01925119972229004, 0.019352575302124024, 0.019353599548339845, 0.019351551055908203, 0.019308544158935546, 0.019557376861572266, 0.019333120346069335, 0.019325952529907226, 0.01945395278930664, 0.0192491512298584, 0.01941606330871582, 0.042439678192138675, 0.01924403190612793, 0.01929523277282715, 0.019312639236450196, 0.019281919479370118, 0.0192542724609375, 0.01923276710510254, 0.019354623794555666, 0.019322879791259767, 0.019300352096557616, 0.019359743118286133, 0.019689472198486328, 0.01945599937438965, 0.01928294372558594, 0.019298303604125978, 0.01926553535461426, 0.01927065658569336, 0.019276800155639647, 0.019347455978393553, 0.01923686408996582, 0.01926553535461426, 0.019277824401855468, 0.01925734329223633, 0.019283967971801756, 0.019293184280395507, 0.01942937660217285, 0.019284992218017577, 0.019292160034179686, 0.01923276710510254, 0.01925529670715332, 0.019273727416992188, 0.019345407485961915, 0.01922047996520996, 0.019306495666503908, 0.019297279357910157, 0.019333120346069335, 0.019298303604125978, 0.01947750473022461, 0.01926144027709961, 0.0194201602935791, 0.01926246452331543, 0.019483648300170898, 0.019372032165527343, 0.01939558410644531, 0.019316736221313476, 0.01927475166320801, 0.0192225284576416, 0.019304447174072266, 0.019285055160522462, 0.019469247817993165, 0.019348480224609374, 0.01966796875, 0.019400703430175782, 0.019303455352783203, 0.019248096466064454, 0.01896044731140137, 0.01892243194580078, 0.019120128631591796, 0.01939455986022949, 0.019363840103149413, 0.019300352096557616, 0.018927616119384767, 0.019292160034179686, 0.041809921264648435, 0.01825279998779297, 0.018301952362060548, 0.018735103607177735, 0.01942118453979492, 0.019508224487304687, 0.019348480224609374, 0.01898700714111328, 0.018920448303222655, 0.019286016464233398, 0.01942425537109375, 0.019322879791259767, 0.019021823883056642, 0.01922355270385742, 0.019260416030883788, 0.01921023941040039, 0.019280960083007812, 0.019331008911132812, 0.01925324821472168, 0.01932598304748535, 0.019331039428710936, 0.01927475166320801, 0.01929523277282715, 0.01921331214904785, 0.019261472702026366, 0.019305440902709962, 0.019301376342773437, 0.019513343811035155, 0.019346431732177736, 0.019397632598876953, 0.019386367797851564, 0.019357696533203125, 0.019310592651367187, 0.019308544158935546, 0.019294208526611328, 0.019280895233154297, 0.01925632095336914, 0.019389440536499023, 0.01942527961730957, 0.01928704071044922, 0.019327999114990235, 0.019321855545043946, 0.019545087814331053, 0.019123199462890626, 0.018877439498901367, 0.01926144027709961, 0.01927065658569336, 0.019338239669799806, 0.01927475166320801, 0.01897881507873535, 0.01899212837219238, 0.019589120864868165, 0.019286016464233398, 0.01926655960083008, 0.01925836753845215, 0.019184640884399414, 0.01924403190612793, 0.01924095916748047, 0.01918156814575195, 0.018952192306518553, 0.019113983154296875, 0.01924300765991211, 0.01927168083190918]",tokens/s,51.3262228910111,,,main,False,False,,,,,,,,,,,,,,,,,,,,,,, float32-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-1.8B,,cuda,0,42,,,,,True,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,217063f5c507ed7cc255df7e1f64c4333a0b4dfe,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,,MB,1175.990272,9272.03328,0.0,8631.877632,8498.684928,s,10,0.5260754814147949,0.0526075481414795,0.0002627683235581081,0.05261451148986816,0.053006309509277344,0.05303499412536621,0.05305794181823731,"[0.052999935150146484, 0.052679134368896485, 0.05233830261230469, 0.05238092803955078, 0.05229375839233399, 0.05262063980102539, 0.05260838317871094, 0.0527611198425293, 0.052329601287841795, 0.05306367874145508]",tokens/s,4866.221845419014,kWh,6.546623378241992e-07,3.5871722440627497e-07,4.525012828110586e-06,5.53839239034106e-06,tokens/kWh,46222799.31744512,MB,1175.990272,9272.03328,0.0,8631.877632,8597.292544,s,10,12.151699951171874,1.2151699951171875,0.008398459603249456,1.213122802734375,1.226181201171875,1.2296250854492188,1.2323801928710938,"[1.2254158935546875, 1.214334716796875, 1.2065869140625, 1.2061162109375, 1.206941162109375, 1.2103851318359375, 1.211910888671875, 1.217904052734375, 1.2190360107421876, 1.2330689697265624]",tokens/s,51.844598083517084,kWh,1.4507108161590668e-05,7.949670775071141e-06,5.187353229069146e-05,7.433031122735327e-05,tokens/kWh,847568.0911291038,,s,629,12.604601356506356,0.020039111854541092,0.006308930534271433,0.019177471160888672,0.01962741775512695,0.02015703048706055,0.07194259490966798,"[0.01963315200805664, 0.01924198341369629, 0.019166208267211913, 0.019269632339477538, 0.01925632095336914, 0.019409919738769533, 0.019563520431518554, 0.019211263656616212, 0.019180543899536134, 0.019219455718994142, 0.019316736221313476, 0.0191016960144043, 0.01944166374206543, 0.019489791870117186, 0.01965977668762207, 0.01962598419189453, 0.01945395278930664, 0.019330047607421876, 0.01937919998168945, 0.01926348876953125, 0.019582975387573243, 0.019396608352661132, 0.01921843147277832, 0.019194879531860352, 0.019138559341430664, 0.01924608039855957, 0.02002943992614746, 0.023557119369506836, 0.021587968826293946, 0.020083711624145507, 0.019610624313354492, 0.01926655960083008, 0.019277824401855468, 0.019773439407348634, 0.02042265510559082, 0.019486719131469727, 0.01920102310180664, 0.019134464263916014, 0.019329023361206055, 0.01919795227050781, 0.019195903778076173, 0.019137535095214844, 0.01912118339538574, 0.019141599655151366, 0.01923788833618164, 0.019132415771484376, 0.019195968627929688, 0.019203008651733397, 0.019166208267211913, 0.01925939178466797, 0.019150848388671874, 0.019163135528564454, 0.01923481559753418, 0.01901875114440918, 0.019138559341430664, 0.019421247482299803, 0.019311552047729493, 0.019100671768188478, 0.01914681625366211, 0.019149759292602538, 0.019158016204833983, 0.01920614433288574, 0.0725032958984375, 0.019342336654663086, 0.019745792388916016, 0.019509248733520508, 0.019284992218017577, 0.01926655960083008, 0.019349504470825195, 0.01925119972229004, 0.019513343811035155, 0.019458112716674806, 0.019256256103515626, 0.01942937660217285, 0.019558399200439454, 0.019612672805786133, 0.019710975646972655, 0.019324928283691405, 0.019334144592285156, 0.019574783325195313, 0.0196177921295166, 0.019340288162231444, 0.019776512145996093, 0.019558399200439454, 0.020486143112182616, 0.019766271591186522, 0.019203071594238282, 0.019134464263916014, 0.01906892776489258, 0.019062784194946288, 0.019207168579101562, 0.01899929618835449, 0.019122175216674805, 0.019135488510131835, 0.019127296447753905, 0.019222591400146486, 0.01914361572265625, 0.01904640007019043, 0.01907302474975586, 0.019170303344726563, 0.019078144073486326, 0.01904844856262207, 0.01925324821472168, 0.019083263397216797, 0.01918156814575195, 0.01925119972229004, 0.019006464004516603, 0.018938880920410156, 0.019167232513427734, 0.019105791091918945, 0.019090431213378906, 0.019184640884399414, 0.019133504867553712, 0.019068864822387694, 0.019092479705810548, 0.019091455459594727, 0.019137535095214844, 0.019108928680419923, 0.019029951095581053, 0.019103744506835937, 0.019159040451049804, 0.019123199462890626, 0.0191345272064209, 0.0191077766418457, 0.019082239151000976, 0.07184896087646485, 0.018969600677490234, 0.019158016204833983, 0.019150848388671874, 0.019013631820678712, 0.018913280487060546, 0.019106815338134766, 0.01912937545776367, 0.01914057540893555, 0.0191016960144043, 0.019163135528564454, 0.019566591262817384, 0.019427391052246095, 0.019072959899902344, 0.019111936569213867, 0.01922355270385742, 0.019133440017700197, 0.019131391525268555, 0.019471359252929688, 0.019162111282348633, 0.019190784454345702, 0.019430400848388672, 0.019136512756347656, 0.019183616638183593, 0.019163135528564454, 0.019147775650024415, 0.019153919219970703, 0.019158016204833983, 0.01911299133300781, 0.01912931251525879, 0.01921843147277832, 0.019187711715698243, 0.019128320693969726, 0.019148799896240236, 0.01921331214904785, 0.019119104385375976, 0.019185663223266602, 0.019144704818725586, 0.019095552444458007, 0.019212287902832033, 0.019178495407104493, 0.019083263397216797, 0.019096576690673828, 0.019078144073486326, 0.019111936569213867, 0.019172351837158205, 0.019094528198242186, 0.019107839584350587, 0.019079168319702147, 0.019088384628295898, 0.019123199462890626, 0.01926246452331543, 0.01906380844116211, 0.0191016960144043, 0.019070976257324217, 0.019056640625, 0.019108863830566408, 0.01927987289428711, 0.01905561637878418, 0.019079200744628905, 0.01906377601623535, 0.01904844856262207, 0.019098623275756836, 0.07197183990478516, 0.01904844856262207, 0.01909350395202637, 0.01906790351867676, 0.019059776306152344, 0.019446720123291017, 0.019116031646728517, 0.01903411293029785, 0.019103744506835937, 0.019136512756347656, 0.01908531188964844, 0.019132415771484376, 0.019159040451049804, 0.019104799270629882, 0.0191231689453125, 0.019107839584350587, 0.019165184020996092, 0.019153919219970703, 0.019092479705810548, 0.019087360382080077, 0.019161088943481445, 0.019091455459594727, 0.01906483268737793, 0.019130367279052735, 0.019124223709106446, 0.019110912322998046, 0.019112960815429687, 0.01905356788635254, 0.019122175216674805, 0.019147775650024415, 0.019102720260620116, 0.019144704818725586, 0.019160064697265625, 0.019113983154296875, 0.019139583587646485, 0.019104768753051758, 0.019102720260620116, 0.019187711715698243, 0.019537984848022463, 0.019148735046386717, 0.019211263656616212, 0.019165184020996092, 0.019149824142456053, 0.01909760093688965, 0.019130367279052735, 0.019118080139160155, 0.019165184020996092, 0.019105791091918945, 0.019192832946777344, 0.01923072052001953, 0.019148799896240236, 0.019110912322998046, 0.019141632080078123, 0.019150848388671874, 0.019137535095214844, 0.019149824142456053, 0.019159040451049804, 0.01908121681213379, 0.019169343948364257, 0.01922553634643555, 0.019090463638305664, 0.019134431838989257, 0.019112960815429687, 0.07186739349365234, 0.019123199462890626, 0.019142656326293944, 0.019130367279052735, 0.01920512008666992, 0.019145727157592773, 0.019124223709106446, 0.019139583587646485, 0.019075071334838867, 0.019109888076782225, 0.019147775650024415, 0.019146751403808594, 0.019162111282348633, 0.019141632080078123, 0.019756032943725587, 0.019247167587280272, 0.01894393539428711, 0.018899967193603515, 0.018892799377441406, 0.018883583068847656, 0.019104768753051758, 0.019087360382080077, 0.019140607833862306, 0.019090463638305664, 0.0191866569519043, 0.019126272201538085, 0.019182592391967773, 0.019160064697265625, 0.019211263656616212, 0.019311616897583008, 0.019565568923950196, 0.019174400329589843, 0.019152896881103516, 0.019135488510131835, 0.019156991958618166, 0.019140607833862306, 0.019183616638183593, 0.019127296447753905, 0.019143680572509765, 0.019165184020996092, 0.019126272201538085, 0.01923072052001953, 0.019164159774780275, 0.019173376083374022, 0.019123199462890626, 0.019138559341430664, 0.019148799896240236, 0.019155967712402345, 0.019216384887695313, 0.019134464263916014, 0.019139583587646485, 0.019124223709106446, 0.019116031646728517, 0.019141632080078123, 0.01919692802429199, 0.019144704818725586, 0.019158016204833983, 0.019155040740966797, 0.019119007110595703, 0.01918976020812988, 0.019145727157592773, 0.019070976257324217, 0.019120128631591796, 0.07306240081787109, 0.019148799896240236, 0.019194879531860352, 0.019175424575805664, 0.019167232513427734, 0.019169279098510742, 0.019141632080078123, 0.019115007400512696, 0.019152896881103516, 0.01924300765991211, 0.019139583587646485, 0.019135488510131835, 0.019277824401855468, 0.019130399703979492, 0.0191364803314209, 0.019147775650024415, 0.019152896881103516, 0.019124223709106446, 0.01923891258239746, 0.01926348876953125, 0.019174400329589843, 0.019100671768188478, 0.019106815338134766, 0.019130367279052735, 0.01920102310180664, 0.01910688018798828, 0.01914054489135742, 0.01922150421142578, 0.01924608039855957, 0.01920102310180664, 0.019176448822021484, 0.01918156814575195, 0.0192194881439209, 0.019194847106933595, 0.019120128631591796, 0.01964543914794922, 0.019177471160888672, 0.019170368194580078, 0.01915897560119629, 0.019125247955322267, 0.019111936569213867, 0.019175424575805664, 0.019140607833862306, 0.019167232513427734, 0.019088384628295898, 0.019171327590942384, 0.019297279357910157, 0.019110912322998046, 0.019132415771484376, 0.019182592391967773, 0.019121152877807617, 0.019203071594238282, 0.019317760467529296, 0.019169376373291015, 0.01918659210205078, 0.019180543899536134, 0.01967001533508301, 0.019216384887695313, 0.01924095916748047, 0.019311616897583008, 0.019350528717041016, 0.019168256759643554, 0.019188735961914064, 0.07226163482666016, 0.019095552444458007, 0.0192225284576416, 0.019208192825317383, 0.019143680572509765, 0.019175424575805664, 0.019224576950073242, 0.019110912322998046, 0.019136512756347656, 0.01923583984375, 0.019131391525268555, 0.019200000762939453, 0.01920518493652344, 0.01913542366027832, 0.019129344940185547, 0.019141632080078123, 0.01923891258239746, 0.019162111282348633, 0.019208192825317383, 0.019113983154296875, 0.019120128631591796, 0.019158016204833983, 0.019170303344726563, 0.019102720260620116, 0.01926553535461426, 0.019116031646728517, 0.019123199462890626, 0.019178495407104493, 0.01911404800415039, 0.01957164764404297, 0.019147775650024415, 0.019171327590942384, 0.019244096755981446, 0.019138496398925783, 0.019166208267211913, 0.019103744506835937, 0.019173376083374022, 0.019116031646728517, 0.01932806396484375, 0.01916102409362793, 0.019163135528564454, 0.01922150421142578, 0.019159040451049804, 0.018988031387329102, 0.019722240447998047, 0.020699136734008788, 0.019927040100097656, 0.01923072052001953, 0.019185663223266602, 0.019136512756347656, 0.01926563262939453, 0.019165088653564453, 0.01908531188964844, 0.01901875114440918, 0.019212287902832033, 0.019283967971801756, 0.01902284812927246, 0.019547136306762695, 0.019326976776123047, 0.019296255111694336, 0.019298303604125978, 0.019212287902832033, 0.01923276710510254, 0.0725401611328125, 0.019340288162231444, 0.019264511108398438, 0.019264511108398438, 0.01923583984375, 0.019244096755981446, 0.019226560592651366, 0.01926044845581055, 0.01947337532043457, 0.019610687255859374, 0.019303359985351563, 0.019207168579101562, 0.019316736221313476, 0.01925836753845215, 0.019352575302124024, 0.019194879531860352, 0.019298303604125978, 0.01924608039855957, 0.01925632095336914, 0.019168256759643554, 0.019177471160888672, 0.01922047996520996, 0.019161088943481445, 0.019173376083374022, 0.019993600845336915, 0.019537919998168944, 0.01946009635925293, 0.01949286460876465, 0.019193920135498047, 0.019314624786376952, 0.019211263656616212, 0.01940787124633789, 0.019467264175415038, 0.019252288818359376, 0.019223487854003907, 0.019207168579101562, 0.01923788833618164, 0.019211263656616212, 0.01924608039855957, 0.019179519653320314, 0.019200000762939453, 0.019167232513427734, 0.01926246452331543, 0.019185663223266602, 0.01902796745300293, 0.01903001594543457, 0.019248159408569335, 0.018996192932128907, 0.019105791091918945, 0.01927993583679199, 0.019248064041137696, 0.01923788833618164, 0.0192225284576416, 0.01903411293029785, 0.019994623184204103, 0.021124095916748048, 0.01987993621826172, 0.019523647308349608, 0.01950611114501953, 0.01925529670715332, 0.01926655960083008, 0.01925324821472168, 0.01922355270385742, 0.07199846649169922, 0.019154943466186524, 0.01923481559753418, 0.01925836753845215, 0.01922355270385742, 0.019203071594238282, 0.01922047996520996, 0.019172351837158205, 0.019140607833862306, 0.01920921516418457, 0.019182592391967773, 0.01925017547607422, 0.019136512756347656, 0.019183616638183593, 0.01916111946105957, 0.01927574348449707, 0.019316736221313476, 0.019363840103149413, 0.01922355270385742, 0.019177471160888672, 0.018966527938842775, 0.01919385528564453, 0.019710975646972655, 0.020303871154785155, 0.0194703369140625, 0.019502111434936523, 0.01950204849243164, 0.019491840362548828, 0.019547136306762695, 0.020298751831054687, 0.019386367797851564, 0.01920102310180664, 0.019178495407104493, 0.019176448822021484, 0.01922764778137207, 0.019155967712402345, 0.01918976020812988, 0.019149824142456053, 0.01904537582397461, 0.019123199462890626, 0.01919795227050781, 0.01902796745300293, 0.019187711715698243, 0.01918156814575195, 0.019308544158935546, 0.019474431991577147, 0.019483648300170898, 0.01923276710510254, 0.019320831298828126, 0.01920921516418457, 0.01942118453979492, 0.01940377616882324, 0.019475456237792968, 0.019429407119750976, 0.019400672912597658, 0.019479551315307618, 0.019552255630493166, 0.019283967971801756, 0.019216415405273437, 0.019967967987060548, 0.01963929557800293, 0.019543039321899415, 0.01993011283874512, 0.07285350036621094, 0.019086336135864256, 0.019083263397216797, 0.019111936569213867, 0.019337215423583985, 0.019299392700195313, 0.019156927108764647, 0.019108863830566408, 0.019182592391967773, 0.019107839584350587, 0.019148799896240236, 0.019185663223266602, 0.019147775650024415, 0.019150848388671874, 0.019140607833862306, 0.019156991958618166, 0.019804159164428712, 0.020758527755737305, 0.019825664520263672, 0.019070976257324217, 0.01919795227050781, 0.018947071075439453, 0.01920921516418457, 0.01922355270385742, 0.019139583587646485, 0.019160064697265625, 0.019166208267211913, 0.019126272201538085, 0.01926348876953125, 0.019743743896484374, 0.019358783721923827, 0.019277759552001953, 0.019212287902832033, 0.01923276710510254, 0.019294208526611328, 0.01908121681213379, 0.01945907211303711, 0.020368383407592772, 0.020304895401000975, 0.020364288330078126, 0.020291584014892578, 0.02040934371948242, 0.02032537651062012, 0.02021683120727539, 0.020247552871704103, 0.020221952438354493, 0.019969024658203126, 0.019164159774780275, 0.020377599716186523, 0.020360191345214843, 0.02026700782775879, 0.02031820869445801, 0.02003558349609375, 0.02001817512512207, 0.02004582405090332, 0.02020249557495117, 0.02008883285522461, 0.01991372871398926, 0.019971071243286134, 0.018980863571166993, 0.01902284812927246, 0.01902284812927246, 0.018976768493652343]",tokens/s,49.902411207579966,,,main,False,False,,,,,,,,,,,,,,,,,,,,,,, float32-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-70m,EleutherAI/pythia-70m,cuda,0,42,,,True,True,True,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.42.1,,0.31.0,,,,1.20.0,,,,0.11.1,,,,MB,1133.932544,1057.488896,0.0,417.333248,391.122432,s,37,0.19297657632827758,0.005215583144007502,0.0002458087771285964,0.005141056060791015,0.005336576080322265,0.005714982414245604,0.006154892845153808,"[0.0060210561752319335, 0.00521727991104126, 0.006230175971984863, 0.005278528213500977, 0.005129407882690429, 0.005105184078216553, 0.005638463973999024, 0.0051446080207824705, 0.005360095977783203, 0.00532089614868164, 0.005159488201141357, 0.005196991920471191, 0.00501043176651001, 0.005087679862976075, 0.005113952159881592, 0.005162655830383301, 0.005273568153381348, 0.005103263854980469, 0.005108384132385254, 0.005093823909759521, 0.005071104049682617, 0.00527235221862793, 0.005074207782745361, 0.005067455768585205, 0.005191167831420898, 0.005079808235168457, 0.005057536125183106, 0.005187647819519043, 0.005092895984649658, 0.005141056060791015, 0.005125216007232666, 0.005157792091369629, 0.005259967803955078, 0.005194176197052002, 0.005051167964935302, 0.005119264125823975, 0.005077824115753173]",tokens/s,49083.677305410005,kWh,6.09346155302334e-08,3.338912513584502e-08,2.3584455858720373e-07,3.3016829925328214e-07,tokens/kWh,775362142.819213,MB,1133.932544,1057.488896,0.0,417.333248,391.124992,s,37,10.039309570312497,0.27133269108952707,0.0032225905494752964,0.26986520385742185,0.2765611999511719,0.27892122802734376,0.27928929443359374,"[0.2789089965820313, 0.27462530517578126, 0.2721040954589844, 0.2695911865234375, 0.2690017395019531, 0.2763359069824219, 0.27269937133789063, 0.2714544677734375, 0.27897015380859375, 0.2749640502929688, 0.27946881103515625, 0.26859445190429687, 0.2702781677246094, 0.2685722351074219, 0.2690583801269531, 0.2702041015625, 0.2699896240234375, 0.26947454833984374, 0.2695362243652344, 0.2696147766113281, 0.26986520385742185, 0.2686041564941406, 0.26915859985351565, 0.27088729858398436, 0.2687685852050781, 0.268348876953125, 0.2715785217285156, 0.2704392700195313, 0.2688677673339844, 0.2687818603515625, 0.2717867736816406, 0.2751473388671875, 0.2696357421875, 0.2692345275878906, 0.26862261962890627, 0.2692366943359375, 0.2768991394042969]",tokens/s,232.1872817721509,kWh,3.154992274549448e-06,1.728773545775669e-06,6.309941458818119e-06,1.1193707279143237e-05,tokens/kWh,5628162.183353253,,s,2330,10.22227967023854,0.00438724449366459,0.0006837052929359699,0.004271103858947754,0.0044329953193664555,0.004488192081451416,0.009577719783782959,"[0.0049827837944030765, 0.004978687763214112, 0.004911104202270508, 0.004926464080810547, 0.004956160068511963, 0.004955135822296143, 0.004894752025604248, 0.004881375789642334, 0.004857855796813965, 0.004827136039733886, 0.005182464122772217, 0.004931583881378174, 0.004383743762969971, 0.004329472064971924, 0.004316160202026367, 0.00431001615524292, 0.004314112186431885, 0.004290559768676758, 0.004287487983703613, 0.004345856189727783, 0.004298751831054688, 0.004290559768676758, 0.004317183971405029, 0.00430185604095459, 0.004338655948638916, 0.00431820821762085, 0.004327424049377441, 0.0045352959632873535, 0.004296703815460205, 0.004289535999298095, 0.00429260778427124, 0.004312064170837402, 0.004328447818756104, 0.004343808174133301, 0.004291584014892578, 0.004324351787567139, 0.004275199890136719, 0.004291584014892578, 0.004255743980407714, 0.004277311801910401, 0.0042659201622009275, 0.004253695964813233, 0.004245503902435303, 0.0042475519180297855, 0.004248576164245605, 0.004269055843353272, 0.004304895877838135, 0.004262911796569824, 0.004262911796569824, 0.0042854399681091304, 0.004254720211029053, 0.004249631881713867, 0.004254687786102295, 0.004265984058380127, 0.004245503902435303, 0.004257791996002197, 0.004248576164245605, 0.0042444801330566405, 0.004249599933624268, 0.004248576164245605, 0.00425164794921875, 0.004270080089569092, 0.009809920310974121, 0.004305920124053955, 0.00430079984664917, 0.004291584014892578, 0.004291584014892578, 0.004267007827758789, 0.004269055843353272, 0.004284416198730469, 0.004279295921325684, 0.004263936042785644, 0.004276224136352539, 0.004253695964813233, 0.004249599933624268, 0.004373504161834717, 0.004385791778564453, 0.00434278392791748, 0.004314112186431885, 0.004324351787567139, 0.00434278392791748, 0.004280320167541504, 0.004334591865539551, 0.004332543849945069, 0.00429260778427124, 0.004299776077270508, 0.004290559768676758, 0.004291584014892578, 0.0043038721084594726, 0.004261888027191162, 0.004264959812164307, 0.004253695964813233, 0.004257791996002197, 0.004269055843353272, 0.004883456230163574, 0.004523007869720459, 0.004361248016357422, 0.004394976139068604, 0.0043407359123229985, 0.0043407359123229985, 0.004255743980407714, 0.004247583866119385, 0.00429257583618164, 0.004253695964813233, 0.0042475519180297855, 0.004233215808868408, 0.004232255935668945, 0.004243391990661621, 0.004323328018188476, 0.004248576164245605, 0.004245503902435303, 0.004245503902435303, 0.004233215808868408, 0.004242432117462158, 0.004250624179840088, 0.004284416198730469, 0.004255743980407714, 0.004298751831054688, 0.004245503902435303, 0.0042434558868408205, 0.004260863780975342, 0.004700160026550293, 0.005328896045684814, 0.0052633600234985355, 0.005254144191741943, 0.011900927543640137, 0.0047923197746276855, 0.004575232028961182, 0.004488192081451416, 0.004382719993591308, 0.0043786239624023435, 0.005092351913452148, 0.004550655841827392, 0.004369408130645752, 0.004297728061676025, 0.004259840011596679, 0.00425164794921875, 0.004240447998046875, 0.004255680084228516, 0.00425164794921875, 0.004379648208618164, 0.0043079681396484375, 0.004290559768676758, 0.00429260778427124, 0.004262911796569824, 0.004276288032531738, 0.004257728099822998, 0.004270080089569092, 0.004269055843353272, 0.0042434558868408205, 0.0042434558868408205, 0.00425267219543457, 0.004237311840057373, 0.004248576164245605, 0.004254720211029053, 0.004241407871246338, 0.004258815765380859, 0.004245503902435303, 0.004258815765380859, 0.004249599933624268, 0.004248576164245605, 0.004246528148651123, 0.004298751831054688, 0.004248576164245605, 0.004250688076019287, 0.00425977611541748, 0.004255743980407714, 0.004239359855651856, 0.004237311840057373, 0.004269055843353272, 0.004253695964813233, 0.004249599933624268, 0.004272128105163574, 0.004248576164245605, 0.004228096008300781, 0.004249599933624268, 0.004232192039489746, 0.004239359855651856, 0.004235263824462891, 0.004248576164245605, 0.0042444801330566405, 0.004283391952514649, 0.0042444801330566405, 0.004248576164245605, 0.004262944221496582, 0.004237279891967773, 0.004263936042785644, 0.004270080089569092, 0.009979904174804688, 0.0043079681396484375, 0.00429260778427124, 0.004295680046081543, 0.004291615962982178, 0.004294623851776123, 0.004288512229919434, 0.004281343936920166, 0.004284416198730469, 0.004289535999298095, 0.004284416198730469, 0.004289535999298095, 0.004291584014892578, 0.004327424049377441, 0.0042741761207580565, 0.004298751831054688, 0.00429260778427124, 0.004346879959106445, 0.004295680046081543, 0.004295680046081543, 0.004290559768676758, 0.0043069438934326175, 0.0043069438934326175, 0.004297728061676025, 0.00430182409286499, 0.004253695964813233, 0.004268032073974609, 0.0042854399681091304, 0.004273151874542237, 0.004264959812164307, 0.00425267219543457, 0.004253695964813233, 0.004246528148651123, 0.004313087940216064, 0.004261888027191162, 0.0042639999389648435, 0.00431609582901001, 0.004256768226623535, 0.004239359855651856, 0.004258815765380859, 0.004261888027191162, 0.0042475519180297855, 0.004276224136352539, 0.004241407871246338, 0.004255743980407714, 0.00425267219543457, 0.004242432117462158, 0.004253695964813233, 0.004238336086273193, 0.004257791996002197, 0.0042475519180297855, 0.004245503902435303, 0.004257791996002197, 0.004253695964813233, 0.0042434558868408205, 0.004263936042785644, 0.004250624179840088, 0.004259840011596679, 0.004270080089569092, 0.004239359855651856, 0.004258815765380859, 0.004254720211029053, 0.004240384101867676, 0.009591808319091797, 0.004273151874542237, 0.004283391952514649, 0.004290559768676758, 0.004291584014892578, 0.00430182409286499, 0.004290559768676758, 0.0042854399681091304, 0.004236288070678711, 0.004260863780975342, 0.004245503902435303, 0.004254720211029053, 0.004230144023895263, 0.004238336086273193, 0.0042434558868408205, 0.004265984058380127, 0.0042475519180297855, 0.004268032073974609, 0.004236288070678711, 0.004259840011596679, 0.004280320167541504, 0.004268032073974609, 0.004263936042785644, 0.004257791996002197, 0.004260863780975342, 0.00425267219543457, 0.004260863780975342, 0.004261951923370361, 0.004249536037445069, 0.004262911796569824, 0.004272128105163574, 0.004257791996002197, 0.004263936042785644, 0.004262911796569824, 0.004248576164245605, 0.0042434558868408205, 0.004262911796569824, 0.004257791996002197, 0.00425267219543457, 0.004250624179840088, 0.004265984058380127, 0.004260863780975342, 0.0042434558868408205, 0.004255743980407714, 0.004246528148651123, 0.004254720211029053, 0.0042741761207580565, 0.004258815765380859, 0.004249599933624268, 0.004260863780975342, 0.004253695964813233, 0.0042772479057312015, 0.004248576164245605, 0.004264959812164307, 0.004298751831054688, 0.004302847862243653, 0.004297728061676025, 0.0043089919090271, 0.004364287853240967, 0.004294655799865723, 0.004304895877838135, 0.004291584014892578, 0.004297791957855224, 0.009573311805725097, 0.004294655799865723, 0.0042854399681091304, 0.004296703815460205, 0.004276224136352539, 0.004261888027191162, 0.004255743980407714, 0.004259840011596679, 0.004235263824462891, 0.004254720211029053, 0.004260863780975342, 0.004245503902435303, 0.0043038721084594726, 0.00429260778427124, 0.00429363203048706, 0.004290559768676758, 0.00429363203048706, 0.004279295921325684, 0.004257791996002197, 0.004232192039489746, 0.004248576164245605, 0.00425267219543457, 0.004242432117462158, 0.004436992168426514, 0.004380671977996826, 0.004290559768676758, 0.004362239837646485, 0.004439040184020996, 0.004428800106048584, 0.004868095874786377, 0.004589568138122559, 0.004495359897613525, 0.004448256015777588, 0.004527103900909424, 0.004495359897613525, 0.0045281281471252445, 0.004482048034667969, 0.004461567878723144, 0.004538368225097656, 0.004467711925506592, 0.004452352046966553, 0.004469759941101074, 0.0044011521339416505, 0.004392960071563721, 0.004621312141418457, 0.004461567878723144, 0.004456448078155518, 0.0044718079566955565, 0.00446668815612793, 0.004458528041839599, 0.004418528079986573, 0.004376575946807861, 0.004380671977996826, 0.004396031856536866, 0.004383743762969971, 0.004386816024780273, 0.004389887809753418, 0.004397056102752686, 0.004387839794158936, 0.004386816024780273, 0.004404223918914795, 0.004380671977996826, 0.0043786239624023435, 0.01031884765625, 0.004459519863128662, 0.004498432159423828, 0.0044349441528320314, 0.004399104118347168, 0.004407296180725098, 0.0043786239624023435, 0.0043786239624023435, 0.004391935825347901, 0.004413440227508545, 0.004389887809753418, 0.004443136215209961, 0.004467711925506592, 0.00444927978515625, 0.004445184230804444, 0.004445184230804444, 0.004440063953399658, 0.004468736171722412, 0.004413440227508545, 0.004383743762969971, 0.004390912055969238, 0.004415487766265869, 0.004402175903320312, 0.004380671977996826, 0.004383808135986328, 0.004287424087524414, 0.004267007827758789, 0.00425267219543457, 0.004245503902435303, 0.00425267219543457, 0.0042434558868408205, 0.004253695964813233, 0.004260863780975342, 0.004275199890136719, 0.004236288070678711, 0.004299776077270508, 0.004255743980407714, 0.004267007827758789, 0.00425164794921875, 0.004265984058380127, 0.004248576164245605, 0.004292672157287598, 0.004302783966064453, 0.004286464214324951, 0.004281343936920166, 0.004250624179840088, 0.004248576164245605, 0.004262911796569824, 0.004254752159118652, 0.004262879848480224, 0.0042782721519470214, 0.004249599933624268, 0.004225024223327637, 0.004258815765380859, 0.0042475519180297855, 0.004264959812164307, 0.004253695964813233, 0.004254720211029053, 0.004253695964813233, 0.0042434558868408205, 0.004269055843353272, 0.004324351787567139, 0.004260863780975342, 0.009602047920227052, 0.004282368183135986, 0.0042772479057312015, 0.0042475519180297855, 0.004259840011596679, 0.004254720211029053, 0.00425267219543457, 0.004259840011596679, 0.004541440010070801, 0.0043069438934326175, 0.0043079681396484375, 0.004304895877838135, 0.00429363203048706, 0.00430079984664917, 0.004297728061676025, 0.0042854399681091304, 0.004267007827758789, 0.004248576164245605, 0.004260863780975342, 0.00425267219543457, 0.004294655799865723, 0.004297728061676025, 0.00429363203048706, 0.00431001615524292, 0.004291584014892578, 0.004313087940216064, 0.00430079984664917, 0.004283391952514649, 0.004297728061676025, 0.004273151874542237, 0.004265984058380127, 0.004230144023895263, 0.004254720211029053, 0.004257791996002197, 0.004268032073974609, 0.004256768226623535, 0.004259840011596679, 0.004294655799865723, 0.004289535999298095, 0.00430182409286499, 0.004298751831054688, 0.004241407871246338, 0.004250624179840088, 0.004248576164245605, 0.004234240055084228, 0.004255743980407714, 0.004239359855651856, 0.004257791996002197, 0.004255807876586914, 0.004262847900390625, 0.004427775859832763, 0.004256768226623535, 0.0042475519180297855, 0.00429363203048706, 0.004352000236511231, 0.004255743980407714, 0.004245535850524902, 0.004248544216156006, 0.004720640182495117, 0.00467251205444336, 0.004570112228393554, 0.004465663909912109, 0.00450764799118042, 0.010275839805603027, 0.004914175987243652, 0.004994048118591309, 0.004544511795043945, 0.004447231769561768, 0.0044718399047851565, 0.004543456077575683, 0.004454400062561035, 0.004473855972290039, 0.00447488021850586, 0.004453375816345215, 0.004538368225097656, 0.0044646401405334475, 0.00445030403137207, 0.004451327800750733, 0.004483071804046631, 0.004440063953399658, 0.004456448078155518, 0.0044308481216430665, 0.0044380159378051755, 0.004467711925506592, 0.004441088199615479, 0.004492288112640381, 0.004391935825347901, 0.004376575946807861, 0.004403200149536133, 0.004379648208618164, 0.004415487766265869, 0.004516863822937011, 0.0045097279548645015, 0.0048455362319946285, 0.004518911838531494, 0.004454400062561035, 0.004451327800750733, 0.004472832202911377, 0.004511744022369385, 0.004492288112640381, 0.004472832202911377, 0.004434976100921631, 0.004391903877258301, 0.004390912055969238, 0.004403200149536133, 0.0044083199501037595, 0.004396031856536866, 0.004406271934509277, 0.004393983840942383, 0.004352000236511231, 0.004248576164245605, 0.004250624179840088, 0.004260863780975342, 0.00429363203048706, 0.004280320167541504, 0.004286464214324951, 0.004242432117462158, 0.004228096008300781, 0.004248576164245605, 0.004242432117462158, 0.004227071762084961, 0.004239359855651856, 0.0042434558868408205, 0.004242432117462158, 0.004246528148651123, 0.004264959812164307, 0.009967616081237793, 0.004446208000183105, 0.004436992168426514, 0.004440063953399658, 0.004436992168426514, 0.00448412799835205, 0.004447199821472168, 0.004448256015777588, 0.004435967922210693, 0.004457471847534179, 0.004447296142578125, 0.004454336166381836, 0.004457471847534179, 0.004476928234100342, 0.0044380159378051755, 0.004457471847534179, 0.0045660161972045895, 0.004478975772857666, 0.00444927978515625, 0.004444159984588623, 0.004440127849578857, 0.0044328317642211915, 0.004470784187316895, 0.004453375816345215, 0.004447231769561768, 0.004443136215209961, 0.004483071804046631, 0.004440063953399658, 0.004439040184020996, 0.004424704074859619, 0.004426752090454102, 0.004415487766265869, 0.0042772479057312015, 0.004291584014892578, 0.004291584014892578, 0.004295680046081543, 0.0043038721084594726, 0.004282368183135986, 0.0042741761207580565, 0.004297728061676025, 0.004296703815460205, 0.004298751831054688, 0.004287487983703613, 0.004260863780975342, 0.004229119777679443, 0.004255743980407714, 0.004236288070678711, 0.004265984058380127, 0.004253695964813233, 0.004264959812164307, 0.004232192039489746, 0.004250624179840088, 0.0042854399681091304, 0.004256768226623535, 0.004261919975280762, 0.004271071910858154, 0.004242432117462158, 0.00429363203048706, 0.004256768226623535, 0.004248576164245605, 0.004267007827758789, 0.004259840011596679, 0.004248576164245605, 0.009666560173034668, 0.004288512229919434, 0.004264959812164307, 0.004295680046081543, 0.004271103858947754, 0.004282368183135986, 0.004236288070678711, 0.004263936042785644, 0.004233215808868408, 0.004264959812164307, 0.004236288070678711, 0.004245503902435303, 0.004262911796569824, 0.004377600193023682, 0.005079040050506592, 0.0048752641677856446, 0.0049827837944030765, 0.006623231887817383, 0.006615039825439453, 0.004719615936279297, 0.004523007869720459, 0.004492288112640381, 0.00451584005355835, 0.0044718079566955565, 0.004448256015777588, 0.004473855972290039, 0.00444927978515625, 0.004455423831939697, 0.004468736171722412, 0.0044308481216430665, 0.004369408130645752, 0.0043038721084594726, 0.004283391952514649, 0.004276224136352539, 0.00429260778427124, 0.004304895877838135, 0.0043069438934326175, 0.004296703815460205, 0.00430079984664917, 0.004305920124053955, 0.004313087940216064, 0.004299776077270508, 0.004299776077270508, 0.0043745279312133786, 0.004314112186431885, 0.0043151359558105465, 0.0042782721519470214, 0.004294655799865723, 0.004291584014892578, 0.004281343936920166, 0.004289535999298095, 0.00430182409286499, 0.004283391952514649, 0.004308032035827637, 0.004431807994842529, 0.004291584014892578, 0.004261888027191162, 0.004258815765380859, 0.0042782721519470214, 0.004255743980407714, 0.004263936042785644, 0.004245503902435303, 0.004260863780975342, 0.009666560173034668, 0.00429260778427124, 0.004280320167541504, 0.004299776077270508, 0.0043069438934326175, 0.00430079984664917, 0.00429260778427124, 0.004289535999298095, 0.00425267219543457, 0.004267007827758789, 0.004257791996002197, 0.004259840011596679, 0.004265984058380127, 0.004255743980407714, 0.0042475519180297855, 0.004258815765380859, 0.004254720211029053, 0.004260863780975342, 0.004260863780975342, 0.004266047954559326, 0.004273087978363037, 0.004254720211029053, 0.004260863780975342, 0.004267007827758789, 0.004257791996002197, 0.004245503902435303, 0.004268032073974609, 0.0042475519180297855, 0.004257791996002197, 0.004288512229919434, 0.004254720211029053, 0.004254720211029053, 0.004258815765380859, 0.004296703815460205, 0.00425164794921875, 0.004253695964813233, 0.004281343936920166, 0.004237311840057373, 0.004245535850524902, 0.004250592231750489, 0.004254720211029053, 0.004253695964813233, 0.004253695964813233, 0.0042475519180297855, 0.00424348783493042, 0.004246496200561524, 0.00424454402923584, 0.004253632068634034, 0.00425171184539795, 0.0042618241310119626, 0.0042557759284973145, 0.004255712032318115, 0.004254720211029053, 0.004258815765380859, 0.004258815765380859, 0.004259840011596679, 0.004242432117462158, 0.004249599933624268, 0.004250624179840088, 0.004281343936920166, 0.004250624179840088, 0.004227071762084961, 0.004237311840057373, 0.009432064056396485, 0.0042475519180297855, 0.0042434558868408205, 0.0042475519180297855, 0.004275199890136719, 0.0043110399246215824, 0.0042772479057312015, 0.004294655799865723, 0.004281343936920166, 0.0042854399681091304, 0.0042854399681091304, 0.00429260778427124, 0.0042741761207580565, 0.004313087940216064, 0.004249599933624268, 0.004267007827758789, 0.004248576164245605, 0.004314112186431885, 0.004655104160308838, 0.004612095832824707, 0.004488192081451416, 0.004463615894317627, 0.004402175903320312, 0.00429363203048706, 0.004249599933624268, 0.004260863780975342, 0.004240384101867676, 0.004246528148651123, 0.004254784107208252, 0.004262847900390625, 0.004233215808868408, 0.004256768226623535, 0.004262911796569824, 0.004268032073974609, 0.004257791996002197, 0.004255743980407714, 0.004246528148651123, 0.004269055843353272, 0.004246528148651123, 0.004287487983703613, 0.004255743980407714, 0.004260863780975342, 0.004241407871246338, 0.004273151874542237, 0.004257823944091797, 0.004263904094696045, 0.004253695964813233, 0.004269055843353272, 0.004270112037658692, 0.004252639770507813, 0.004262911796569824, 0.004255743980407714, 0.0042475519180297855, 0.004256768226623535, 0.004267007827758789, 0.004255743980407714, 0.004260863780975342, 0.0042639999389648435, 0.0042618241310119626, 0.004287487983703613, 0.004343808174133301, 0.004288512229919434, 0.004298751831054688, 0.009564160346984863, 0.00431001615524292, 0.004286464214324951, 0.004286464214324951, 0.004289535999298095, 0.00430079984664917, 0.00429260778427124, 0.004298751831054688, 0.004298751831054688, 0.004298751831054688, 0.00425267219543457, 0.004254720211029053, 0.004250688076019287, 0.004249536037445069, 0.0042475519180297855, 0.004276224136352539, 0.004249599933624268, 0.004258815765380859, 0.004255743980407714, 0.004258815765380859, 0.00425267219543457, 0.004256800174713134, 0.004249567985534668, 0.004246528148651123, 0.004248576164245605, 0.0042475519180297855, 0.004270080089569092, 0.00423529577255249, 0.004248544216156006, 0.004232192039489746, 0.004245503902435303, 0.00425164794921875, 0.00435814380645752, 0.004242432117462158, 0.004241407871246338, 0.004242432117462158, 0.0042444801330566405, 0.004242464065551758, 0.004232160091400147, 0.004245503902435303, 0.004246528148651123, 0.004283391952514649, 0.004260863780975342, 0.0042475519180297855, 0.0042219839096069335, 0.004248544216156006, 0.004226047992706299, 0.004283391952514649, 0.004263936042785644, 0.004245503902435303, 0.004245503902435303, 0.004260863780975342, 0.004258815765380859, 0.004265984058380127, 0.00425267219543457, 0.00425164794921875, 0.004250624179840088, 0.004257791996002197, 0.004245503902435303, 0.004257791996002197, 0.0042444801330566405, 0.004256768226623535, 0.004299776077270508, 0.009579520225524902, 0.004276224136352539, 0.004276224136352539, 0.004248576164245605, 0.004241407871246338, 0.00425267219543457, 0.004245503902435303, 0.004239359855651856, 0.004229119777679443, 0.004256768226623535, 0.004264959812164307, 0.004250624179840088, 0.004241407871246338, 0.00425267219543457, 0.004241407871246338, 0.004253695964813233, 0.004239359855651856, 0.004255743980407714, 0.0042444801330566405, 0.004242432117462158, 0.004256768226623535, 0.004262911796569824, 0.004246528148651123, 0.004249599933624268, 0.004239359855651856, 0.004238336086273193, 0.004272128105163574, 0.004245503902435303, 0.004253695964813233, 0.0043489279747009274, 0.004260863780975342, 0.0042475519180297855, 0.0042475519180297855, 0.004407296180725098, 0.004241407871246338, 0.004263936042785644, 0.004237311840057373, 0.004246528148651123, 0.004249599933624268, 0.00425164794921875, 0.004249599933624268, 0.004275263786315918, 0.004246496200561524, 0.004239327907562256, 0.004250624179840088, 0.00425267219543457, 0.004262911796569824, 0.004291584014892578, 0.004283391952514649, 0.004284416198730469, 0.00429472017288208, 0.004297664165496826, 0.0042854399681091304, 0.004283391952514649, 0.004284448146820068, 0.004318175792694091, 0.004295680046081543, 0.004424704074859619, 0.004250624179840088, 0.004303904056549073, 0.004303840160369873, 0.004329472064971924, 0.004290559768676758, 0.009687040328979492, 0.004316160202026367, 0.0042854399681091304, 0.004288512229919434, 0.004284416198730469, 0.004289567947387695, 0.004303840160369873, 0.00429260778427124, 0.004280320167541504, 0.0043089919090271, 0.00430182409286499, 0.0042772479057312015, 0.0042854399681091304, 0.004281343936920166, 0.0043069438934326175, 0.00429260778427124, 0.004260863780975342, 0.004264959812164307, 0.004261888027191162, 0.00425164794921875, 0.004241407871246338, 0.004260863780975342, 0.004261888027191162, 0.0042475519180297855, 0.004260863780975342, 0.004254720211029053, 0.004270080089569092, 0.004264959812164307, 0.004286464214324951, 0.004233215808868408, 0.004271103858947754, 0.004238336086273193, 0.004245503902435303, 0.004234240055084228, 0.004239391803741455, 0.004245471954345703, 0.004255743980407714, 0.004233215808868408, 0.004233215808868408, 0.004259840011596679, 0.004260863780975342, 0.004238336086273193, 0.0042434558868408205, 0.004248576164245605, 0.004230144023895263, 0.0042434558868408205, 0.004265984058380127, 0.004260863780975342, 0.0042444801330566405, 0.004267007827758789, 0.0042854399681091304, 0.00429363203048706, 0.00429260778427124, 0.004284416198730469, 0.004294655799865723, 0.004305952072143555, 0.004285408020019531, 0.004297728061676025, 0.004288512229919434, 0.0042824001312255855, 0.004725728034973144, 0.0046561279296875, 0.004433919906616211, 0.009876480102539062, 0.004404223918914795, 0.004412415981292724, 0.0043919677734375, 0.0043703999519348145, 0.004267007827758789, 0.004232192039489746, 0.004241407871246338, 0.004235263824462891, 0.004287487983703613, 0.004290559768676758, 0.004296703815460205, 0.004282368183135986, 0.0042854399681091304, 0.004288512229919434, 0.004262911796569824, 0.004233215808868408, 0.004258815765380859, 0.004249663829803467, 0.004240320205688476, 0.004257791996002197, 0.004259840011596679, 0.004299776077270508, 0.004289535999298095, 0.004291584014892578, 0.004312064170837402, 0.004296703815460205, 0.004257791996002197, 0.0042444801330566405, 0.004248576164245605, 0.004261888027191162, 0.004253695964813233, 0.004241407871246338, 0.004259840011596679, 0.004286464214324951, 0.004290559768676758, 0.004281375885009766, 0.00429257583618164, 0.004280320167541504, 0.004304895877838135, 0.00430079984664917, 0.00429260778427124, 0.004322303771972656, 0.004294655799865723, 0.00429363203048706, 0.004283391952514649, 0.004270080089569092, 0.004281343936920166, 0.004291584014892578, 0.004276224136352539, 0.004283391952514649, 0.004271103858947754, 0.004263936042785644, 0.004261888027191162, 0.004239359855651856, 0.00425164794921875, 0.004249599933624268, 0.0042444801330566405, 0.004246528148651123, 0.004276224136352539, 0.004289535999298095, 0.0043110399246215824, 0.004290559768676758, 0.00952627182006836, 0.004260863780975342, 0.004241407871246338, 0.004398079872131347, 0.004416512012481689, 0.004443136215209961, 0.004411392211914063, 0.00435097599029541, 0.004286464214324951, 0.004298751831054688, 0.004287487983703613, 0.004249599933624268, 0.004246528148651123, 0.004257791996002197, 0.004245503902435303, 0.004325376033782959, 0.004248576164245605, 0.004254752159118652, 0.004245471954345703, 0.004267007827758789, 0.004253695964813233, 0.004245503902435303, 0.00425267219543457, 0.004256768226623535, 0.0042434558868408205, 0.004249599933624268, 0.0042475519180297855, 0.004239359855651856, 0.0042475519180297855, 0.004282368183135986, 0.0042741761207580565, 0.004388864040374756, 0.004245503902435303, 0.004226079940795898, 0.004257760047912598, 0.004241407871246338, 0.004241407871246338, 0.004240384101867676, 0.0042444801330566405, 0.0043089919090271, 0.004291584014892578, 0.004305920124053955, 0.004288512229919434, 0.004302847862243653, 0.004284416198730469, 0.0042854399681091304, 0.004245503902435303, 0.004256768226623535, 0.004250624179840088, 0.004269055843353272, 0.004260863780975342, 0.004263936042785644, 0.004264959812164307, 0.004260863780975342, 0.00425267219543457, 0.004263936042785644, 0.004258815765380859, 0.004234240055084228, 0.004256768226623535, 0.004257791996002197, 0.004258815765380859, 0.004254720211029053, 0.004259840011596679, 0.009850879669189454, 0.00425164794921875, 0.004239359855651856, 0.004250624179840088, 0.0042772479057312015, 0.0043079681396484375, 0.004294655799865723, 0.004249631881713867, 0.004240352153778076, 0.004250624179840088, 0.0042434558868408205, 0.0043089919090271, 0.004279327869415283, 0.004287456035614014, 0.004277279853820801, 0.0042956480979919436, 0.004288512229919434, 0.004313087940216064, 0.0042741761207580565, 0.004253695964813233, 0.004261888027191162, 0.004262911796569824, 0.004261888027191162, 0.004255743980407714, 0.0042527041435241695, 0.0042772159576416014, 0.0042516798973083495, 0.004257760047912598, 0.004271103858947754, 0.004262911796569824, 0.004260863780975342, 0.004258815765380859, 0.0042516798973083495, 0.004260831832885742, 0.004261919975280762, 0.004260831832885742, 0.004257823944091797, 0.0042772159576416014, 0.004245503902435303, 0.0042741761207580565, 0.0042782721519470214, 0.0042782721519470214, 0.004262911796569824, 0.004249599933624268, 0.004291584014892578, 0.0043038721084594726, 0.004294655799865723, 0.004298751831054688, 0.004279295921325684, 0.004260863780975342, 0.004265984058380127, 0.004298783779144287, 0.004282336235046386, 0.004290559768676758, 0.004272128105163574, 0.0043089919090271, 0.004335616111755371, 0.004279295921325684, 0.0042854399681091304, 0.004284416198730469, 0.004248576164245605, 0.004248576164245605, 0.004236288070678711, 0.009549823760986328, 0.004295680046081543, 0.004242432117462158, 0.00423529577255249, 0.004239327907562256, 0.0042434558868408205, 0.004255743980407714, 0.004237311840057373, 0.004237311840057373, 0.004267007827758789, 0.004246528148651123, 0.0042434558868408205, 0.004233215808868408, 0.004245503902435303, 0.004250624179840088, 0.004253727912902832, 0.004242400169372559, 0.004290559768676758, 0.00425164794921875, 0.004232192039489746, 0.004259840011596679, 0.004241407871246338, 0.004248576164245605, 0.004246528148651123, 0.004248576164245605, 0.004227071762084961, 0.004253695964813233, 0.004265984058380127, 0.004286464214324951, 0.0042772479057312015, 0.004302847862243653, 0.0042772479057312015, 0.004282368183135986, 0.0043069438934326175, 0.00430079984664917, 0.004287487983703613, 0.004295680046081543, 0.0042854399681091304, 0.004988927841186524, 0.0044011521339416505, 0.0043673601150512695, 0.004291584014892578, 0.0042444801330566405, 0.004329472064971924, 0.004235263824462891, 0.004239359855651856, 0.004233215808868408, 0.004238336086273193, 0.004253727912902832, 0.004286431789398194, 0.0042772479057312015, 0.004287487983703613, 0.004295680046081543, 0.004296703815460205, 0.00430079984664917, 0.004282368183135986, 0.004248576164245605, 0.0042444801330566405, 0.004240384101867676, 0.004257791996002197, 0.004235263824462891, 0.004226047992706299, 0.004255743980407714, 0.009521151542663574, 0.0042772479057312015, 0.004271103858947754, 0.004279295921325684, 0.004237343788146972, 0.004240352153778076, 0.004253695964813233, 0.0042434558868408205, 0.004230144023895263, 0.004246528148651123, 0.004289535999298095, 0.004289535999298095, 0.00430182409286499, 0.00429260778427124, 0.004261888027191162, 0.004272128105163574, 0.004262911796569824, 0.004258815765380859, 0.004255743980407714, 0.004270080089569092, 0.004249599933624268, 0.004290559768676758, 0.004249599933624268, 0.004259840011596679, 0.0042444801330566405, 0.004302847862243653, 0.004284416198730469, 0.004286464214324951, 0.004296703815460205, 0.004295711994171143, 0.00429257583618164, 0.00430079984664917, 0.0042854399681091304, 0.004326399803161621, 0.004281375885009766, 0.0043294401168823245, 0.004294655799865723, 0.004302847862243653, 0.0043151359558105465, 0.004261888027191162, 0.0042444801330566405, 0.0042475519180297855, 0.004260863780975342, 0.004261888027191162, 0.004298751831054688, 0.00425164794921875, 0.004250624179840088, 0.004289535999298095, 0.004250624179840088, 0.0042445120811462405, 0.004279263973236084, 0.004239359855651856, 0.004250624179840088, 0.00425267219543457, 0.004326399803161621, 0.004255743980407714, 0.004239359855651856, 0.004240384101867676, 0.0042444801330566405, 0.0046356477737426755, 0.004426752090454102, 0.004297728061676025, 0.004325376033782959, 0.009834495544433594, 0.004294655799865723, 0.00429260778427124, 0.00429260778427124, 0.004290559768676758, 0.004284416198730469, 0.00430182409286499, 0.0042444801330566405, 0.004231167793273926, 0.004265984058380127, 0.004253695964813233, 0.004247583866119385, 0.004250592231750489, 0.004257791996002197, 0.004250624179840088, 0.00425267219543457, 0.0042772479057312015, 0.004317183971405029, 0.004320256233215332, 0.004280320167541504, 0.004263936042785644, 0.004258815765380859, 0.004260863780975342, 0.00425267219543457, 0.004290559768676758, 0.004246528148651123, 0.004250624179840088, 0.0042444801330566405, 0.004255743980407714, 0.004249599933624268, 0.004248576164245605, 0.004240384101867676, 0.004249599933624268, 0.004256768226623535, 0.004261888027191162, 0.004248576164245605, 0.00425164794921875, 0.00425267219543457, 0.004254720211029053, 0.004272128105163574, 0.004259840011596679, 0.004245503902435303, 0.0042475519180297855, 0.00424351978302002, 0.004244416236877441, 0.004253695964813233, 0.0042444801330566405, 0.004253695964813233, 0.0042527041435241695, 0.00425980806350708, 0.004256768226623535, 0.0042782721519470214, 0.004260863780975342, 0.004249599933624268, 0.004268032073974609, 0.004279295921325684, 0.004299776077270508, 0.004238336086273193, 0.004238336086273193, 0.004228096008300781, 0.004246528148651123, 0.0042147841453552244, 0.004238336086273193, 0.009485312461853027, 0.00425267219543457, 0.004232192039489746, 0.004264959812164307, 0.00425267219543457, 0.004248576164245605, 0.004229119777679443, 0.004240384101867676, 0.004220928192138672, 0.00425164794921875, 0.004233215808868408, 0.004242432117462158, 0.004226047992706299, 0.004259840011596679, 0.0042434558868408205, 0.004249599933624268, 0.004270080089569092, 0.0043110399246215824, 0.004286464214324951, 0.004345856189727783, 0.00429363203048706, 0.00429260778427124, 0.004319231986999511, 0.004287487983703613, 0.004271103858947754, 0.004284416198730469, 0.0042772479057312015, 0.004302847862243653, 0.004288512229919434, 0.00429472017288208, 0.004290495872497559, 0.004295680046081543, 0.004287487983703613, 0.004324351787567139, 0.004275199890136719, 0.004255743980407714, 0.004256768226623535, 0.004257791996002197, 0.004282368183135986, 0.00429363203048706, 0.004286464214324951, 0.0043038721084594726, 0.0042854399681091304, 0.004288512229919434, 0.004237311840057373, 0.004246528148651123, 0.004250624179840088, 0.004276224136352539, 0.004228096008300781, 0.004232192039489746, 0.004258815765380859, 0.004239359855651856, 0.004250624179840088, 0.004234240055084228, 0.004239359855651856, 0.004237311840057373, 0.004263936042785644, 0.004289535999298095, 0.004296703815460205, 0.004294655799865723, 0.004298751831054688, 0.0043110399246215824, 0.004286464214324951, 0.009535488128662109, 0.004289535999298095, 0.004274208068847657, 0.004286431789398194, 0.004289535999298095, 0.004289535999298095, 0.004296703815460205, 0.004299776077270508, 0.004280320167541504, 0.004319231986999511, 0.004284480094909668, 0.00428332805633545, 0.004294655799865723, 0.0043079681396484375, 0.004294655799865723, 0.004897791862487793, 0.004454400062561035, 0.004456448078155518, 0.004420608043670654, 0.004456448078155518, 0.004380671977996826, 0.004263936042785644, 0.004263936042785644, 0.004264959812164307, 0.0042741761207580565, 0.004257791996002197, 0.004271103858947754, 0.004267072200775147, 0.00425977611541748, 0.0042444801330566405, 0.004262911796569824, 0.004239359855651856, 0.0042444801330566405, 0.004272128105163574, 0.004257823944091797, 0.004238304138183594, 0.00425164794921875, 0.004232192039489746, 0.004267007827758789, 0.004263936042785644, 0.004254720211029053, 0.004260863780975342, 0.004258815765380859, 0.004268032073974609, 0.0042475519180297855, 0.004261888027191162, 0.004246528148651123, 0.004254720211029053, 0.00425267219543457, 0.004287487983703613, 0.004245503902435303, 0.004286464214324951, 0.0043110399246215824, 0.004304959774017334, 0.004296639919281006, 0.004289535999298095, 0.004298751831054688, 0.00430182409286499, 0.004287487983703613, 0.004296703815460205, 0.0043079681396484375, 0.004291584014892578, 0.004316160202026367, 0.00970751953125, 0.00429472017288208, 0.004286399841308594, 0.00429363203048706, 0.004270080089569092, 0.004258815765380859, 0.004232192039489746, 0.004239359855651856, 0.004230144023895263, 0.004282368183135986, 0.004242432117462158, 0.004248576164245605, 0.0042475519180297855, 0.004262911796569824, 0.004249599933624268, 0.004266047954559326, 0.004239295959472656, 0.004268032073974609, 0.004242432117462158, 0.004271168231964112, 0.004258783817291259, 0.004390880107879639, 0.004332543849945069, 0.004257791996002197, 0.0042772479057312015, 0.004261888027191162, 0.004237311840057373, 0.004242432117462158, 0.004245503902435303, 0.004398079872131347, 0.00425267219543457, 0.004241407871246338, 0.004231167793273926, 0.004270080089569092, 0.0042434558868408205, 0.004279295921325684, 0.004302847862243653, 0.004246528148651123, 0.004240384101867676, 0.004270080089569092, 0.004245503902435303, 0.00425267219543457, 0.00425164794921875, 0.004248576164245605, 0.004256768226623535, 0.004241407871246338, 0.004233215808868408, 0.004248576164245605, 0.004245503902435303, 0.00425267219543457, 0.004254720211029053, 0.004248576164245605, 0.004248576164245605, 0.004282368183135986, 0.004237311840057373, 0.0042741761207580565, 0.0042434558868408205, 0.004316160202026367, 0.004299776077270508, 0.004262911796569824, 0.004249599933624268, 0.004245503902435303, 0.004229119777679443, 0.009516032218933105, 0.00425267219543457, 0.004289535999298095, 0.004268095970153808, 0.004246463775634765, 0.00425267219543457, 0.0042434558868408205, 0.004249599933624268, 0.004256768226623535, 0.004268032073974609, 0.004283391952514649, 0.004238336086273193, 0.004287487983703613, 0.004254720211029053, 0.004235263824462891, 0.004249599933624268, 0.004239359855651856, 0.004259840011596679, 0.004313087940216064, 0.0042475519180297855, 0.004245503902435303, 0.0042434558868408205, 0.00425164794921875, 0.004248576164245605, 0.004246528148651123, 0.004246528148651123, 0.004259840011596679, 0.004258815765380859, 0.00425267219543457, 0.004242432117462158, 0.004242432117462158, 0.0042434558868408205, 0.00425164794921875, 0.004262911796569824, 0.004250624179840088, 0.004294655799865723, 0.004275199890136719, 0.004250624179840088, 0.004246528148651123, 0.004259840011596679, 0.004241407871246338, 0.004240384101867676, 0.004254720211029053, 0.00425267219543457, 0.004242432117462158, 0.004246528148651123, 0.004233215808868408, 0.004254720211029053, 0.004271135807037353, 0.004284383773803711, 0.004258815765380859, 0.004254720211029053, 0.004269055843353272, 0.004259840011596679, 0.004271103858947754, 0.004250624179840088, 0.004260863780975342, 0.004259840011596679, 0.00425267219543457, 0.004279295921325684, 0.004257791996002197, 0.004257791996002197, 0.0042475519180297855, 0.009614336013793945, 0.00430079984664917, 0.004298751831054688, 0.0043069438934326175, 0.0043110399246215824, 0.004295680046081543, 0.004275199890136719, 0.004279295921325684, 0.004283391952514649, 0.004455423831939697, 0.004312064170837402, 0.004287487983703613, 0.004289535999298095, 0.004282368183135986, 0.004288512229919434, 0.00430079984664917, 0.004282368183135986, 0.00429363203048706, 0.004317183971405029, 0.0042772479057312015, 0.00430182409286499, 0.004297728061676025, 0.004302847862243653, 0.004319231986999511, 0.004286464214324951, 0.004283391952514649, 0.004279295921325684, 0.004284416198730469, 0.004299776077270508, 0.004288512229919434, 0.00429260778427124, 0.004289535999298095, 0.004616191864013672, 0.004299808025360108, 0.004286431789398194, 0.004316160202026367, 0.004304895877838135, 0.004281343936920166, 0.004283391952514649, 0.004284416198730469, 0.0042772479057312015, 0.0043089919090271, 0.004283391952514649, 0.004287487983703613, 0.0043079681396484375, 0.004286464214324951, 0.004284416198730469, 0.004296703815460205, 0.0042854399681091304, 0.004283391952514649, 0.004324351787567139, 0.004289535999298095, 0.004291584014892578, 0.0042741761207580565, 0.004271135807037353, 0.004293600082397461, 0.00444927978515625, 0.004370463848114014, 0.004327392101287842, 0.004321280002593994, 0.004328447818756104, 0.004332543849945069, 0.004381696224212647, 0.00972390365600586, 0.0042772479057312015, 0.004279295921325684, 0.004273151874542237, 0.004284416198730469, 0.004280320167541504, 0.0042854399681091304, 0.004290559768676758, 0.004304895877838135, 0.00430182409286499, 0.004280320167541504, 0.0042782721519470214, 0.004282368183135986, 0.004294655799865723, 0.004279295921325684, 0.0043069438934326175, 0.0043038721084594726, 0.004305920124053955, 0.0043069438934326175, 0.004287487983703613, 0.004287487983703613, 0.004329472064971924, 0.0043079681396484375, 0.0043079681396484375, 0.004305920124053955, 0.004295680046081543, 0.004313087940216064, 0.0042854399681091304, 0.00430182409286499, 0.004254752159118652, 0.004316127777099609, 0.004295680046081543, 0.004295680046081543, 0.00429260778427124, 0.004268032073974609, 0.004272128105163574, 0.004259840011596679, 0.0042444801330566405, 0.004248576164245605, 0.004296703815460205, 0.004296703815460205, 0.0043038721084594726, 0.00429363203048706, 0.004323328018188476, 0.004326399803161621, 0.00431820821762085, 0.00434278392791748, 0.004289535999298095, 0.004328447818756104, 0.004312064170837402, 0.004268032073974609, 0.004253695964813233, 0.00425267219543457, 0.004242432117462158, 0.004261888027191162, 0.004264959812164307, 0.0042475519180297855, 0.004335616111755371, 0.004250624179840088, 0.004253695964813233, 0.004263936042785644, 0.00430079984664917, 0.004296703815460205, 0.009579520225524902, 0.004355072021484375, 0.004281343936920166, 0.00430079984664917, 0.004291584014892578, 0.004256800174713134, 0.004261856079101563, 0.0042444801330566405, 0.004240384101867676, 0.00425267219543457, 0.0043203201293945314, 0.004257728099822998, 0.004248576164245605, 0.0042444801330566405, 0.004233215808868408, 0.0042475519180297855, 0.004238336086273193, 0.004253695964813233, 0.0042434558868408205, 0.004302879810333252, 0.0042659521102905275, 0.00425267219543457, 0.0042475519180297855, 0.004257791996002197, 0.004250624179840088, 0.0043151359558105465, 0.004258815765380859, 0.004236288070678711, 0.004276224136352539, 0.00431001615524292, 0.0043038721084594726, 0.004284416198730469, 0.004242432117462158, 0.00425267219543457, 0.004258815765380859, 0.0043089919090271, 0.00431820821762085, 0.004334591865539551, 0.004268032073974609, 0.004329472064971924, 0.004270080089569092, 0.004265984058380127, 0.0042444801330566405, 0.004250624179840088, 0.004240384101867676, 0.004242432117462158, 0.004253695964813233, 0.004255743980407714, 0.004281343936920166, 0.004255743980407714, 0.004261888027191162, 0.004255743980407714, 0.004239359855651856, 0.00425164794921875, 0.004233215808868408, 0.004268032073974609, 0.0042434558868408205, 0.0042434558868408205, 0.004240384101867676, 0.0042475519180297855, 0.004254720211029053, 0.004249599933624268, 0.0042444801330566405, 0.009563136100769042, 0.004237311840057373, 0.004250624179840088, 0.004317183971405029, 0.004286464214324951, 0.004283423900604248, 0.004293600082397461, 0.004258815765380859, 0.004239359855651856, 0.00425164794921875, 0.004249631881713867, 0.004240352153778076, 0.004270080089569092, 0.004256768226623535, 0.00425267219543457, 0.004259840011596679, 0.004239359855651856, 0.004263936042785644, 0.004284416198730469, 0.004257791996002197, 0.004253695964813233, 0.0042444801330566405, 0.004257791996002197, 0.004258815765380859, 0.004255743980407714, 0.0042527360916137695, 0.004262847900390625, 0.004256768226623535, 0.004258815765380859, 0.004253695964813233, 0.004254720211029053, 0.004241407871246338, 0.004248576164245605, 0.004247583866119385, 0.004282336235046386, 0.004281343936920166, 0.004267007827758789, 0.004287487983703613, 0.004263936042785644, 0.004267007827758789, 0.00425164794921875, 0.004250656127929688, 0.004263904094696045, 0.004276224136352539, 0.0042741761207580565, 0.004305920124053955, 0.004325376033782959, 0.004304895877838135, 0.004287487983703613, 0.004275199890136719, 0.00425164794921875, 0.004242432117462158, 0.004259840011596679, 0.004255743980407714, 0.004283391952514649, 0.004269055843353272, 0.004245503902435303, 0.004276224136352539, 0.004253695964813233, 0.004257791996002197, 0.004271103858947754, 0.0042444801330566405, 0.004254720211029053, 0.009606143951416016, 0.00429260778427124, 0.00429260778427124, 0.004264959812164307, 0.004242432117462158, 0.004283391952514649, 0.0042434558868408205, 0.004226047992706299, 0.004232192039489746, 0.004259840011596679, 0.004289535999298095, 0.004280320167541504, 0.004280320167541504, 0.004320256233215332, 0.004294655799865723, 0.004268032073974609, 0.0045015039443969725, 0.004333568096160889, 0.004319231986999511, 0.004295680046081543, 0.00430079984664917, 0.004289535999298095, 0.004281343936920166, 0.00431001615524292, 0.004291584014892578, 0.004295680046081543, 0.004463615894317627, 0.004312064170837402, 0.004319231986999511, 0.004320256233215332, 0.0043110399246215824, 0.004334591865539551, 0.00431820821762085, 0.004325376033782959, 0.004298751831054688, 0.004330495834350586, 0.004299776077270508, 0.00430079984664917, 0.0042854399681091304, 0.004299776077270508, 0.004257791996002197, 0.004294655799865723, 0.004249599933624268, 0.004257791996002197, 0.004289535999298095, 0.0042854399681091304, 0.004295680046081543, 0.004298751831054688, 0.004282368183135986, 0.004480000019073486, 0.004305920124053955, 0.004308032035827637, 0.004289472103118896, 0.004290559768676758, 0.004279295921325684, 0.00425164794921875, 0.004259840011596679, 0.004246528148651123, 0.005114880084991455, 0.0043448319435119625, 0.0043069438934326175, 0.004296703815460205, 0.004245503902435303, 0.009639936447143555, 0.004305920124053955, 0.004323328018188476, 0.004287487983703613, 0.004346943855285644, 0.004304895877838135, 0.0043253121376037595, 0.004289535999298095, 0.004238336086273193, 0.004245503902435303, 0.004302847862243653, 0.004249599933624268, 0.004246528148651123, 0.0042475519180297855, 0.0042782721519470214, 0.004314112186431885, 0.004302847862243653, 0.00431001615524292, 0.0043089919090271, 0.004296703815460205, 0.00431001615524292, 0.004298751831054688, 0.004297728061676025, 0.00431001615524292, 0.004302847862243653, 0.004295680046081543, 0.0043038721084594726, 0.004287487983703613, 0.0047216639518737795, 0.00446668815612793, 0.004457471847534179, 0.00449945592880249, 0.004589568138122559, 0.004776959896087647, 0.004459519863128662, 0.004476928234100342, 0.004470784187316895, 0.004465663909912109, 0.004444159984588623, 0.004354047775268555, 0.004333568096160889, 0.00435203218460083, 0.004346848011016846, 0.004357120037078857, 0.004453375816345215, 0.004474944114685059, 0.004467648029327392, 0.004444159984588623, 0.004473855972290039, 0.0044380159378051755, 0.0044421439170837405, 0.0044338879585266115, 0.0044380159378051755, 0.004473855972290039, 0.0044421119689941405, 0.004316160202026367, 0.0042772479057312015, 0.0042475519180297855, 0.004258815765380859, 0.004268032073974609, 0.004265984058380127, 0.004298751831054688, 0.004268032073974609, 0.009706496238708496, 0.0043079681396484375, 0.004286464214324951, 0.004289535999298095, 0.004279295921325684, 0.004368383884429932, 0.0042854399681091304, 0.004281343936920166, 0.004275199890136719, 0.004259840011596679, 0.004248576164245605, 0.004299776077270508, 0.0042741761207580565, 0.004302847862243653, 0.004287487983703613, 0.004286464214324951, 0.00430182409286499, 0.00429363203048706, 0.004253695964813233, 0.004282368183135986, 0.004270080089569092, 0.004260863780975342, 0.00425164794921875, 0.004253695964813233, 0.004257791996002197, 0.0043110399246215824, 0.004289535999298095, 0.0043079681396484375, 0.004290559768676758, 0.004261888027191162, 0.004256768226623535, 0.004331520080566406, 0.004256768226623535, 0.004288512229919434, 0.004255743980407714, 0.004245503902435303, 0.004286464214324951, 0.004298751831054688, 0.004294655799865723, 0.00431820821762085, 0.00429260778427124, 0.0042741761207580565, 0.00430079984664917, 0.004299776077270508, 0.004283391952514649, 0.00430182409286499, 0.004289535999298095, 0.004275199890136719, 0.004246528148651123, 0.004242432117462158, 0.00425164794921875, 0.004257791996002197, 0.004241407871246338, 0.0042568321228027345, 0.004245440006256104, 0.004267007827758789, 0.004262911796569824, 0.0042741761207580565, 0.0042772479057312015, 0.004250624179840088, 0.004260863780975342, 0.00425164794921875, 0.0042741761207580565, 0.009634816169738769, 0.004260863780975342, 0.004297728061676025, 0.0043038721084594726, 0.004288512229919434, 0.00429363203048706, 0.004288512229919434, 0.0042854399681091304, 0.004280320167541504, 0.0043203201293945314, 0.00431712007522583, 0.004263936042785644, 0.004253695964813233, 0.0042444801330566405, 0.004258815765380859, 0.004267039775848388, 0.00425980806350708, 0.004276224136352539, 0.004326399803161621, 0.004260863780975342, 0.004256768226623535, 0.004269055843353272, 0.004250624179840088, 0.004279295921325684, 0.004265984058380127, 0.004255743980407714, 0.004284416198730469, 0.004254720211029053, 0.004272128105163574, 0.004245503902435303, 0.004275199890136719, 0.004249599933624268, 0.004259840011596679, 0.004284416198730469, 0.004279295921325684, 0.004283391952514649, 0.004265984058380127, 0.00425164794921875, 0.004270080089569092, 0.004280320167541504, 0.004264959812164307, 0.004253695964813233, 0.004261888027191162, 0.004256768226623535, 0.004249599933624268, 0.004259840011596679, 0.004253695964813233, 0.004254720211029053, 0.004263936042785644, 0.004262911796569824, 0.004248576164245605, 0.004254720211029053, 0.004304895877838135, 0.0043745279312133786, 0.0043089919090271, 0.00430182409286499, 0.004270080089569092, 0.004260863780975342, 0.004232192039489746, 0.004256768226623535, 0.00425164794921875, 0.004245503902435303, 0.004260863780975342, 0.009439231872558594, 0.004265984058380127, 0.004242432117462158, 0.004250624179840088, 0.004245503902435303, 0.004249599933624268, 0.004265984058380127, 0.004236288070678711, 0.004234240055084228, 0.004258815765380859, 0.0043089919090271, 0.0043407359123229985, 0.0043079681396484375, 0.004327424049377441, 0.0043069438934326175, 0.004407296180725098, 0.004354047775268555, 0.004262911796569824, 0.004246528148651123, 0.004245503902435303, 0.004261888027191162, 0.004250624179840088, 0.004254720211029053, 0.00425164794921875, 0.004272128105163574, 0.0042475519180297855, 0.004248576164245605, 0.004237311840057373, 0.00425267219543457, 0.004250624179840088, 0.004253695964813233, 0.004259871959686279, 0.004250592231750489, 0.004246528148651123, 0.004256768226623535, 0.004253695964813233, 0.004249599933624268, 0.004250624179840088, 0.004250624179840088, 0.004298751831054688, 0.004283391952514649, 0.004248576164245605, 0.004232192039489746, 0.004238336086273193, 0.004233215808868408, 0.004234303951263427, 0.004260799884796143, 0.004240384101867676, 0.004235263824462891, 0.004235263824462891, 0.004231167793273926, 0.004240384101867676, 0.004242432117462158, 0.004294655799865723, 0.004254720211029053, 0.004259871959686279, 0.004272096157073975, 0.004257791996002197, 0.004254720211029053, 0.004253695964813233, 0.004254720211029053, 0.004270080089569092, 0.0042741761207580565, 0.009589759826660157, 0.004289535999298095, 0.004291584014892578, 0.004290559768676758, 0.004269055843353272, 0.004240384101867676, 0.0042444801330566405, 0.004273151874542237, 0.004385791778564453, 0.004514815807342529, 0.004255743980407714, 0.004270080089569092, 0.004241407871246338, 0.004249599933624268, 0.004262911796569824, 0.004253695964813233, 0.004267007827758789, 0.004302847862243653, 0.004254720211029053, 0.00425164794921875, 0.004250624179840088, 0.004254720211029053, 0.004257791996002197, 0.004314112186431885, 0.004271103858947754, 0.004260863780975342, 0.004271103858947754, 0.00430182409286499, 0.004269055843353272, 0.004238336086273193, 0.004253727912902832, 0.004244448184967041, 0.004298751831054688, 0.004241407871246338, 0.004231167793273926, 0.004250656127929688, 0.004246496200561524, 0.004229119777679443, 0.004265984058380127, 0.004235263824462891, 0.004240384101867676, 0.004245503902435303, 0.004263936042785644, 0.004248576164245605, 0.00425267219543457, 0.004261888027191162, 0.004282368183135986, 0.004275199890136719, 0.004297728061676025, 0.004245503902435303, 0.0042475519180297855, 0.004269055843353272, 0.004250624179840088, 0.004355072021484375, 0.004334591865539551, 0.00429363203048706, 0.004276224136352539, 0.004269055843353272, 0.004264959812164307, 0.0042444801330566405, 0.004259840011596679, 0.004256768226623535, 0.004261888027191162, 0.009488384246826171, 0.004260863780975342, 0.004225024223327637, 0.004245503902435303, 0.0047554559707641605, 0.004586495876312256, 0.004453375816345215, 0.004427775859832763, 0.004620287895202637, 0.004905983924865722, 0.004455423831939697, 0.004432896137237549, 0.004432896137237549, 0.0044421119689941405, 0.004451327800750733, 0.004447231769561768, 0.004425727844238282, 0.004492288112640381, 0.00445849609375, 0.004445184230804444, 0.004460544109344483, 0.004426752090454102, 0.004304895877838135, 0.00431001615524292, 0.00429260778427124, 0.004283391952514649, 0.004305920124053955, 0.004428800106048584, 0.004477952003479004, 0.004412415981292724, 0.00434278392791748, 0.0043120961189270016, 0.004302815914154053, 0.004313087940216064, 0.004270080089569092, 0.004345856189727783, 0.0043089919090271, 0.0043038721084594726, 0.0043079681396484375, 0.0043069438934326175, 0.004333568096160889, 0.00430079984664917, 0.00430182409286499, 0.004296703815460205, 0.004299776077270508, 0.004298783779144287, 0.004301792144775391, 0.004268032073974609, 0.004314112186431885, 0.004454400062561035, 0.004443136215209961, 0.004363264083862305, 0.004384768009185791, 0.004405248165130615, 0.004526080131530762, 0.004476928234100342, 0.004429823875427246, 0.004429823875427246, 0.004399104118347168, 0.004415487766265869, 0.00440934419631958, 0.004429823875427246, 0.004431871891021728]",tokens/s,227.93350164187387,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,google/gemma-2b,google/gemma-2b,cuda,0,42,,,True,,,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.223-212.873.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.0,,,,1.21.4,,,,0.12.0,,,gemma,MB,1352.482816,12778.471424,0.0,12138.315776,10311.21664,s,10,0.8171145324707031,0.08171145324707031,0.000294027769370703,0.08162539291381836,0.08211953735351563,0.0821870002746582,0.08224097061157226,"[0.08225446319580078, 0.08152393341064453, 0.08210454559326172, 0.08187353515625, 0.08153174591064453, 0.08158211517333984, 0.08188502502441407, 0.08139276885986328, 0.08166867065429688, 0.0812977294921875]",tokens/s,3132.9757313939176,kWh,9.966263269841565e-07,5.460741374583997e-07,6.989319316941198e-06,8.532019781383755e-06,tokens/kWh,30004618.667032786,MB,1388.716032,12784.76288,0.0,12144.607232,10311.2192,s,10,15.36981787109375,1.5369817871093752,0.0017074170528178286,1.5365200195312498,1.5389403442382812,1.539103106689453,1.5392333166503906,"[1.5357572021484376, 1.535958251953125, 1.53530126953125, 1.5350706787109376, 1.5348577880859375, 1.539265869140625, 1.5387503662109374, 1.5389041748046874, 1.5388704833984375, 1.537081787109375]",tokens/s,40.989425202288864,kWh,1.8100047610723435e-05,9.918975342816827e-06,7.573158019305863e-05,0.00010375060314659888,tokens/kWh,607225.3855814355,,s,630,15.368101884841911,0.02439381251562209,0.0001430601343260939,0.024365071296691895,0.024468786430358887,0.024661965560913088,0.025054679660797118,"[0.025093120574951173, 0.025062400817871092, 0.024408063888549804, 0.024343551635742186, 0.02433024024963379, 0.02431078338623047, 0.02430259132385254, 0.024300544738769532, 0.02430771255493164, 0.024384511947631835, 0.02431795120239258, 0.02431488037109375, 0.02429849624633789, 0.02428108787536621, 0.024300544738769532, 0.024307775497436523, 0.024266687393188477, 0.024266752243041992, 0.024258560180664062, 0.024300544738769532, 0.02427801513671875, 0.02428620719909668, 0.024292352676391602, 0.02430771255493164, 0.024349695205688478, 0.024321023941040038, 0.02430975914001465, 0.02428211212158203, 0.02428006362915039, 0.024270847320556642, 0.024275968551635742, 0.02429644775390625, 0.024368127822875976, 0.024373247146606446, 0.024375295639038085, 0.024358911514282225, 0.024382463455200197, 0.024379392623901368, 0.024363008499145508, 0.024758272171020508, 0.024450048446655274, 0.024381439208984376, 0.024374271392822267, 0.024376319885253905, 0.024365055084228517, 0.024377344131469726, 0.024394752502441407, 0.024382463455200197, 0.024378368377685547, 0.02434867286682129, 0.024360960006713867, 0.024361984252929687, 0.024370176315307617, 0.02451456069946289, 0.02453094482421875, 0.024379392623901368, 0.024375295639038085, 0.024360960006713867, 0.024403968811035157, 0.024374271392822267, 0.02436403274536133, 0.024360960006713867, 0.02434252738952637, 0.025257984161376954, 0.024645631790161132, 0.024433664321899414, 0.024353792190551758, 0.024329216003417968, 0.02430259132385254, 0.024328191757202147, 0.0243056640625, 0.02430463981628418, 0.02428313636779785, 0.02429132843017578, 0.024333311080932618, 0.02430975914001465, 0.02430771255493164, 0.024311807632446288, 0.024374271392822267, 0.024381439208984376, 0.02429439926147461, 0.0242872314453125, 0.024337408065795898, 0.0243056640625, 0.024308736801147462, 0.024439807891845702, 0.02427289581298828, 0.024266752243041992, 0.024333311080932618, 0.024378368377685547, 0.024327167510986326, 0.024368127822875976, 0.02430463981628418, 0.024319999694824217, 0.0243056640625, 0.024384511947631835, 0.024468479156494142, 0.02433843231201172, 0.024370176315307617, 0.024353792190551758, 0.02433433532714844, 0.0243507194519043, 0.024370176315307617, 0.02436403274536133, 0.024351743698120116, 0.024358911514282225, 0.024345600128173828, 0.024341503143310548, 0.02434867286682129, 0.024359935760498046, 0.02434662437438965, 0.0243558406829834, 0.024377344131469726, 0.024374271392822267, 0.024370176315307617, 0.024442880630493165, 0.024412160873413087, 0.024499200820922853, 0.024441856384277344, 0.024385536193847656, 0.024371200561523438, 0.02471833610534668, 0.024407039642333983, 0.024371200561523438, 0.0243507194519043, 0.024419328689575196, 0.025017343521118163, 0.02472550392150879, 0.024460287094116212, 0.024366079330444337, 0.02432614326477051, 0.024341503143310548, 0.024308736801147462, 0.02430463981628418, 0.02429849624633789, 0.024327167510986326, 0.024333311080932618, 0.02430463981628418, 0.024335359573364256, 0.02431795120239258, 0.02429849624633789, 0.02429132843017578, 0.02426470375061035, 0.024259584426879883, 0.02427187156677246, 0.024263679504394533, 0.02428825569152832, 0.024275968551635742, 0.02427801513671875, 0.0242872314453125, 0.02428620719909668, 0.024292352676391602, 0.02429747200012207, 0.02429747200012207, 0.02430668830871582, 0.02428108787536621, 0.02426982307434082, 0.02427903938293457, 0.02434867286682129, 0.024357887268066408, 0.024846431732177734, 0.024430496215820312, 0.024366079330444337, 0.024386560440063477, 0.024376319885253905, 0.02436403274536133, 0.02434662437438965, 0.024344575881958007, 0.024371200561523438, 0.024468479156494142, 0.024360960006713867, 0.02437222480773926, 0.024468479156494142, 0.024365087509155274, 0.024432607650756836, 0.024376319885253905, 0.024344575881958007, 0.02434867286682129, 0.0243507194519043, 0.024345600128173828, 0.024440832138061523, 0.024401920318603516, 0.02436403274536133, 0.024378368377685547, 0.02444697570800781, 0.024379392623901368, 0.024368127822875976, 0.024375295639038085, 0.024353792190551758, 0.02490777587890625, 0.024656896591186524, 0.02443059158325195, 0.024360960006713867, 0.02431078338623047, 0.02431283187866211, 0.02429542350769043, 0.02430259132385254, 0.02428825569152832, 0.02432204818725586, 0.024790016174316407, 0.024361984252929687, 0.02430668830871582, 0.02429542350769043, 0.024292352676391602, 0.024285184860229493, 0.024275968551635742, 0.024293376922607423, 0.024501247406005858, 0.02430668830871582, 0.024303615570068358, 0.0243507194519043, 0.02430463981628418, 0.02428313636779785, 0.02427801513671875, 0.024267776489257813, 0.02429030418395996, 0.02429132843017578, 0.024268800735473633, 0.024258560180664062, 0.024284160614013672, 0.024293376922607423, 0.024360960006713867, 0.02435481643676758, 0.024414207458496092, 0.024377344131469726, 0.024360960006713867, 0.024335359573364256, 0.0243507194519043, 0.024344575881958007, 0.02435686492919922, 0.0243558406829834, 0.024343551635742186, 0.024373247146606446, 0.024368127822875976, 0.024392704010009765, 0.024359935760498046, 0.02434764862060547, 0.024365055084228517, 0.024344575881958007, 0.024344575881958007, 0.024374271392822267, 0.024374271392822267, 0.02467737579345703, 0.024365055084228517, 0.024391679763793944, 0.024367103576660155, 0.024368127822875976, 0.024351743698120116, 0.024336383819580077, 0.024358911514282225, 0.024363008499145508, 0.02435481643676758, 0.024871936798095705, 0.02462822341918945, 0.024418304443359375, 0.02432512092590332, 0.0243189754486084, 0.0243056640625, 0.02430156707763672, 0.02432307243347168, 0.024328191757202147, 0.02431078338623047, 0.02430668830871582, 0.02428927993774414, 0.02427801513671875, 0.024292352676391602, 0.02429849624633789, 0.024275968551635742, 0.02427494430541992, 0.0243240966796875, 0.02429439926147461, 0.024285184860229493, 0.024266752243041992, 0.02428006362915039, 0.024293376922607423, 0.02428006362915039, 0.02428006362915039, 0.02428825569152832, 0.0242739200592041, 0.02431488037109375, 0.02427801513671875, 0.024666112899780275, 0.024341503143310548, 0.02438860893249512, 0.024410112380981445, 0.024460287094116212, 0.02449715232849121, 0.024351743698120116, 0.024352767944335937, 0.02435686492919922, 0.024353792190551758, 0.02433433532714844, 0.024371200561523438, 0.024377344131469726, 0.0243558406829834, 0.02437222480773926, 0.02437222480773926, 0.024386560440063477, 0.024349695205688478, 0.024367103576660155, 0.02433843231201172, 0.024366079330444337, 0.024345600128173828, 0.024376319885253905, 0.024365055084228517, 0.0243558406829834, 0.02438041687011719, 0.024371200561523438, 0.02437222480773926, 0.0243507194519043, 0.024443904876708986, 0.024343551635742186, 0.024363008499145508, 0.02438041687011719, 0.02446233558654785, 0.02508697509765625, 0.024648704528808595, 0.02444697570800781, 0.024436735153198243, 0.02469990348815918, 0.024885248184204102, 0.024411136627197266, 0.02433843231201172, 0.0244715518951416, 0.024359935760498046, 0.024357887268066408, 0.024429567337036134, 0.02434662437438965, 0.024369152069091796, 0.024361984252929687, 0.02429747200012207, 0.02430156707763672, 0.024319999694824217, 0.024311807632446288, 0.02429644775390625, 0.0243056640625, 0.024285184860229493, 0.02431488037109375, 0.024450048446655274, 0.024321023941040038, 0.024375295639038085, 0.024361984252929687, 0.02433945655822754, 0.02429644775390625, 0.024303615570068358, 0.02431385612487793, 0.024300544738769532, 0.024382463455200197, 0.024429567337036134, 0.024458240509033204, 0.024441856384277344, 0.024458240509033204, 0.02443059158325195, 0.02443059158325195, 0.024369152069091796, 0.024395776748657227, 0.024414207458496092, 0.024370176315307617, 0.024367103576660155, 0.024450048446655274, 0.024395776748657227, 0.024435712814331056, 0.024823808670043947, 0.024450048446655274, 0.024444927215576173, 0.024417280197143554, 0.024392704010009765, 0.024793088912963866, 0.02453094482421875, 0.02455244827270508, 0.024449024200439453, 0.024396799087524415, 0.024357887268066408, 0.024387584686279298, 0.02437222480773926, 0.02446233558654785, 0.024402944564819336, 0.02448076820373535, 0.025224191665649414, 0.024747007369995116, 0.024483840942382814, 0.024357887268066408, 0.024360960006713867, 0.024336383819580077, 0.02433945655822754, 0.02467430305480957, 0.024383487701416014, 0.024414207458496092, 0.0243558406829834, 0.024376319885253905, 0.024435712814331056, 0.024343551635742186, 0.02429439926147461, 0.024284160614013672, 0.024293376922607423, 0.0243558406829834, 0.024329216003417968, 0.02433024024963379, 0.024303615570068358, 0.024361984252929687, 0.024416255950927734, 0.024823808670043947, 0.024370176315307617, 0.024293376922607423, 0.02429849624633789, 0.024376319885253905, 0.024412160873413087, 0.024336416244506835, 0.02435273551940918, 0.02433433532714844, 0.02447667121887207, 0.024422399520874022, 0.024389631271362306, 0.024379392623901368, 0.024426496505737305, 0.024399871826171874, 0.024389631271362306, 0.024394752502441407, 0.024389631271362306, 0.024433664321899414, 0.024429567337036134, 0.02449715232849121, 0.024396799087524415, 0.024415231704711913, 0.024389631271362306, 0.024419328689575196, 0.024461311340332033, 0.024415231704711913, 0.024415231704711913, 0.024401920318603516, 0.024406015396118166, 0.024525823593139647, 0.024376319885253905, 0.024427520751953126, 0.024411136627197266, 0.02450022315979004, 0.024466432571411133, 0.024413183212280275, 0.02444697570800781, 0.024423423767089843, 0.024440832138061523, 0.025456640243530275, 0.024698879241943358, 0.024491008758544923, 0.024498176574707032, 0.024417280197143554, 0.0243507194519043, 0.02430259132385254, 0.024328191757202147, 0.024351743698120116, 0.024344575881958007, 0.024361984252929687, 0.024439807891845702, 0.024406015396118166, 0.02431078338623047, 0.024351743698120116, 0.024333311080932618, 0.024343551635742186, 0.02429849624633789, 0.02430771255493164, 0.024327167510986326, 0.024385536193847656, 0.024311807632446288, 0.024358911514282225, 0.024369152069091796, 0.024376319885253905, 0.024276992797851563, 0.02429952049255371, 0.02429030418395996, 0.02433126449584961, 0.02429747200012207, 0.0243189754486084, 0.02431283187866211, 0.024426496505737305, 0.024366079330444337, 0.024457216262817383, 0.024434688568115235, 0.024358911514282225, 0.025035776138305665, 0.02444697570800781, 0.0245483512878418, 0.024457216262817383, 0.024802303314208983, 0.02451148796081543, 0.024397823333740236, 0.024447999954223632, 0.024390655517578123, 0.024376319885253905, 0.024376319885253905, 0.024457216262817383, 0.024424448013305664, 0.024450048446655274, 0.024420352935791017, 0.024378368377685547, 0.024379392623901368, 0.024385536193847656, 0.024389631271362306, 0.024390655517578123, 0.02450432014465332, 0.024389631271362306, 0.024370176315307617, 0.024429567337036134, 0.024466432571411133, 0.024407039642333983, 0.025586687088012695, 0.024697856903076174, 0.02449407958984375, 0.02448486328125, 0.02433945655822754, 0.02434867286682129, 0.024385536193847656, 0.02435686492919922, 0.024371200561523438, 0.024395776748657227, 0.024370176315307617, 0.024353792190551758, 0.02429644775390625, 0.024351743698120116, 0.024465408325195313, 0.024382463455200197, 0.02430156707763672, 0.02462822341918945, 0.024457216262817383, 0.024419328689575196, 0.02438860893249512, 0.024308736801147462, 0.024266752243041992, 0.02429747200012207, 0.02432307243347168, 0.02429132843017578, 0.0243189754486084, 0.02433945655822754, 0.024328191757202147, 0.02431283187866211, 0.02433228874206543, 0.024366079330444337, 0.024411136627197266, 0.0243558406829834, 0.024358911514282225, 0.024353792190551758, 0.024397823333740236, 0.024422399520874022, 0.024396799087524415, 0.024976383209228514, 0.024597503662109374, 0.024433664321899414, 0.024365055084228517, 0.02437222480773926, 0.024379392623901368, 0.024392704010009765, 0.02437222480773926, 0.024360960006713867, 0.024406015396118166, 0.024403968811035157, 0.024461311340332033, 0.024421375274658205, 0.024414207458496092, 0.024378368377685547, 0.024393728256225586, 0.024424448013305664, 0.024407039642333983, 0.024426496505737305, 0.024370176315307617, 0.024403968811035157, 0.024607744216918945, 0.024456192016601562, 0.024409088134765625, 0.025010175704956054, 0.0247511043548584, 0.02452377510070801, 0.024345600128173828, 0.024404991149902345, 0.024352767944335937, 0.024461311340332033, 0.024328191757202147, 0.0243189754486084, 0.024369152069091796, 0.024422399520874022, 0.02435481643676758, 0.0243240966796875, 0.024381439208984376, 0.024373247146606446, 0.02426982307434082, 0.02429132843017578, 0.02428211212158203, 0.024352767944335937, 0.02430771255493164, 0.02431078338623047, 0.024311807632446288, 0.0243189754486084, 0.024319999694824217, 0.024442880630493165, 0.024292352676391602, 0.024284160614013672, 0.024270847320556642, 0.0242872314453125, 0.024308736801147462, 0.024436735153198243, 0.024432640075683593, 0.024369152069091796, 0.024459264755249024, 0.024435712814331056, 0.024395776748657227, 0.024772607803344726, 0.024439807891845702, 0.024370176315307617, 0.024434688568115235, 0.024404991149902345, 0.02436403274536133, 0.024422399520874022, 0.02448793601989746, 0.024357887268066408, 0.024353792190551758, 0.0243507194519043, 0.02438860893249512, 0.024375295639038085, 0.024370176315307617, 0.024371200561523438, 0.024447999954223632, 0.024400896072387695, 0.02449612808227539, 0.02438860893249512, 0.024370176315307617, 0.02438041687011719, 0.02436403274536133, 0.024349695205688478, 0.024375295639038085, 0.024385536193847656, 0.024383487701416014, 0.024387584686279298]",tokens/s,40.9940020388198,,,,,,MB,934.436864,12753.3056,0.0,12123.635712,12121.851904,s,1,6.919205078125,6.919205078125,0.0,6.919205078125,6.919205078125,6.919205078125,6.919205078125,[6.919205078125],,kWh,1.8626668020784918e-06,9.847741043013395e-07,3.842503074003023e-06,6.689943980382855e-06, float32-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,s,s,cuda,0,42,,,True,True,True,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/s/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6694735c-34752f973fd1e62729838657;8cd1b69a-5f46-4541-a0cf-0a08b36db210) Repository Not Found for url: https://huggingface.co/s/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: s is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-40b,tiiuae/falcon-40b,cuda,0,42,,,True,True,True,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ self.load_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 250, in load_model_with_no_weights self.pretrained_model = self.automodel_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 559, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3704, in from_pretrained config = cls._autoset_attn_implementation( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1490, in _autoset_attn_implementation config = cls._check_and_enable_sdpa( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1656, in _check_and_enable_sdpa raise ValueError( ValueError: FalconForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,google/gemma-7b,google/gemma-7b,cuda,0,42,,,True,,,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",gemma,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-125m,facebook/opt-125m,cuda,0,42,,,True,,,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",opt,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,databricks/dbrx-base,databricks/dbrx-base,cuda,0,42,,,True,,,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",dbrx,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-350m,facebook/opt-350m,cuda,0,42,,,True,,,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",opt,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,-,-,cuda,0,42,,,True,True,True,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 106, in _inner_fn validate_repo_id(arg_value) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 160, in validate_repo_id raise HFValidationError( huggingface_hub.errors.HFValidationError: Repo id must use alphanumeric chars or '-', '_', '.', '--' and '..' are forbidden, '-' and '.' cannot start or end the name, max length is 96: '-'. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 466, in cached_file raise EnvironmentError( OSError: Incorrect path_or_model_id: '-'. Please provide either the path to a local folder or the repo_id of a model on the Hub. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,stabilityai/stablelm-2-12b,stabilityai/stablelm-2-12b,cuda,0,42,,,True,True,True,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ self.load_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 250, in load_model_with_no_weights self.pretrained_model = self.automodel_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3710, in from_pretrained model = cls(config, *model_args, **model_kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/stablelm/modeling_stablelm.py"", line 1188, in __init__ self.model = StableLmModel(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/stablelm/modeling_stablelm.py"", line 966, in __init__ [StableLmDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/stablelm/modeling_stablelm.py"", line 966, in [StableLmDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/stablelm/modeling_stablelm.py"", line 748, in __init__ self.mlp = StableLmMLP(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/stablelm/modeling_stablelm.py"", line 204, in __init__ self.gate_proj = nn.Linear(self.hidden_size, self.intermediate_size, bias=False) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/linear.py"", line 98, in __init__ self.weight = Parameter(torch.empty((out_features, in_features), **factory_kwargs)) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_device.py"", line 78, in __torch_function__ return func(*args, **kwargs) torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 270.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,v,v,cuda,0,42,,,True,True,True,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/v/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66932568-005036bd6322c2c50deee463;bc5166fb-1326-4450-bfc0-2e90f0113859) Repository Not Found for url: https://huggingface.co/v/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: v is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,openai-community/gpt2,openai-community/gpt2,cuda,0,42,,,True,True,True,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.42.3,,0.31.0,,,,1.20.0,,,,0.11.1,,,,MB,1053.3888,1420.296192,0.0,780.140544,601.373184,s,25,0.18852665615081787,0.007541066246032715,0.00015444865038906702,0.0075185918807983395,0.007637945556640625,0.00771103343963623,0.008090321884155273,"[0.00820633602142334, 0.007599775791168213, 0.007441728115081787, 0.0074460477828979495, 0.007518847942352295, 0.007422624111175537, 0.007475488185882569, 0.007544640064239502, 0.007455904006958008, 0.007430912017822266, 0.007444352149963379, 0.0075185918807983395, 0.007453055858612061, 0.007586400032043457, 0.007526559829711914, 0.007479008197784424, 0.0074737281799316406, 0.007440703868865967, 0.0077229437828063964, 0.007541664123535156, 0.007537024021148682, 0.007561823844909668, 0.007663392066955567, 0.0074661440849304195, 0.007568960189819336]",tokens/s,33947.45406654917,kWh,8.757900842164379e-08,4.798957402127395e-08,5.078331630407976e-07,6.434017454837154e-07,tokens/kWh,397885149.98127156,MB,1053.3888,1420.296192,0.0,780.140544,601.375744,s,25,9.990148040771484,0.3996059216308594,0.006300418674750675,0.3998371887207031,0.4023217956542969,0.4038839172363281,0.4198370434570312,"[0.424822265625, 0.39957925415039064, 0.39524063110351565, 0.3914546203613281, 0.4009781494140625, 0.39893215942382815, 0.3998371887207031, 0.4008005065917969, 0.40078765869140626, 0.3996658020019531, 0.3878934020996094, 0.40016326904296873, 0.39710504150390624, 0.4006253967285156, 0.399701171875, 0.4005265197753906, 0.40069314575195314, 0.400043212890625, 0.39953118896484374, 0.4032175598144531, 0.40060714721679686, 0.3923962707519531, 0.40405050659179687, 0.39701882934570315, 0.3944771423339844]",tokens/s,157.65532137983925,kWh,4.630829495936061e-06,2.537316286878802e-06,1.0025348340420571e-05,1.7193494123235434e-05,tokens/kWh,3664176.6675489927,,s,1574,10.164045824050913,0.006457462404098414,0.0009624600741983297,0.006333487987518311,0.006491852712631226,0.006745907306671142,0.0139994624710083,"[0.007158783912658692, 0.007185535907745361, 0.007174015998840332, 0.007103487968444824, 0.006779903888702392, 0.006807551860809326, 0.006730751991271973, 0.006762495994567871, 0.006638591766357422, 0.006540287971496582, 0.006519807815551758, 0.006601727962493896, 0.006623263835906982, 0.006700128078460693, 0.006765439987182618, 0.0067041277885437015, 0.0066109437942504885, 0.006616064071655273, 0.006682720184326172, 0.006706143856048584, 0.006728640079498291, 0.006717440128326416, 0.00662937593460083, 0.006676479816436768, 0.0068853759765625, 0.006685696125030518, 0.006526976108551025, 0.006557695865631104, 0.006550528049468994, 0.006535168170928955, 0.006511616230010986, 0.006625279903411865, 0.007200767993927002, 0.006907936096191406, 0.006951903820037842, 0.006940671920776367, 0.006894591808319092, 0.006841343879699707, 0.0068249602317810056, 0.006790143966674805, 0.006680575847625733, 0.006609920024871826, 0.006515711784362793, 0.0065781760215759275, 0.0066078720092773435, 0.0068853759765625, 0.006841343879699707, 0.006915071964263916, 0.006622208118438721, 0.00684441614151001, 0.0066078720092773435, 0.006562848091125489, 0.0066928319931030275, 0.00669593620300293, 0.00667852783203125, 0.006738944053649902, 0.006591487884521485, 0.006650879859924317, 0.006578271865844727, 0.006761375904083252, 0.006756351947784424, 0.006729728221893311, 0.014393343925476074, 0.006553599834442139, 0.006392831802368164, 0.006140927791595459, 0.006153215885162353, 0.006142975807189942, 0.006149119853973388, 0.006170720100402832, 0.006630303859710693, 0.006564864158630371, 0.006562816143035889, 0.006796288013458252, 0.0070563840866088865, 0.0064839677810668945, 0.006450175762176514, 0.0063907837867736815, 0.00638976001739502, 0.006456319808959961, 0.006425600051879882, 0.00638976001739502, 0.006947840213775635, 0.006441984176635742, 0.006421504020690918, 0.0063272957801818845, 0.0064245758056640625, 0.006367231845855713, 0.006289408206939697, 0.006421504020690918, 0.006326272010803223, 0.006386688232421875, 0.006336512088775635, 0.006276095867156982, 0.006309887886047363, 0.006311935901641846, 0.006301695823669433, 0.006275072097778321, 0.006304768085479737, 0.006345727920532227, 0.006301695823669433, 0.006606847763061524, 0.00637440013885498, 0.006338560104370118, 0.006449151992797852, 0.006411263942718506, 0.006347775936126709, 0.006311935901641846, 0.006106112003326416, 0.006102015972137451, 0.006092895984649658, 0.006089632034301758, 0.006072319984436035, 0.006211584091186524, 0.006049791812896729, 0.006118400096893311, 0.0060702719688415525, 0.0061337599754333495, 0.0060661759376525876, 0.006137856006622314, 0.0060928001403808595, 0.00619212818145752, 0.006338560104370118, 0.006387712001800537, 0.006392831802368164, 0.014010368347167968, 0.006355967998504639, 0.006358016014099121, 0.006368256092071533, 0.0063201279640197755, 0.006346752166748047, 0.006377471923828125, 0.006323200225830078, 0.006831103801727295, 0.006316031932830811, 0.006333471775054932, 0.006378464221954346, 0.006303743839263916, 0.006899712085723877, 0.0065669121742248536, 0.006326272010803223, 0.006354944229125976, 0.006323200225830078, 0.0063610877990722655, 0.006347775936126709, 0.006375423908233643, 0.006344704151153564, 0.006347775936126709, 0.0063498239517211915, 0.006301695823669433, 0.006303743839263916, 0.006325247764587402, 0.006351871967315674, 0.0063170561790466305, 0.006348800182342529, 0.006312960147857666, 0.006339583873748779, 0.006330368041992188, 0.006323200225830078, 0.006089727878570556, 0.00612556791305542, 0.006086656093597412, 0.006127615928649902, 0.006135807991027832, 0.00618393611907959, 0.006144000053405762, 0.006106112003326416, 0.006194176197052002, 0.006138879776000977, 0.0062740478515625, 0.006095871925354004, 0.006142975807189942, 0.0061296639442443845, 0.006144000053405762, 0.006069248199462891, 0.006162432193756104, 0.006093823909759521, 0.006141952037811279, 0.006107264041900635, 0.0061049599647521975, 0.006081535816192627, 0.006094848155975342, 0.006103040218353272, 0.006102015972137451, 0.006121471881866455, 0.006209536075592041, 0.006250495910644531, 0.006465536117553711, 0.01382198429107666, 0.006111199855804443, 0.006151167869567871, 0.006100992202758789, 0.006123680114746094, 0.0061060161590576175, 0.0060845441818237304, 0.006108160018920898, 0.006095871925354004, 0.00611737585067749, 0.006155263900756836, 0.006108160018920898, 0.0061337599754333495, 0.006110239982604981, 0.006078432083129883, 0.006636576175689697, 0.006080480098724366, 0.006098944187164307, 0.006114304065704346, 0.006124544143676758, 0.0060999999046325686, 0.006070240020751953, 0.006132736206054688, 0.006116352081298828, 0.006090752124786377, 0.006067327976226807, 0.006171520233154297, 0.00620851182937622, 0.006179840087890625, 0.006198272228240966, 0.006437888145446777, 0.006264832019805908, 0.0063508481979370115, 0.0063201279640197755, 0.006072351932525635, 0.006109151840209961, 0.006055935859680176, 0.006097919940948486, 0.006072319984436035, 0.006139904022216797, 0.006096896171569824, 0.0061337599754333495, 0.006202367782592774, 0.006103040218353272, 0.006131711959838867, 0.006108160018920898, 0.006104063987731933, 0.006396927833557129, 0.006354944229125976, 0.006302720069885254, 0.006359039783477783, 0.006379519939422608, 0.007003136157989502, 0.006412288188934326, 0.0063099517822265624, 0.006379456043243408, 0.006301695823669433, 0.006330368041992188, 0.006366208076477051, 0.006301695823669433, 0.006405119895935059, 0.006326272010803223, 0.006398975849151611, 0.014101504325866699, 0.006304768085479737, 0.0063569917678833006, 0.00636518383026123, 0.006434815883636475, 0.006346752166748047, 0.006322175979614258, 0.006364160060882569, 0.006375423908233643, 0.006388735771179199, 0.006318079948425293, 0.006343679904937744, 0.006393919944763184, 0.006326208114624024, 0.006358016014099121, 0.006305791854858398, 0.006304768085479737, 0.006428671836853027, 0.006328320026397705, 0.00633241605758667, 0.00631715202331543, 0.0063405117988586425, 0.006303743839263916, 0.006315008163452148, 0.00638156795501709, 0.006300672054290772, 0.006329343795776367, 0.006386688232421875, 0.006322175979614258, 0.006329343795776367, 0.006333439826965332, 0.006377471923828125, 0.006298624038696289, 0.006367263793945313, 0.006322144031524658, 0.006340608119964599, 0.006391808032989502, 0.006423552036285401, 0.006362112045288086, 0.006815743923187256, 0.006351871967315674, 0.006491136074066162, 0.006404096126556396, 0.006392831802368164, 0.006393856048583985, 0.006323200225830078, 0.006376512050628662, 0.00630675220489502, 0.006335584163665772, 0.00629033613204956, 0.006308896064758301, 0.006367199897766114, 0.006277120113372803, 0.006376448154449463, 0.006302720069885254, 0.006346752166748047, 0.0064778242111206055, 0.006284287929534912, 0.0064204797744750975, 0.006433792114257812, 0.006303743839263916, 0.006461503982543945, 0.006367167949676514, 0.013757568359375, 0.006122399806976318, 0.006209504127502441, 0.006111231803894043, 0.006161407947540283, 0.006091775894165039, 0.006157311916351318, 0.006115327835083008, 0.006104191780090332, 0.006103936195373535, 0.006155263900756836, 0.006394879817962646, 0.006392831802368164, 0.006401023864746094, 0.006699103832244873, 0.00639785623550415, 0.006351871967315674, 0.0064124159812927245, 0.006336383819580078, 0.006329504013061524, 0.006325088024139404, 0.00638259220123291, 0.006313983917236328, 0.006328447818756104, 0.0063517441749572755, 0.006351871967315674, 0.00636518383026123, 0.006342656135559082, 0.006561791896820069, 0.006425600051879882, 0.006367231845855713, 0.0063610877990722655, 0.006329343795776367, 0.006364160060882569, 0.00637440013885498, 0.006335487842559814, 0.006347775936126709, 0.006354944229125976, 0.006403103828430175, 0.006304863929748535, 0.0064971518516540525, 0.006394879817962646, 0.0063805441856384275, 0.0063907837867736815, 0.006335487842559814, 0.006452223777770996, 0.006354944229125976, 0.006372352123260498, 0.006348800182342529, 0.006307839870452881, 0.006342688083648682, 0.006318048000335693, 0.006325247764587402, 0.006391808032989502, 0.006328320026397705, 0.0063499841690063475, 0.006297440052032471, 0.006323200225830078, 0.006326272010803223, 0.006335487842559814, 0.006371327877044678, 0.006395904064178467, 0.006345727920532227, 0.014003199577331543, 0.006343679904937744, 0.0063272957801818845, 0.006318111896514892, 0.006322144031524658, 0.006278207778930664, 0.006339583873748779, 0.006328256130218506, 0.006359039783477783, 0.0063201279640197755, 0.0063498239517211915, 0.006385663986206055, 0.006363135814666748, 0.006450175762176514, 0.0063498239517211915, 0.0063324799537658695, 0.006324160099029541, 0.006875135898590088, 0.006376448154449463, 0.006366208076477051, 0.006342656135559082, 0.006348800182342529, 0.0062863359451293946, 0.006406144142150879, 0.006346752166748047, 0.006336512088775635, 0.0063907837867736815, 0.006303743839263916, 0.006418432235717773, 0.0063170561790466305, 0.006323200225830078, 0.0063272957801818845, 0.0062975997924804685, 0.0063211522102355954, 0.006284287929534912, 0.006335487842559814, 0.006464511871337891, 0.0063990077972412105, 0.00638153600692749, 0.006363135814666748, 0.00636518383026123, 0.00633241605758667, 0.006369279861450195, 0.006362112045288086, 0.00638259220123291, 0.006542335987091064, 0.006351871967315674, 0.006266880035400391, 0.006149119853973388, 0.006121471881866455, 0.0061224961280822755, 0.006079487800598145, 0.006095871925354004, 0.006094848155975342, 0.0066406397819519045, 0.006220831871032715, 0.006374368190765381, 0.006448128223419189, 0.006425600051879882, 0.006372352123260498, 0.0063805441856384275, 0.006378496170043945, 0.0063508481979370115, 0.014110719680786133, 0.006492159843444824, 0.006341631889343262, 0.006342656135559082, 0.0064204797744750975, 0.006336512088775635, 0.006369279861450195, 0.006343679904937744, 0.006324351787567139, 0.006398848056793213, 0.006621183872222901, 0.006392831802368164, 0.006294528007507324, 0.00643071985244751, 0.006336512088775635, 0.006364160060882569, 0.00648089599609375, 0.006307839870452881, 0.006370304107666015, 0.006304768085479737, 0.006310912132263183, 0.006372352123260498, 0.006291456222534179, 0.006355967998504639, 0.006335487842559814, 0.006316031932830811, 0.00626585578918457, 0.006328320026397705, 0.0062975997924804685, 0.006305791854858398, 0.006329343795776367, 0.006311935901641846, 0.006331391811370849, 0.006346752166748047, 0.006289408206939697, 0.006464511871337891, 0.006313983917236328, 0.0064174079895019534, 0.006338560104370118, 0.006360064029693604, 0.0063272957801818845, 0.006301695823669433, 0.006355072021484375, 0.006388607978820801, 0.00633241605758667, 0.0063201279640197755, 0.006296576023101807, 0.00636518383026123, 0.006402048110961914, 0.006310912132263183, 0.006343679904937744, 0.006295551776885986, 0.00631820821762085, 0.0062882561683654786, 0.0066416640281677245, 0.006322175979614258, 0.006311935901641846, 0.00658739185333252, 0.006364160060882569, 0.006358016014099121, 0.006337535858154297, 0.0063907837867736815, 0.006358016014099121, 0.013998080253601074, 0.006340608119964599, 0.006363135814666748, 0.006319104194641113, 0.006388735771179199, 0.006310912132263183, 0.006377471923828125, 0.006325247764587402, 0.00636518383026123, 0.006334496021270752, 0.006306784152984619, 0.006345727920532227, 0.006352896213531494, 0.006402048110961914, 0.006379519939422608, 0.006326272010803223, 0.006408192157745361, 0.006356063842773437, 0.006373280048370361, 0.0063272957801818845, 0.006360064029693604, 0.006363135814666748, 0.0063170561790466305, 0.006415359973907471, 0.006301695823669433, 0.00638259220123291, 0.006331391811370849, 0.006300672054290772, 0.0063508481979370115, 0.0063201279640197755, 0.0063498239517211915, 0.006306816101074219, 0.006314112186431885, 0.0063619837760925295, 0.006308864116668702, 0.006359039783477783, 0.00643891191482544, 0.00632422399520874, 0.006319104194641113, 0.006308864116668702, 0.006364160060882569, 0.00637337589263916, 0.006433792114257812, 0.006323200225830078, 0.006363135814666748, 0.006347775936126709, 0.006308864116668702, 0.006948863983154297, 0.006366208076477051, 0.0063201279640197755, 0.0063508481979370115, 0.006281216144561768, 0.006322175979614258, 0.006303743839263916, 0.006407167911529541, 0.0063805441856384275, 0.006344704151153564, 0.006391808032989502, 0.0063508481979370115, 0.006313983917236328, 0.006347775936126709, 0.006419456005096436, 0.006378496170043945, 0.013956095695495606, 0.00632422399520874, 0.0063272957801818845, 0.006339583873748779, 0.006379519939422608, 0.006393856048583985, 0.006340608119964599, 0.006309887886047363, 0.006312960147857666, 0.006294528007507324, 0.006300672054290772, 0.006305888175964356, 0.006317984104156494, 0.006366208076477051, 0.006331391811370849, 0.006341728210449218, 0.006335391998291016, 0.006299647808074951, 0.006333504199981689, 0.006286272048950195, 0.006284287929534912, 0.006306816101074219, 0.006296576023101807, 0.006519807815551758, 0.006328415870666504, 0.006351776123046875, 0.006328320026397705, 0.006415359973907471, 0.006341631889343262, 0.006279295921325684, 0.006320000171661377, 0.006292479991912842, 0.006369279861450195, 0.006313983917236328, 0.006334464073181153, 0.006328320026397705, 0.006336512088775635, 0.006344704151153564, 0.0063211841583251955, 0.00633135986328125, 0.006403071880340576, 0.006347775936126709, 0.006331391811370849, 0.0062997441291809085, 0.0063968319892883305, 0.0062956161499023435, 0.006309823989868164, 0.006323200225830078, 0.006339615821838379, 0.0063385281562805176, 0.00637440013885498, 0.006301695823669433, 0.006478847980499268, 0.0063170561790466305, 0.006384640216827392, 0.006308864116668702, 0.0063508481979370115, 0.006308864116668702, 0.006278143882751465, 0.006524928092956543, 0.006360064029693604, 0.006451200008392334, 0.0063508481979370115, 0.013963264465332031, 0.006314112186431885, 0.006084479808807373, 0.00611737585067749, 0.006082560062408447, 0.0060999679565429685, 0.006100992202758789, 0.006113279819488526, 0.006149119853973388, 0.006124544143676758, 0.0060999679565429685, 0.006123519897460937, 0.0061337599754333495, 0.006127615928649902, 0.006091775894165039, 0.00612556791305542, 0.006120448112487793, 0.006098944187164307, 0.006114304065704346, 0.0060999679565429685, 0.006130688190460205, 0.006098944187164307, 0.006095871925354004, 0.006237184047698975, 0.006166528224945069, 0.0060928001403808595, 0.006109183788299561, 0.006087679862976075, 0.006091807842254639, 0.00608355188369751, 0.006255616188049316, 0.006113279819488526, 0.00608358383178711, 0.006092959880828857, 0.006047584056854248, 0.006159359931945801, 0.006073408126831055, 0.006123456001281738, 0.006104063987731933, 0.006160384178161621, 0.006093823909759521, 0.006106112003326416, 0.006080512046813965, 0.006191103935241699, 0.006108160018920898, 0.006100992202758789, 0.006088704109191895, 0.006116352081298828, 0.006128640174865723, 0.006097919940948486, 0.006093823909759521, 0.006082560062408447, 0.006166528224945069, 0.006375423908233643, 0.006330368041992188, 0.006349855899810791, 0.00633238410949707, 0.006289408206939697, 0.0063211522102355954, 0.0063201279640197755, 0.006262784004211426, 0.006366208076477051, 0.006280191898345947, 0.014090239524841308, 0.006319104194641113, 0.007004159927368164, 0.006492159843444824, 0.006312960147857666, 0.006388735771179199, 0.006351871967315674, 0.006375423908233643, 0.006801407814025879, 0.006617087841033936, 0.006343679904937744, 0.00637337589263916, 0.006319104194641113, 0.006378496170043945, 0.006966271877288818, 0.00638259220123291, 0.0064204797744750975, 0.006323328018188476, 0.0062984957695007325, 0.006330368041992188, 0.006288383960723877, 0.006343711853027343, 0.006277088165283203, 0.00635097599029541, 0.00630675220489502, 0.006292448043823242, 0.00632534408569336, 0.006466432094573975, 0.006372352123260498, 0.006300672054290772, 0.006310912132263183, 0.0063211522102355954, 0.006313983917236328, 0.006366208076477051, 0.006289408206939697, 0.006318079948425293, 0.006377471923828125, 0.0063498239517211915, 0.006336512088775635, 0.006299647808074951, 0.006313983917236328, 0.006348800182342529, 0.006325247764587402, 0.0063632321357727055, 0.006314911842346191, 0.0063201279640197755, 0.006364160060882569, 0.006358016014099121, 0.0063786239624023435, 0.006294400215148926, 0.006319104194641113, 0.006326367855072021, 0.0063640642166137695, 0.0064471039772033695, 0.006364160060882569, 0.006196224212646485, 0.00608358383178711, 0.006113279819488526, 0.006095871925354004, 0.006132736206054688, 0.006130688190460205, 0.006146048069000244, 0.006105088233947754, 0.013883392333984374, 0.006110208034515381, 0.006149119853973388, 0.006168575763702393, 0.006135807991027832, 0.006154240131378174, 0.006458367824554443, 0.006419456005096436, 0.006360064029693604, 0.006450175762176514, 0.00632422399520874, 0.006395904064178467, 0.006366208076477051, 0.007089248180389404, 0.006367136001586914, 0.006367231845855713, 0.0064102401733398436, 0.006390848159790039, 0.006321087837219238, 0.00632428789138794, 0.006352831840515136, 0.006292479991912842, 0.006292479991912842, 0.006288383960723877, 0.006351871967315674, 0.006384640216827392, 0.006306816101074219, 0.006351871967315674, 0.006286431789398193, 0.006356895923614502, 0.006363135814666748, 0.0063201279640197755, 0.006502399921417237, 0.006285312175750732, 0.006360064029693604, 0.006339583873748779, 0.006335487842559814, 0.006312960147857666, 0.006330368041992188, 0.006319104194641113, 0.006087679862976075, 0.006127679824829101, 0.006292416095733643, 0.006137856006622314, 0.006130688190460205, 0.006154240131378174, 0.006114304065704346, 0.006145023822784424, 0.006107135772705078, 0.006136832237243653, 0.006330368041992188, 0.006067200183868408, 0.006102015972137451, 0.006067200183868408, 0.006408192157745361, 0.006350944042205811, 0.006323103904724121, 0.006281216144561768, 0.006326335906982422, 0.00634771203994751, 0.0063610877990722655, 0.006481919765472412, 0.0063170561790466305, 0.014160896301269531, 0.00632422399520874, 0.006325247764587402, 0.006347775936126709, 0.006310912132263183, 0.006371327877044678, 0.0063272957801818845, 0.006377471923828125, 0.006345727920532227, 0.006393856048583985, 0.006323200225830078, 0.006296576023101807, 0.0064440321922302245, 0.00637337589263916, 0.006310912132263183, 0.006367231845855713, 0.006326303958892822, 0.006294496059417724, 0.006385663986206055, 0.006266880035400391, 0.006334464073181153, 0.006297696113586426, 0.006341536045074463, 0.006295551776885986, 0.006318079948425293, 0.006301695823669433, 0.006278143882751465, 0.0065443840026855465, 0.006315008163452148, 0.006351871967315674, 0.006527999877929688, 0.0062904319763183595, 0.006328320026397705, 0.00637337589263916, 0.006402048110961914, 0.006343711853027343, 0.0063200960159301755, 0.006347775936126709, 0.006304768085479737, 0.006319104194641113, 0.006304831981658936, 0.0064009599685668946, 0.006371327877044678, 0.006311935901641846, 0.006945792198181152, 0.00637440013885498, 0.006355967998504639, 0.006354944229125976, 0.0063580479621887205, 0.006346720218658447, 0.0063170561790466305, 0.006315008163452148, 0.006316031932830811, 0.006319104194641113, 0.006423552036285401, 0.006327424049377441, 0.006320000171661377, 0.006352896213531494, 0.006310912132263183, 0.006387712001800537, 0.00637337589263916, 0.006384640216827392, 0.0063569917678833006, 0.01408614444732666, 0.006334464073181153, 0.006312960147857666, 0.006392831802368164, 0.006348800182342529, 0.0064102401733398436, 0.006394879817962646, 0.006402048110961914, 0.0064031357765197755, 0.006316991806030273, 0.006310912132263183, 0.00637337589263916, 0.006352896213531494, 0.006358016014099121, 0.006313055992126465, 0.006336415767669678, 0.006303743839263916, 0.006328320026397705, 0.006345727920532227, 0.006358016014099121, 0.0063610877990722655, 0.006299647808074951, 0.0062863359451293946, 0.006305791854858398, 0.006293504238128662, 0.006345727920532227, 0.006411263942718506, 0.006367231845855713, 0.006340608119964599, 0.006306816101074219, 0.006318079948425293, 0.006292479991912842, 0.006319104194641113, 0.006346752166748047, 0.006323200225830078, 0.006313983917236328, 0.006337535858154297, 0.006391808032989502, 0.006303743839263916, 0.0063407039642333985, 0.0063640642166137695, 0.0063836159706115725, 0.006354944229125976, 0.0063170561790466305, 0.006395904064178467, 0.0063201279640197755, 0.006316127777099609, 0.006357920169830322, 0.0063017277717590334, 0.006343647956848144, 0.006303743839263916, 0.00643174409866333, 0.006332543849945069, 0.006295423984527588, 0.0063498239517211915, 0.0063069119453430175, 0.006380447864532471, 0.006296576023101807, 0.0063272957801818845, 0.006326272010803223, 0.006347775936126709, 0.006377471923828125, 0.006303743839263916, 0.014117888450622559, 0.006348800182342529, 0.00643174409866333, 0.006333439826965332, 0.006316031932830811, 0.006360064029693604, 0.006311935901641846, 0.0063539199829101565, 0.006328320026397705, 0.006304768085479737, 0.006360064029693604, 0.006292479991912842, 0.006308864116668702, 0.006395904064178467, 0.006336512088775635, 0.006394879817962646, 0.006311935901641846, 0.006344704151153564, 0.006830080032348633, 0.0066007041931152345, 0.006389887809753418, 0.006358975887298584, 0.006452159881591797, 0.006323200225830078, 0.006298624038696289, 0.006309887886047363, 0.006271999835968017, 0.006338560104370118, 0.00636518383026123, 0.006316031932830811, 0.006313983917236328, 0.006280191898345947, 0.006304768085479737, 0.006276095867156982, 0.006328320026397705, 0.00633241605758667, 0.0063201279640197755, 0.0063621759414672855, 0.006302656173706055, 0.006334464073181153, 0.006299647808074951, 0.006363135814666748, 0.006398975849151611, 0.006294528007507324, 0.006364160060882569, 0.006284287929534912, 0.006339615821838379, 0.006463456153869629, 0.006337535858154297, 0.006347775936126709, 0.0063170561790466305, 0.006351871967315674, 0.006302720069885254, 0.006316031932830811, 0.006364160060882569, 0.006313024044036865, 0.006370240211486816, 0.006333439826965332, 0.006335487842559814, 0.006516736030578613, 0.006364287853240967, 0.006372223854064942, 0.006368256092071533, 0.014109696388244629, 0.006354944229125976, 0.006462463855743408, 0.006344704151153564, 0.006330368041992188, 0.006337535858154297, 0.006304768085479737, 0.0063498239517211915, 0.006352896213531494, 0.006967296123504638, 0.006395904064178467, 0.006334464073181153, 0.006401023864746094, 0.0063610877990722655, 0.006305791854858398, 0.006366208076477051, 0.00632422399520874, 0.0063805441856384275, 0.006342656135559082, 0.006289408206939697, 0.006343679904937744, 0.006304768085479737, 0.006330368041992188, 0.006408192157745361, 0.006303743839263916, 0.006338560104370118, 0.0062873601913452145, 0.006411263942718506, 0.006301695823669433, 0.006489088058471679, 0.006350944042205811, 0.006288288116455078, 0.006346752166748047, 0.006323200225830078, 0.0063272957801818845, 0.006341631889343262, 0.006326272010803223, 0.0063907837867736815, 0.006336512088775635, 0.0063569917678833006, 0.0062873601913452145, 0.00632422399520874, 0.0063498239517211915, 0.006319104194641113, 0.006415359973907471, 0.006310912132263183, 0.00632528018951416, 0.00633135986328125, 0.006416384220123291, 0.006342688083648682, 0.006367263793945313, 0.006320064067840576, 0.006301695823669433, 0.006363135814666748, 0.006328320026397705, 0.006364160060882569, 0.006352928161621094, 0.006459360122680664, 0.006334464073181153, 0.006313983917236328, 0.006316031932830811, 0.0063272957801818845, 0.006310912132263183, 0.014356479644775391, 0.006428671836853027, 0.006339583873748779, 0.006329343795776367, 0.006312960147857666, 0.0063610877990722655, 0.00638976001739502, 0.006327328205108643, 0.0063426241874694825, 0.006387712001800537, 0.0063508481979370115, 0.006354944229125976, 0.0063610877990722655, 0.006363135814666748, 0.006323200225830078, 0.006334464073181153, 0.006312960147857666, 0.0063211841583251955, 0.0063190717697143555, 0.006312960147857666, 0.006378496170043945, 0.006329343795776367, 0.006330368041992188, 0.006312992095947266, 0.006376416206359863, 0.006457344055175781, 0.00633241605758667, 0.006342656135559082, 0.006311935901641846, 0.006319104194641113, 0.006421504020690918, 0.006288383960723877, 0.006358016014099121, 0.006296671867370605, 0.006348703861236572, 0.006348800182342529, 0.006337600231170654, 0.006162367820739746, 0.006102015972137451, 0.00612556791305542, 0.006142975807189942, 0.0061337599754333495, 0.006121471881866455, 0.006168575763702393, 0.006108160018920898, 0.0061296639442443845, 0.006102015972137451, 0.006151167869567871, 0.006177792072296143, 0.006559743881225586, 0.006378496170043945, 0.006305791854858398, 0.006348800182342529, 0.006306816101074219, 0.00638259220123291, 0.006401023864746094, 0.006797311782836914, 0.0064849920272827145, 0.0064542717933654785, 0.006751232147216797, 0.0070522880554199216, 0.006492159843444824, 0.006348800182342529, 0.01454092788696289, 0.006382463932037353, 0.006375423908233643, 0.006393856048583985, 0.006354944229125976, 0.006335487842559814, 0.006359039783477783, 0.006340608119964599, 0.006330368041992188, 0.006344704151153564, 0.006426623821258545, 0.006354944229125976, 0.006336512088775635, 0.006394879817962646, 0.0063089919090271, 0.00635481595993042, 0.006304768085479737, 0.006358016014099121, 0.006311935901641846, 0.006323200225830078, 0.006348800182342529, 0.006323264122009277, 0.006324160099029541, 0.006319104194641113, 0.0063201279640197755, 0.006432864189147949, 0.006322080135345459, 0.006325247764587402, 0.006304768085479737, 0.0063610877990722655, 0.006315008163452148, 0.006331391811370849, 0.0063539199829101565, 0.006461503982543945, 0.006154176235198975, 0.006088704109191895, 0.006115327835083008, 0.006088704109191895, 0.006100063800811768, 0.006924191951751709, 0.006343679904937744, 0.006405119895935059, 0.006359072208404541, 0.006334432125091553, 0.006373439788818359, 0.006297535896301269, 0.006340608119964599, 0.00638156795501709, 0.006322207927703857, 0.006290400028228759, 0.006309887886047363, 0.00636518383026123, 0.006288383960723877, 0.006308864116668702, 0.006343679904937744, 0.006334464073181153, 0.006347775936126709, 0.006309887886047363, 0.006394879817962646, 0.006355967998504639, 0.006307839870452881, 0.006305791854858398, 0.006330368041992188, 0.01386291217803955, 0.006079487800598145, 0.006160384178161621, 0.006077439785003662, 0.006139904022216797, 0.006077439785003662, 0.0060999679565429685, 0.006090752124786377, 0.006131711959838867, 0.00612556791305542, 0.006136832237243653, 0.006083680152893066, 0.006180768013000488, 0.006146048069000244, 0.0070256638526916505, 0.006468607902526856, 0.007489535808563232, 0.007550975799560547, 0.006666240215301514, 0.0064102401733398436, 0.006331391811370849, 0.006417503833770752, 0.006405024051666259, 0.006338560104370118, 0.006398079872131347, 0.006327167987823486, 0.006369279861450195, 0.006356031894683838, 0.006265791893005371, 0.006545407772064209, 0.006386688232421875, 0.006304768085479737, 0.006341631889343262, 0.006294528007507324, 0.006302720069885254, 0.006339583873748779, 0.006849535942077637, 0.006423552036285401, 0.006347775936126709, 0.006335584163665772, 0.00669379186630249, 0.0071526398658752445, 0.0065177597999572755, 0.006359072208404541, 0.006400991916656494, 0.0064471039772033695, 0.0063498239517211915, 0.006440959930419922, 0.006511616230010986, 0.006398975849151611, 0.006379519939422608, 0.006331391811370849, 0.006423552036285401, 0.006333439826965332, 0.006335487842559814, 0.006400000095367431, 0.006343679904937744, 0.006346752166748047, 0.006345727920532227, 0.006340608119964599, 0.006427680015563965, 0.0063497920036315915, 0.006466559886932373, 0.014119935989379882, 0.006392831802368164, 0.006354944229125976, 0.006333439826965332, 0.006343679904937744, 0.006301695823669433, 0.006364160060882569, 0.006313983917236328, 0.0063283519744873045, 0.006335455894470215, 0.006295551776885986, 0.006432767868041992, 0.0063508481979370115, 0.006718463897705078, 0.006359039783477783, 0.006350912094116211, 0.0063415679931640625, 0.006341631889343262, 0.006339583873748779, 0.006322175979614258, 0.0063272957801818845, 0.006369279861450195, 0.006298624038696289, 0.006370304107666015, 0.006294528007507324, 0.006358016014099121, 0.006326272010803223, 0.006306816101074219, 0.00636630392074585, 0.006334368228912353, 0.006345727920532227, 0.006364192008972168, 0.006336480140686035, 0.006372352123260498, 0.006550528049468994, 0.006393856048583985, 0.006618144035339356, 0.00637334394454956, 0.006377471923828125, 0.006316031932830811, 0.006319104194641113, 0.0063805441856384275, 0.006323200225830078, 0.006384640216827392, 0.006315008163452148, 0.006328320026397705, 0.006333439826965332, 0.006376448154449463, 0.006362112045288086, 0.0062904319763183595, 0.006362112045288086, 0.006306816101074219, 0.006302720069885254, 0.006346752166748047, 0.006302720069885254, 0.006337535858154297, 0.006362112045288086, 0.006328320026397705, 0.006325247764587402, 0.0063569917678833006, 0.006354944229125976, 0.006309887886047363, 0.006328320026397705, 0.014107647895812989, 0.006323200225830078, 0.0063498239517211915, 0.00675328016281128, 0.006523903846740723, 0.006344704151153564, 0.006330368041992188, 0.006370304107666015, 0.006299647808074951, 0.006342656135559082, 0.006323200225830078, 0.006330368041992188, 0.006305791854858398, 0.0063539199829101565, 0.006342688083648682, 0.006325215816497803, 0.006328320026397705, 0.0062873601913452145, 0.006328320026397705, 0.006323200225830078, 0.0063805441856384275, 0.006367263793945313, 0.006295519828796387, 0.006311935901641846, 0.006299647808074951, 0.006115327835083008, 0.006139904022216797, 0.006103040218353272, 0.006088704109191895, 0.0060661759376525876, 0.006115327835083008, 0.006090752124786377, 0.00612556791305542, 0.006288383960723877, 0.006171648025512695, 0.006089727878570556, 0.006127615928649902, 0.0061265921592712404, 0.006203392028808594, 0.006114304065704346, 0.006170783996582031, 0.006146912097930908, 0.00609494400024414, 0.006137760162353516, 0.0061521921157836916, 0.006170688152313232, 0.006285247802734375, 0.006127615928649902, 0.006108160018920898, 0.006119423866271972, 0.006264832019805908, 0.006139904022216797, 0.006095871925354004, 0.006096896171569824, 0.00612556791305542, 0.006073344230651856, 0.006123519897460937, 0.006069248199462891, 0.006113279819488526, 0.006153215885162353, 0.006199295997619629, 0.006097919940948486, 0.006316031932830811, 0.014223360061645507, 0.0063211522102355954, 0.006501376152038574, 0.006308864116668702, 0.006329472064971924, 0.006300543785095215, 0.006308864116668702, 0.006336512088775635, 0.006338592052459717, 0.006377439975738525, 0.006291456222534179, 0.006339583873748779, 0.006321248054504395, 0.006334368228912353, 0.00635097599029541, 0.00631385612487793, 0.0063272957801818845, 0.006304768085479737, 0.006351967811584472, 0.006440864086151123, 0.006337535858154297, 0.006364160060882569, 0.0063201279640197755, 0.006400000095367431, 0.006329343795776367, 0.006348800182342529, 0.006351871967315674, 0.006338560104370118, 0.006337535858154297, 0.0062975997924804685, 0.0063805441856384275, 0.006376448154449463, 0.006343679904937744, 0.0063539199829101565, 0.006343679904937744, 0.006348800182342529, 0.006331391811370849, 0.006334464073181153, 0.00632422399520874, 0.006270976066589356, 0.0063170561790466305, 0.0063805441856384275, 0.006333439826965332, 0.006360064029693604, 0.007351295948028564, 0.0070830078125, 0.006966271877288818, 0.007106560230255127, 0.006434815883636475, 0.006427648067474365, 0.006407167911529541, 0.006352928161621094, 0.006345695972442627, 0.006310912132263183, 0.006391808032989502, 0.006418432235717773, 0.006384640216827392, 0.006913023948669434, 0.007067647933959961, 0.006223872184753418, 0.006182911872863769, 0.006173759937286377, 0.006243264198303223, 0.013827072143554688, 0.006154240131378174, 0.0061337599754333495, 0.006166528224945069, 0.00616755199432373, 0.006109183788299561, 0.006077439785003662, 0.0061859841346740725, 0.006093823909759521, 0.006152256011962891, 0.0061848959922790525, 0.006139904022216797, 0.006162432193756104, 0.006113279819488526, 0.006181888103485108, 0.006318079948425293, 0.006364160060882569, 0.006419456005096436, 0.00638259220123291, 0.006416416168212891, 0.006309855937957764, 0.00636518383026123, 0.006316031932830811, 0.006322175979614258, 0.006393856048583985, 0.006284287929534912, 0.006312960147857666, 0.006304768085479737, 0.0062791681289672855, 0.006309887886047363, 0.006293504238128662, 0.006401023864746094, 0.0062904319763183595, 0.006358016014099121, 0.0062791681289672855, 0.006348800182342529, 0.00658841609954834, 0.00638976001739502, 0.00638156795501709, 0.006323200225830078, 0.006330368041992188, 0.006343679904937744, 0.006337535858154297, 0.006398975849151611, 0.006366208076477051, 0.006346752166748047, 0.00638976001739502, 0.0062863359451293946, 0.006369279861450195, 0.006310912132263183, 0.0065372161865234375, 0.006305823802947998, 0.006299615859985352, 0.0063211522102355954, 0.006299647808074951, 0.006326272010803223, 0.006337535858154297, 0.006326272010803223, 0.006303743839263916, 0.0063201279640197755, 0.006366208076477051, 0.00634278392791748, 0.006356863975524902, 0.014119935989379882, 0.006462463855743408, 0.0063272957801818845, 0.0063211522102355954, 0.006334464073181153, 0.006313983917236328, 0.006336512088775635, 0.006469632148742676, 0.006356031894683838, 0.006343616008758545, 0.00632422399520874, 0.006336703777313233, 0.006304575920104981, 0.00636518383026123, 0.006362112045288086, 0.006308864116668702, 0.006289408206939697, 0.00612556791305542, 0.0060999679565429685, 0.006106112003326416, 0.0061296639442443845, 0.006190080165863037, 0.006167679786682129, 0.006168511867523194, 0.0063129281997680664, 0.0063426241874694825, 0.006743040084838867, 0.006146048069000244, 0.006088704109191895, 0.006104063987731933, 0.006388735771179199, 0.0063201279640197755, 0.006309887886047363, 0.0064778242111206055, 0.006313983917236328, 0.006128640174865723, 0.006212607860565185, 0.006402048110961914, 0.006282368183135986, 0.006367167949676514, 0.006296512126922608, 0.006305791854858398, 0.006323200225830078, 0.006312960147857666, 0.006325247764587402, 0.006232063770294189, 0.006285312175750732, 0.006204480171203613, 0.006052800178527832, 0.005981184005737304, 0.0059996161460876465, 0.0059996161460876465, 0.005992608070373535, 0.006159200191497803, 0.006232063770294189, 0.0062341117858886715, 0.00620959997177124, 0.006194111824035645, 0.006237184047698975, 0.006203392028808594, 0.0062269759178161625, 0.00624128007888794, 0.00623305606842041]",tokens/s,154.8595930446798,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Salesforce/codegen-6B-nl,Salesforce/codegen-6B-nl,cuda,0,42,,,True,,,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",codegen,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-410m,EleutherAI/pythia-410m,cuda,0,42,,,True,True,True,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.42.1,,0.31.0,,,,1.20.0,,,,0.11.1,,,,MB,1127.718912,2563.244032,0.0,1923.088384,1891.20256,s,11,0.18864259338378905,0.01714932667125355,0.00017200955493621296,0.01710428810119629,0.017189599990844726,0.017428239822387694,0.01761915168762207,"[0.017666879653930663, 0.017092159271240234, 0.017189599990844726, 0.01710428810119629, 0.016975744247436524, 0.017066688537597657, 0.017127424240112304, 0.017155231475830077, 0.017123552322387697, 0.017078208923339843, 0.017062816619873047]",tokens/s,14927.699781305024,kWh,2.0373009602000431e-07,1.1162646050524405e-07,1.3940729543379313e-06,1.7094295108631795e-06,tokens/kWh,149757564.364694,MB,1127.718912,2563.244032,0.0,1923.088384,1895.80288,s,11,9.91951788330078,0.9017743530273438,0.014426219656692317,0.8999153442382812,0.9109261474609375,0.9240941162109375,0.9346284912109375,"[0.9372620849609375, 0.9109261474609375, 0.8886005859375, 0.8956931762695313, 0.88610693359375, 0.8879814453125, 0.8894957885742187, 0.903033935546875, 0.91068603515625, 0.90981640625, 0.8999153442382812]",tokens/s,69.86226630697902,kWh,1.0631748260972354e-05,5.824595549962069e-06,2.588354987784384e-05,4.233989368877824e-05,tokens/kWh,1487958.3889153106,,s,692,10.075111425399784,0.01455940957427714,0.0020905949694838347,0.014394368171691894,0.01461637134552002,0.015043430423736567,0.03154311212539673,"[0.015462400436401368, 0.015341567993164062, 0.015394816398620606, 0.015101951599121094, 0.015127552032470704, 0.01528934383392334, 0.01517568016052246, 0.01524227237701416, 0.015224800109863282, 0.015193087577819824, 0.0156692476272583, 0.015429632186889648, 0.015118335723876953, 0.014877696037292481, 0.014721023559570312, 0.014592000007629394, 0.014565376281738282, 0.014499839782714843, 0.014818304061889649, 0.014446592330932618, 0.014541824340820313, 0.014625791549682618, 0.014954496383666992, 0.017670143127441407, 0.015337471961975097, 0.015158271789550782, 0.015095808029174805, 0.01531289577484131, 0.014941184043884278, 0.01489510440826416, 0.015244288444519043, 0.015176704406738281, 0.014929920196533204, 0.01457151985168457, 0.0160819206237793, 0.015273983955383302, 0.014771200180053711, 0.014451711654663087, 0.01465446376800537, 0.01451417636871338, 0.014486528396606446, 0.01445683193206787, 0.014512127876281738, 0.0144650239944458, 0.014556159973144531, 0.014552063941955566, 0.01451417636871338, 0.014485504150390625, 0.014501888275146485, 0.014474240303039551, 0.014510080337524415, 0.014402560234069824, 0.01428889560699463, 0.014323712348937988, 0.01454694366455078, 0.014468095779418945, 0.014502911567687989, 0.014376959800720214, 0.014523391723632812, 0.014467071533203125, 0.014403583526611329, 0.014230527877807618, 0.03168460845947266, 0.014433279991149902, 0.014376992225646973, 0.014491616249084472, 0.014156800270080566, 0.014551072120666505, 0.014512096405029297, 0.014746623992919922, 0.014657535552978516, 0.015766528129577635, 0.014889984130859376, 0.01453878402709961, 0.014492639541625977, 0.014500864028930664, 0.01446399974822998, 0.014535679817199706, 0.014450688362121582, 0.01448038387298584, 0.014458911895751953, 0.014467040061950683, 0.014433279991149902, 0.014330880165100097, 0.014178303718566895, 0.01425100803375244, 0.014332927703857423, 0.014432255744934081, 0.014466048240661621, 0.014306367874145508, 0.014176192283630372, 0.014341119766235352, 0.014358528137207031, 0.014442496299743653, 0.01447935962677002, 0.014397439956665039, 0.014438400268554688, 0.014491647720336913, 0.01448857593536377, 0.014560256004333496, 0.0144650239944458, 0.014454784393310547, 0.01436569595336914, 0.01447321605682373, 0.014491647720336913, 0.014425087928771972, 0.014358528137207031, 0.014223360061645507, 0.014240768432617188, 0.014461952209472656, 0.014476287841796874, 0.014234623908996581, 0.0142673921585083, 0.014431232452392579, 0.014401535987854003, 0.014557184219360352, 0.01444863986968994, 0.014529536247253418, 0.014441472053527832, 0.014468095779418945, 0.014451711654663087, 0.014433279991149902, 0.01449062442779541, 0.01448038387298584, 0.014416895866394042, 0.03176345634460449, 0.014429183959960937, 0.01435750389099121, 0.014426143646240234, 0.014475232124328613, 0.014327808380126953, 0.014345215797424317, 0.014352383613586426, 0.01429196834564209, 0.013853695869445801, 0.013875200271606445, 0.013864959716796875, 0.01386393642425537, 0.013839360237121581, 0.01387929630279541, 0.01386393642425537, 0.013835264205932616, 0.013840383529663085, 0.013868063926696777, 0.013826016426086427, 0.013840383529663085, 0.01386291217803955, 0.013821951866149903, 0.013851648330688476, 0.013806591987609864, 0.013829119682312012, 0.013796352386474609, 0.013899776458740234, 0.013865983963012696, 0.013820927619934082, 0.013780991554260253, 0.013853695869445801, 0.013818880081176758, 0.013816831588745117, 0.01386393642425537, 0.01386393642425537, 0.013797375679016113, 0.013828096389770507, 0.013881343841552735, 0.013858816146850587, 0.013839360237121581, 0.013850624084472657, 0.013841407775878906, 0.014162943840026856, 0.014418944358825684, 0.014440447807312011, 0.014544896125793457, 0.014397472381591796, 0.01442403221130371, 0.01454694366455078, 0.014616576194763184, 0.014366720199584961, 0.014439423561096192, 0.01447219181060791, 0.01429196834564209, 0.014232576370239258, 0.014445568084716797, 0.014425087928771972, 0.01437388801574707, 0.014430208206176758, 0.01444863986968994, 0.014460927963256836, 0.014418944358825684, 0.031718399047851564, 0.014278656005859374, 0.014171135902404786, 0.014431232452392579, 0.014378016471862792, 0.01445680046081543, 0.014715935707092285, 0.0145448637008667, 0.014417920112609863, 0.014468095779418945, 0.01447116756439209, 0.014458880424499512, 0.014515199661254884, 0.014444543838500976, 0.014223360061645507, 0.014621696472167968, 0.014617600440979005, 0.014527487754821777, 0.01439641571044922, 0.014416928291320802, 0.014376928329467773, 0.01437491226196289, 0.014051327705383301, 0.013880319595336914, 0.01389363193511963, 0.013892607688903809, 0.013855744361877441, 0.013911040306091308, 0.014095359802246094, 0.014223360061645507, 0.014278656005859374, 0.014305279731750489, 0.013789183616638183, 0.01384447956085205, 0.013829119682312012, 0.013800448417663574, 0.013791232109069825, 0.013858816146850587, 0.013855744361877441, 0.013891584396362304, 0.0138854398727417, 0.014626815795898437, 0.014376959800720214, 0.014402560234069824, 0.01446399974822998, 0.014368767738342286, 0.01439027214050293, 0.014476287841796874, 0.014403583526611329, 0.014375935554504395, 0.014415871620178223, 0.014394368171691894, 0.01439948844909668, 0.014443519592285157, 0.014425087928771972, 0.013849599838256836, 0.013821951866149903, 0.013845503807067871, 0.013822976112365723, 0.01388748836517334, 0.014053376197814941, 0.013826047897338867, 0.013825023651123047, 0.030918655395507814, 0.013806591987609864, 0.013807616233825683, 0.01466982364654541, 0.014433279991149902, 0.014407679557800293, 0.014355456352233887, 0.014423040390014649, 0.014354432106018066, 0.014187520027160644, 0.013974528312683105, 0.01386905574798584, 0.014221311569213867, 0.014565376281738282, 0.014050304412841797, 0.013966336250305175, 0.013908991813659668, 0.013846528053283692, 0.013809663772583008, 0.013897727966308594, 0.013865983963012696, 0.013941760063171387, 0.013786111831665039, 0.013730815887451172, 0.01375436782836914, 0.013845503807067871, 0.013837311744689941, 0.013874176025390626, 0.01384447956085205, 0.013900799751281738, 0.013849599838256836, 0.013833215713500976, 0.01384447956085205, 0.013873151779174805, 0.014631936073303223, 0.014494720458984376, 0.014382080078125, 0.014453760147094727, 0.01439846420288086, 0.01440665626525879, 0.01438310432434082, 0.014402560234069824, 0.014414848327636719, 0.014253055572509766, 0.013902848243713378, 0.01395206356048584, 0.013827008247375488, 0.013868032455444336, 0.013868032455444336, 0.013868032455444336, 0.01389363193511963, 0.013832192420959472, 0.014134271621704102, 0.014328831672668458, 0.01416710376739502, 0.013690815925598145, 0.013716480255126954, 0.013822976112365723, 0.013801471710205078, 0.013888511657714844, 0.014272512435913086, 0.01448857593536377, 0.014298111915588378, 0.03153408050537109, 0.01376153564453125, 0.013732864379882812, 0.01374617576599121, 0.013821951866149903, 0.01385267162322998, 0.013937664031982423, 0.01451417636871338, 0.014143487930297852, 0.013949952125549316, 0.014017536163330077, 0.014255104064941406, 0.014167039871215821, 0.013645824432373046, 0.013806591987609864, 0.01399295997619629, 0.013820927619934082, 0.014095359802246094, 0.014462976455688477, 0.014565376281738282, 0.014573568344116212, 0.01407692813873291, 0.0139683837890625, 0.013924351692199707, 0.013832192420959472, 0.013923328399658203, 0.013820927619934082, 0.013867008209228515, 0.013692928314208984, 0.013813759803771973, 0.014642175674438476, 0.014358528137207031, 0.014354432106018066, 0.014405632019042968, 0.014366720199584961, 0.014394368171691894, 0.014347264289855957, 0.014348287582397461, 0.01439846420288086, 0.014043135643005371, 0.013828096389770507, 0.0138854398727417, 0.013829119682312012, 0.01387110424041748, 0.014414848327636719, 0.014552063941955566, 0.014405632019042968, 0.014178303718566895, 0.014354432106018066, 0.01439948844909668, 0.014228480339050293, 0.01387007999420166, 0.013854720115661622, 0.01384447956085205, 0.01384447956085205, 0.013913087844848633, 0.01383020782470703, 0.013903807640075684, 0.013857791900634766, 0.01377280044555664, 0.014414848327636719, 0.014529536247253418, 0.014503935813903808, 0.031153152465820313, 0.01376153564453125, 0.013775872230529786, 0.013831168174743653, 0.014312447547912598, 0.014362624168395996, 0.013944831848144532, 0.013853695869445801, 0.013825023651123047, 0.013798399925231934, 0.013851648330688476, 0.013828096389770507, 0.013804544448852539, 0.013824000358581542, 0.01377894401550293, 0.013847552299499511, 0.013797375679016113, 0.013964287757873535, 0.014458880424499512, 0.014413824081420898, 0.014600192070007324, 0.014376959800720214, 0.01397760009765625, 0.01376972770690918, 0.013800448417663574, 0.01366220760345459, 0.013817855834960938, 0.01401961612701416, 0.013734880447387696, 0.014232576370239258, 0.014817279815673828, 0.014410752296447754, 0.014441472053527832, 0.014498815536499024, 0.014404607772827148, 0.014310400009155273, 0.014358528137207031, 0.014415871620178223, 0.014478336334228516, 0.014364671707153321, 0.014328831672668458, 0.014310400009155273, 0.014321663856506347, 0.01439033603668213, 0.014382016181945801, 0.014396448135375977, 0.014342144012451171, 0.014423007965087891, 0.014379008293151856, 0.014261247634887696, 0.013841407775878906, 0.013884415626525879, 0.013824000358581542, 0.013850624084472657, 0.013830143928527832, 0.013946880340576171, 0.013805567741394043, 0.013851648330688476, 0.013858816146850587, 0.0142673921585083, 0.014475263595581055, 0.01456332778930664, 0.014455807685852052, 0.03163443183898926, 0.014171135902404786, 0.014007295608520508, 0.01435750389099121, 0.01437286376953125, 0.014336000442504883, 0.01437286376953125, 0.014384127616882325, 0.014321663856506347, 0.014386176109313965, 0.014359552383422852, 0.014452735900878906, 0.014007295608520508, 0.01430016040802002, 0.014466048240661621, 0.014362624168395996, 0.013825023651123047, 0.01384447956085205, 0.013813759803771973, 0.013856767654418945, 0.013820927619934082, 0.013858816146850587, 0.013834239959716797, 0.013822976112365723, 0.013828096389770507, 0.014356479644775391, 0.014405632019042968, 0.014381119728088379, 0.014400447845458984, 0.014394368171691894, 0.014449664115905762, 0.014496767997741699, 0.014450688362121582, 0.014682111740112304, 0.014422016143798828, 0.014279680252075195, 0.014244864463806153, 0.01420083236694336, 0.014423040390014649, 0.014418944358825684, 0.014244864463806153, 0.01420083236694336, 0.014765055656433105, 0.014847999572753906, 0.014449664115905762, 0.014482432365417481, 0.01450598430633545, 0.014443519592285157, 0.01528217601776123, 0.014567423820495605, 0.014495743751525878, 0.014493696212768555, 0.014455807685852052, 0.014451711654663087, 0.01447935962677002, 0.014445568084716797, 0.01448960018157959, 0.01448038387298584, 0.014419967651367188, 0.014494720458984376, 0.014328864097595215, 0.014204895973205567, 0.014379008293151856, 0.03175833511352539, 0.014338047981262207, 0.014314496040344238, 0.014389280319213868, 0.014451680183410644, 0.01452239990234375, 0.014338015556335449, 0.014297087669372559, 0.014343168258666992, 0.014311424255371094, 0.014429183959960937, 0.014467071533203125, 0.014508031845092773, 0.014414848327636719, 0.014444543838500976, 0.014873600006103516, 0.015532032012939453, 0.014921728134155274, 0.014475263595581055, 0.014466048240661621, 0.014478336334228516, 0.01444863986968994, 0.0139683837890625, 0.014498815536499024, 0.014220288276672363, 0.014324735641479493, 0.014476287841796874, 0.014537759780883789, 0.014244832038879395, 0.014223360061645507, 0.014180352210998535, 0.014256128311157227, 0.014469120025634765, 0.014537728309631348, 0.014642175674438476, 0.01449779224395752, 0.01447116756439209, 0.014400511741638184, 0.014424063682556153, 0.014539775848388671, 0.01438310432434082, 0.014408703804016113, 0.014441472053527832, 0.014500864028930664, 0.014609408378601075, 0.01439641571044922, 0.014510080337524415, 0.014402560234069824, 0.014255135536193848, 0.014333919525146484, 0.014558208465576173, 0.014621696472167968, 0.014438400268554688, 0.014454784393310547, 0.01437183952331543, 0.014440447807312011, 0.014409728050231933, 0.014439423561096192, 0.014362624168395996, 0.014440447807312011, 0.01448755168914795, 0.014624768257141114, 0.014520319938659668, 0.03166720008850098, 0.014429183959960937, 0.014387200355529785, 0.014322688102722168, 0.014350336074829101, 0.014332927703857423, 0.014256128311157227, 0.014343168258666992, 0.014448703765869141, 0.014441408157348633, 0.014440447807312011, 0.014458880424499512, 0.014478336334228516, 0.014444543838500976, 0.014394368171691894, 0.01438822364807129, 0.014385184288024903, 0.01452336025238037, 0.014492671966552734, 0.014447615623474122, 0.014614527702331542, 0.014412799835205077, 0.01445683193206787, 0.01447116756439209, 0.014420991897583007, 0.014498815536499024, 0.014439423561096192, 0.014459967613220215, 0.014474176406860351, 0.014433279991149902, 0.01448857593536377, 0.014485504150390625, 0.01448038387298584, 0.014419967651367188, 0.014377984046936035, 0.014449664115905762, 0.014431232452392579, 0.014462976455688477, 0.014412799835205077, 0.014486528396606446, 0.014408703804016113, 0.014431232452392579, 0.014405632019042968, 0.014495743751525878, 0.014424063682556153, 0.014478336334228516, 0.014433279991149902, 0.0142673921585083, 0.014307328224182129, 0.014431232452392579, 0.014457856178283691, 0.014499839782714843, 0.014512127876281738, 0.0145797119140625, 0.014460927963256836, 0.014438400268554688, 0.014468095779418945, 0.014585856437683106, 0.014424063682556153, 0.01437183952331543, 0.014418944358825684, 0.01450598430633545, 0.014470144271850586, 0.031678464889526366, 0.01447219181060791, 0.01500057601928711, 0.01477734375, 0.01461353588104248, 0.01440764808654785, 0.014780415534973144, 0.014534655570983887, 0.014525440216064453, 0.014603263854980468, 0.014508031845092773, 0.014407679557800293, 0.014339072227478027, 0.014451711654663087, 0.014377984046936035, 0.014780415534973144, 0.014504960060119629, 0.014447615623474122, 0.014511103630065919, 0.014577664375305176, 0.014368767738342286, 0.01438310432434082, 0.014311424255371094, 0.013890560150146485, 0.013831168174743653, 0.01386393642425537, 0.013812735557556152, 0.014443519592285157, 0.014343168258666992, 0.014011391639709473, 0.013880319595336914, 0.013898752212524413, 0.013848575592041015, 0.013898752212524413, 0.013900799751281738, 0.013880319595336914, 0.013739007949829102, 0.013768704414367675, 0.013811712265014648, 0.013855744361877441, 0.013858816146850587, 0.013830143928527832, 0.013888511657714844, 0.014297087669372559, 0.014194687843322755, 0.01435750389099121, 0.014460927963256836, 0.014029824256896972, 0.014056447982788087, 0.014891008377075195, 0.01458073616027832, 0.014408703804016113, 0.014436351776123046, 0.014377984046936035, 0.014368767738342286, 0.014411808013916016, 0.014487520217895507, 0.014362655639648437, 0.014323679924011231, 0.014304256439208985, 0.014364671707153321, 0.014354432106018066, 0.014363648414611817]",tokens/s,68.68410390533636,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,.,.,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 373, in cached_file raise EnvironmentError( OSError: . does not appear to have a file named config.json. Checkout 'https://huggingface.co/./tree/None' for available files. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,a,a,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/a/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6694749f-490bdf8678dde2fd46980025;1caa4945-183e-4ba8-ab68-cbe5091d9404) Repository Not Found for url: https://huggingface.co/a/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: a is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-7B,Qwen/Qwen-7B,cuda,0,42,,,True,,,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",qwen,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-66b,facebook/opt-66b,cuda,0,42,,,True,,,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",opt,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,stabilityai/stablelm-base-alpha-7b,stabilityai/stablelm-base-alpha-7b,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1283.559424,17197.170688,0.0,16550.723584,16322.776064,s,10,0.6767843856811524,0.06767843856811524,0.0002437592060178596,0.06760496139526367,0.06790468215942383,0.06812058067321777,0.06829329948425293,"[0.06833647918701172, 0.06748908996582031, 0.06751113891601562, 0.06763350677490235, 0.067576416015625, 0.06785670471191406, 0.06768377685546875, 0.0676588134765625, 0.06748751831054688, 0.06755094146728516]",tokens/s,3782.5931776240336,kWh,7.997584716604279e-07,4.381849315315186e-07,4.951503585824165e-06,6.189446989016111e-06,tokens/kWh,41360722.606446356,MB,1283.559424,17197.170688,0.0,16550.723584,16467.883008,s,10,22.375432373046873,2.2375432373046875,0.0034930341260981986,2.2367978515624998,2.23843525390625,2.2430009521484373,2.2466535107421874,"[2.247566650390625, 2.236116455078125, 2.23733154296875, 2.237420654296875, 2.237342041015625, 2.236286376953125, 2.23705517578125, 2.235997314453125, 2.233775634765625, 2.23654052734375]",tokens/s,28.15588049859939,kWh,2.6347697339840122e-05,1.443942008482269e-05,0.00012762788025537844,0.00016841499768004123,tokens/kWh,374075.94850720407,,s,629,22.94772020721437,0.03648286201464922,0.00807406808590342,0.03549491119384766,0.03560673370361328,0.03568964996337891,0.10337087768554687,"[0.03605606460571289, 0.03609088134765625, 0.03603353500366211, 0.03602227020263672, 0.03604582214355469, 0.03605299377441406, 0.03602841567993164, 0.03605196762084961, 0.03604684829711914, 0.03602534484863281, 0.03601203155517578, 0.03605401611328125, 0.03599564743041992, 0.03587788772583008, 0.03570073699951172, 0.03551846313476562, 0.035454975128173825, 0.035460094451904296, 0.03543654251098633, 0.0355676155090332, 0.03556454467773437, 0.035460094451904296, 0.035501056671142575, 0.035620864868164064, 0.03560140609741211, 0.03551641464233399, 0.03550310516357422, 0.03551846313476562, 0.03549190521240234, 0.03553273773193359, 0.03557580947875977, 0.03558399963378906, 0.03549497604370117, 0.03553273773193359, 0.03553484725952148, 0.03547750473022461, 0.03547340774536133, 0.03549798583984375, 0.03554611206054688, 0.035707904815673826, 0.03577753448486328, 0.0355491828918457, 0.03552460861206055, 0.0355860481262207, 0.035532958984375, 0.03554083251953125, 0.03552358245849609, 0.035525630950927735, 0.03550217437744141, 0.03552963256835937, 0.035560447692871096, 0.03558092880249023, 0.035585025787353515, 0.03558911895751953, 0.0356126708984375, 0.035576831817626955, 0.03567103958129883, 0.03565055847167969, 0.035597343444824216, 0.03558601760864258, 0.03561983871459961, 0.03559219360351563, 0.10344652557373046, 0.03541507339477539, 0.03545289611816406, 0.03541196823120117, 0.035388416290283206, 0.03539763259887695, 0.03538431930541992, 0.03540582275390625, 0.03541299057006836, 0.03542015838623047, 0.035451904296875, 0.03545600128173828, 0.03546316909790039, 0.035422206878662106, 0.03546316909790039, 0.035432449340820314, 0.03543756866455078, 0.035422206878662106, 0.03543756866455078, 0.03549388885498047, 0.03548057556152344, 0.03544268798828125, 0.03546931076049804, 0.03543654251098633, 0.03541708755493164, 0.03544268798828125, 0.03545292663574219, 0.03544575881958008, 0.03546828842163086, 0.0354785270690918, 0.03548262405395508, 0.03546828842163086, 0.03544780731201172, 0.035434497833251956, 0.03550207901000976, 0.03547955322265625, 0.035492862701416016, 0.03547443389892578, 0.035509246826171875, 0.03548262405395508, 0.03553792190551758, 0.03548672103881836, 0.03551340866088867, 0.0355214729309082, 0.035504127502441404, 0.03551641464233399, 0.035576831817626955, 0.0354856948852539, 0.03551027297973633, 0.035550209045410154, 0.03556147384643555, 0.0355491828918457, 0.03556966400146484, 0.03555942535400391, 0.035535873413085936, 0.03553996658325195, 0.03553494262695313, 0.03560563278198242, 0.03565852737426758, 0.03556876754760742, 0.03561663818359375, 0.035585025787353515, 0.03564748764038086, 0.10337689971923827, 0.03535974502563476, 0.035372032165527346, 0.03537408065795898, 0.035432449340820314, 0.03542732620239258, 0.03549593734741211, 0.035434497833251956, 0.035399681091308595, 0.035388416290283206, 0.03544473648071289, 0.03549593734741211, 0.03545804977416992, 0.0354856948852539, 0.03546112060546875, 0.035408927917480466, 0.03543856048583984, 0.035451904296875, 0.03541196823120117, 0.03546726226806641, 0.03544473648071289, 0.03545600128173828, 0.03561779022216797, 0.035732479095458985, 0.03550822448730469, 0.03545292663574219, 0.03544166564941406, 0.03546931076049804, 0.03549593734741211, 0.03548672103881836, 0.035522560119628906, 0.03549184036254883, 0.035504161834716795, 0.03547030258178711, 0.035484672546386715, 0.03552665710449219, 0.03549388885498047, 0.03551129531860352, 0.03550207901000976, 0.03551334381103516, 0.03551129531860352, 0.03568025588989258, 0.03555737686157227, 0.03554099273681641, 0.035527679443359376, 0.035542015075683595, 0.03553177642822265, 0.035550209045410154, 0.035550209045410154, 0.035538944244384765, 0.035552257537841796, 0.035536895751953124, 0.035575870513916016, 0.0355860481262207, 0.03568531036376953, 0.03558607864379883, 0.0355972785949707, 0.03557891082763672, 0.03559625625610351, 0.03557785415649414, 0.03559731292724609, 0.03557580947875977, 0.035593215942382815, 0.1035182113647461, 0.03547545623779297, 0.03552870559692383, 0.035643390655517575, 0.0355788803100586, 0.03545702362060547, 0.03541708755493164, 0.03540172958374024, 0.03540172958374024, 0.03543552017211914, 0.035422206878662106, 0.03540070343017578, 0.03544166564941406, 0.03547545623779297, 0.03543247985839844, 0.035445728302001954, 0.03545395278930664, 0.035440673828125, 0.035432415008544924, 0.0354334716796875, 0.03544063949584961, 0.035434497833251956, 0.03544268798828125, 0.0355082893371582, 0.03550508880615234, 0.03549388885498047, 0.03548160171508789, 0.03549900817871094, 0.03546931076049804, 0.03547750473022461, 0.03546828842163086, 0.03549184036254883, 0.03549491119384766, 0.03551027297973633, 0.035471359252929685, 0.03547955322265625, 0.0354969596862793, 0.035576831817626955, 0.035527679443359376, 0.035558399200439454, 0.03552262496948242, 0.03553683090209961, 0.03551232147216797, 0.03552358245849609, 0.035530750274658206, 0.03553484725952148, 0.0355676155090332, 0.035542015075683595, 0.035588096618652344, 0.035550209045410154, 0.03557273483276367, 0.035542015075683595, 0.03556147384643555, 0.035566593170166014, 0.03556966400146484, 0.035591167449951173, 0.035574783325195314, 0.03559731292724609, 0.0356280632019043, 0.03562697601318359, 0.035588096618652344, 0.03560755157470703, 0.035588096618652344, 0.10342400360107422, 0.03538431930541992, 0.0353966064453125, 0.03540889739990234, 0.035410945892333984, 0.03540377426147461, 0.035416065216064455, 0.03544985580444336, 0.03548057556152344, 0.03544678497314453, 0.035460094451904296, 0.03542425537109375, 0.035434497833251956, 0.035444862365722654, 0.0354384651184082, 0.035432449340820314, 0.035416065216064455, 0.03544166564941406, 0.035438591003417966, 0.03544678497314453, 0.03544678497314453, 0.03542732620239258, 0.03546316909790039, 0.03542937469482422, 0.03548262405395508, 0.03547955322265625, 0.0354703369140625, 0.035460094451904296, 0.03545600128173828, 0.03544473648071289, 0.035476478576660156, 0.035484672546386715, 0.035550209045410154, 0.03551641464233399, 0.035522560119628906, 0.0354785270690918, 0.03549491119384766, 0.035501056671142575, 0.035520511627197264, 0.035490848541259765, 0.03552560043334961, 0.035517440795898435, 0.03551129531860352, 0.03552153778076172, 0.035527679443359376, 0.03563008117675781, 0.03583180618286133, 0.03573350524902344, 0.03559936141967773, 0.03563008117675781, 0.03554816055297851, 0.03554816055297851, 0.03556454467773437, 0.03554508972167969, 0.03556556701660156, 0.03557580947875977, 0.035563518524169925, 0.035643390655517575, 0.03557785415649414, 0.03567308807373047, 0.03562905502319336, 0.03564031982421875, 0.03562393569946289, 0.10381926727294923, 0.035381248474121094, 0.03538534545898438, 0.03541708755493164, 0.035422206878662106, 0.03545292663574219, 0.03540787124633789, 0.03539763259887695, 0.035451904296875, 0.03540377426147461, 0.03541708755493164, 0.035404800415039066, 0.03544784164428711, 0.03544060897827148, 0.03542015838623047, 0.03542323303222656, 0.03542015838623047, 0.03545907211303711, 0.035432449340820314, 0.03544985580444336, 0.03545907211303711, 0.035419136047363284, 0.03544063949584961, 0.03543142318725586, 0.03540582275390625, 0.03546419143676758, 0.035484672546386715, 0.03552460861206055, 0.03548876953125, 0.03547750473022461, 0.03544985580444336, 0.03546015930175781, 0.035496929168701175, 0.0354662094116211, 0.03550310516357422, 0.03549388885498047, 0.03547443389892578, 0.035492862701416016, 0.0355491828918457, 0.03550003051757813, 0.03556249618530274, 0.03552358245849609, 0.03554611206054688, 0.03552153778076172, 0.03552870559692383, 0.035517440795898435, 0.035506175994873046, 0.03552358245849609, 0.03551232147216797, 0.035527679443359376, 0.03556147384643555, 0.035585025787353515, 0.03558195114135742, 0.03555430221557617, 0.03556966400146484, 0.03560345458984375, 0.035563518524169925, 0.03558911895751953, 0.03560345458984375, 0.035574817657470705, 0.03557577514648438, 0.03560140609741211, 0.03563520050048828, 0.10356428527832032, 0.03548160171508789, 0.03555123138427734, 0.035399681091308595, 0.03537715148925781, 0.035425312042236326, 0.03540886306762695, 0.03540377426147461, 0.03537919998168945, 0.035399681091308595, 0.03545910263061523, 0.035416030883789064, 0.03544985580444336, 0.035404800415039066, 0.03544371032714844, 0.03545292663574219, 0.03542425537109375, 0.03547955322265625, 0.03544473648071289, 0.035507198333740234, 0.03545395278930664, 0.03543756866455078, 0.03544985580444336, 0.03546623992919922, 0.03542835235595703, 0.03549593734741211, 0.0354856948852539, 0.03549388885498047, 0.03547443389892578, 0.035434497833251956, 0.03545600128173828, 0.03546214294433594, 0.03546931076049804, 0.035504127502441404, 0.03555123138427734, 0.035538944244384765, 0.03549388885498047, 0.035490814208984374, 0.03551129531860352, 0.03548672103881836, 0.03550003051757813, 0.03552972793579102, 0.03552870559692383, 0.03552972793579102, 0.03563212966918945, 0.035722240447998044, 0.035606529235839846, 0.03553177642822265, 0.03556966400146484, 0.03553996658325195, 0.03555327987670898, 0.035538944244384765, 0.03559833526611328, 0.03558297729492187, 0.03555942535400391, 0.03556249618530274, 0.03556454467773437, 0.035568641662597655, 0.0355788803100586, 0.035563518524169925, 0.03562496185302735, 0.03564646530151367, 0.0356577262878418, 0.10359500885009766, 0.035389438629150394, 0.03537612915039062, 0.03538739013671875, 0.0353966064453125, 0.03540582275390625, 0.03549491119384766, 0.03544371032714844, 0.0354150390625, 0.035410945892333984, 0.035407966613769534, 0.03538832092285156, 0.035394561767578124, 0.03542015838623047, 0.03545292663574219, 0.03540889739990234, 0.03543756866455078, 0.03546316909790039, 0.03546214294433594, 0.03548262405395508, 0.03546828842163086, 0.035451904296875, 0.03545702362060547, 0.03544268798828125, 0.035536895751953124, 0.03568025588989258, 0.03563724899291992, 0.035506175994873046, 0.03550822448730469, 0.03550207901000976, 0.03548672103881836, 0.03544678497314453, 0.03545702362060547, 0.035438591003417966, 0.035454975128173825, 0.03544473648071289, 0.03546726226806641, 0.035461185455322265, 0.03546822357177734, 0.03545916748046875, 0.03546928024291992, 0.035480510711669924, 0.03548057556152344, 0.035476478576660156, 0.03548876953125, 0.035501056671142575, 0.03551846313476562, 0.03553385543823242, 0.03550511932373047, 0.035506175994873046, 0.035527679443359376, 0.03552880096435547, 0.03562998580932617, 0.03562393569946289, 0.035555328369140625, 0.035536895751953124, 0.035544063568115236, 0.035538944244384765, 0.035563518524169925, 0.035566593170166014, 0.03556249618530274, 0.035585025787353515, 0.0355788803100586, 0.10331954956054687, 0.03534035110473633, 0.03536684799194336, 0.035366912841796876, 0.035361793518066405, 0.035361793518066405, 0.03537510299682617, 0.03537510299682617, 0.03536076736450195, 0.035382335662841796, 0.03539244842529297, 0.03539865493774414, 0.03541401672363281, 0.03540070343017578, 0.03540991973876953, 0.03539148712158203, 0.03539148712158203, 0.035399681091308595, 0.03542937469482422, 0.03541609573364258, 0.03541705703735352, 0.035410945892333984, 0.035406913757324215, 0.035426239013671874, 0.0354150390625, 0.035451904296875, 0.035451904296875, 0.0354252815246582, 0.035419136047363284, 0.03544268798828125, 0.0354334716796875, 0.035432449340820314, 0.035465217590332034, 0.03545087814331055, 0.03545091247558594, 0.03546015930175781, 0.03545491027832031, 0.03545391845703125, 0.0354856948852539, 0.03547443389892578, 0.035487743377685545, 0.035477664947509764, 0.03547939300537109, 0.03545907211303711, 0.0354703369140625, 0.0354969596862793, 0.035487743377685545, 0.0355153923034668, 0.03548876953125, 0.035504127502441404, 0.0354856948852539, 0.035504127502441404, 0.03551027297973633, 0.035517440795898435, 0.03552665710449219, 0.035550209045410154, 0.03555123138427734, 0.03553279876708984, 0.03555123138427734, 0.035576831817626955, 0.035547134399414065, 0.035536895751953124, 0.035563518524169925, 0.10335539245605468, 0.035381248474121094, 0.035537982940673826, 0.035843006134033205, 0.035692543029785154, 0.03540070343017578, 0.03537919998168945, 0.03537510299682617, 0.03540172958374024, 0.0354252815246582, 0.03539558410644531, 0.03537612915039062, 0.03540787124633789, 0.03544063949584961, 0.03545292663574219, 0.0354068489074707, 0.03544166564941406, 0.03541299057006836, 0.035438591003417966, 0.03543142318725586, 0.03543961715698242, 0.035591167449951173, 0.03558707046508789, 0.03547443389892578, 0.0354703369140625, 0.035429439544677734, 0.035461055755615235, 0.03545292663574219, 0.03547750473022461, 0.03546524810791016, 0.03544982528686524, 0.03551334381103516, 0.03551846313476562, 0.035465217590332034, 0.03545292663574219, 0.03544371032714844, 0.03547750473022461, 0.035476478576660156, 0.03551027297973633, 0.035530750274658206, 0.035522560119628906, 0.03549593734741211, 0.035492897033691406, 0.03552355194091797, 0.035504127502441404, 0.03552153778076172, 0.03553177642822265, 0.035558399200439454, 0.035527679443359376, 0.035525630950927735, 0.035525630950927735, 0.03552159881591797, 0.03552761459350586, 0.035571712493896485, 0.03553279876708984, 0.03558399963378906, 0.03557580947875977, 0.03555123138427734, 0.03554611206054688, 0.035563518524169925, 0.035574783325195314, 0.03558297729492187, 0.035568641662597655]",tokens/s,27.410130257830733,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/gpt-neo-2.7B,EleutherAI/gpt-neo-2.7B,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1224.43776,6568.804352,0.0,5922.357248,5577.085952,s,10,0.32041270256042487,0.03204127025604248,0.0002779190315387718,0.03194198417663574,0.032281372070312495,0.032553661346435545,0.03277149276733399,"[0.032825950622558595, 0.03194095993041992, 0.03185081672668457, 0.03190025520324707, 0.03190140724182129, 0.0319354248046875, 0.032220863342285154, 0.03194300842285156, 0.03194601631164551, 0.031947999954223635]",tokens/s,7989.695725366019,kWh,3.7901002192004796e-07,2.0767856692425713e-07,2.126865127416729e-06,2.713553716261034e-06,tokens/kWh,94341231.74563085,MB,1224.43776,6568.804352,0.0,5922.357248,5675.07712,s,10,11.959532958984374,1.1959532958984376,0.01941915108676585,1.2044320678710938,1.2144798828125,1.2167063720703124,1.2184875634765624,"[1.218932861328125, 1.1679766845703126, 1.16572265625, 1.17552587890625, 1.213985107421875, 1.2078804931640625, 1.21256005859375, 1.1847091064453126, 1.2112564697265624, 1.200983642578125]",tokens/s,52.67764235949735,kWh,1.392984951843648e-05,7.633236873941243e-06,5.260681754838236e-05,7.416990394076011e-05,tokens/kWh,849401.1270436379,,s,629,12.227076101303119,0.019438912720672656,0.003871969311859255,0.01920614433288574,0.019390669250488282,0.019668352127075194,0.05093101486206055,"[0.019633216857910155, 0.019659711837768556, 0.01919795227050781, 0.019481599807739256, 0.019385343551635743, 0.020373504638671876, 0.020290559768676757, 0.02045747184753418, 0.0200468807220459, 0.02011747169494629, 0.020320255279541014, 0.02020966339111328, 0.020152320861816408, 0.020154367446899413, 0.020065280914306642, 0.019543039321899415, 0.01919692802429199, 0.01968639945983887, 0.020368383407592772, 0.020282367706298828, 0.020129791259765627, 0.019965951919555663, 0.019359743118286133, 0.01903104019165039, 0.01933523178100586, 0.019611583709716798, 0.019963903427124022, 0.019519487380981446, 0.01934236717224121, 0.019398624420166016, 0.01938332748413086, 0.019250143051147462, 0.01923072052001953, 0.019269664764404296, 0.01931260871887207, 0.01924710464477539, 0.01923072052001953, 0.018485248565673826, 0.01844223976135254, 0.01845248031616211, 0.01846272087097168, 0.018488319396972656, 0.01846067237854004, 0.01846886444091797, 0.018478080749511717, 0.018526208877563476, 0.01849241638183594, 0.018482175827026368, 0.018463743209838866, 0.01848940849304199, 0.018513856887817384, 0.018504703521728515, 0.018495487213134765, 0.01918976020812988, 0.019307519912719725, 0.019375104904174805, 0.019289087295532227, 0.019390464782714844, 0.019380224227905272, 0.019301376342773437, 0.01924710464477539, 0.01922867202758789, 0.050971649169921876, 0.018421760559082033, 0.018495487213134765, 0.018521087646484375, 0.01845043182373047, 0.01844223976135254, 0.018519039154052733, 0.01847500801086426, 0.018511871337890624, 0.018490367889404297, 0.018481151580810547, 0.01848320007324219, 0.01843404769897461, 0.018510847091674804, 0.018473983764648438, 0.018509824752807616, 0.01842278480529785, 0.018495487213134765, 0.018495487213134765, 0.018531328201293946, 0.018437120437622072, 0.018499584197998048, 0.01841459274291992, 0.01840025520324707, 0.01845964813232422, 0.018515968322753908, 0.018774015426635742, 0.01846886444091797, 0.018516992568969725, 0.018519039154052733, 0.018529279708862305, 0.018503679275512695, 0.018537504196166992, 0.018545631408691407, 0.018485248565673826, 0.018585599899291993, 0.0184453125, 0.018563072204589845, 0.018488319396972656, 0.018463743209838866, 0.018505727767944336, 0.018482175827026368, 0.018532352447509767, 0.01857539176940918, 0.01851593589782715, 0.01848422431945801, 0.018606111526489257, 0.01849443244934082, 0.018555904388427736, 0.018506752014160157, 0.018581504821777343, 0.018511871337890624, 0.018511871337890624, 0.018576383590698242, 0.018528255462646484, 0.018482175827026368, 0.018512895584106445, 0.018508800506591795, 0.01846272087097168, 0.018513919830322266, 0.018618368148803712, 0.019286016464233398, 0.01930342483520508, 0.050871295928955076, 0.018464832305908202, 0.018494400024414062, 0.018516992568969725, 0.01848320007324219, 0.018509824752807616, 0.01844633674621582, 0.018482175827026368, 0.018465791702270508, 0.01845964813232422, 0.019247135162353515, 0.018760671615600587, 0.01845043182373047, 0.018503679275512695, 0.018554880142211915, 0.018509824752807616, 0.018530303955078126, 0.018494464874267577, 0.018602048873901367, 0.01856812858581543, 0.01850476837158203, 0.018445247650146483, 0.01846784019470215, 0.018537471771240235, 0.01847603225708008, 0.018463743209838866, 0.018473983764648438, 0.018507776260375978, 0.018506752014160157, 0.01847091293334961, 0.018522111892700196, 0.018447359085083007, 0.018473983764648438, 0.01847091293334961, 0.01840025520324707, 0.018496511459350586, 0.018481151580810547, 0.018503679275512695, 0.018514944076538087, 0.01844121551513672, 0.018463743209838866, 0.018473983764648438, 0.018497535705566406, 0.01860095977783203, 0.018571264266967775, 0.018464767456054687, 0.018489343643188477, 0.018530303955078126, 0.018469888687133788, 0.0186296329498291, 0.018538496017456055, 0.01850268745422363, 0.018548704147338866, 0.018526208877563476, 0.018490367889404297, 0.018338815689086914, 0.018366464614868162, 0.018301952362060548, 0.018234367370605468, 0.018501632690429686, 0.01844633674621582, 0.018309120178222657, 0.018298879623413086, 0.05072076797485352, 0.01849241638183594, 0.01843916893005371, 0.018301952362060548, 0.018315263748168945, 0.01824051284790039, 0.01826304054260254, 0.018260000228881836, 0.018261983871459962, 0.018249727249145507, 0.01829478454589844, 0.0184586238861084, 0.018456575393676757, 0.018547712326049806, 0.018495487213134765, 0.0185743350982666, 0.01867263984680176, 0.01904742431640625, 0.01859686470031738, 0.01855897521972656, 0.018321407318115233, 0.018335744857788085, 0.018257919311523436, 0.01820467185974121, 0.018330623626708984, 0.01828659248352051, 0.018388992309570314, 0.01841152000427246, 0.01843916893005371, 0.018496511459350586, 0.018463743209838866, 0.018655231475830078, 0.01842278480529785, 0.018510847091674804, 0.018594816207885743, 0.018489343643188477, 0.018513919830322266, 0.018469888687133788, 0.018464767456054687, 0.018515968322753908, 0.019348480224609374, 0.019319807052612305, 0.01927168083190918, 0.019171327590942384, 0.019359743118286133, 0.019219455718994142, 0.019886079788208007, 0.01941196823120117, 0.01920614433288574, 0.01864806365966797, 0.018577407836914063, 0.01863270378112793, 0.018490367889404297, 0.018547712326049806, 0.018511871337890624, 0.018585664749145508, 0.01853129577636719, 0.018781152725219727, 0.019284992218017577, 0.01924403190612793, 0.01919795227050781, 0.019192832946777344, 0.019188735961914064, 0.05095423889160156, 0.01927475166320801, 0.01922150421142578, 0.019185663223266602, 0.019252223968505858, 0.019144704818725586, 0.01921023941040039, 0.019171327590942384, 0.019269632339477538, 0.01920614433288574, 0.01925529670715332, 0.01923891258239746, 0.01918156814575195, 0.01923583984375, 0.01925939178466797, 0.019401727676391603, 0.01926246452331543, 0.019290111541748048, 0.01925017547607422, 0.019535871505737306, 0.019169279098510742, 0.019252223968505858, 0.019346431732177736, 0.019178495407104493, 0.019521535873413084, 0.019319807052612305, 0.01927577590942383, 0.01929113578796387, 0.019364864349365234, 0.01938640022277832, 0.019386335372924803, 0.019280895233154297, 0.019182592391967773, 0.01927065658569336, 0.01927884864807129, 0.019312639236450196, 0.019169279098510742, 0.01926144027709961, 0.019215360641479492, 0.01923072052001953, 0.019301376342773437, 0.019384319305419923, 0.019352575302124024, 0.019258399963378907, 0.01932796859741211, 0.019284992218017577, 0.019300352096557616, 0.01926348876953125, 0.019232831954956054, 0.019264448165893556, 0.019298303604125978, 0.01925324821472168, 0.01928704071044922, 0.01927987289428711, 0.019269632339477538, 0.01926144027709961, 0.01925529670715332, 0.019368959426879884, 0.019288063049316406, 0.0192225284576416, 0.01923686408996582, 0.01922150421142578, 0.019224607467651367, 0.05106172943115234, 0.0186296329498291, 0.01863884735107422, 0.018530303955078126, 0.018486272811889647, 0.01842483139038086, 0.01863987159729004, 0.01860403251647949, 0.018501632690429686, 0.018457599639892578, 0.018505727767944336, 0.018511871337890624, 0.018565120697021483, 0.018719743728637696, 0.020320255279541014, 0.01967411231994629, 0.01928294372558594, 0.019363840103149413, 0.019325952529907226, 0.01924812889099121, 0.01925939178466797, 0.01927577590942383, 0.019277824401855468, 0.01924198341369629, 0.01920614433288574, 0.01920102310180664, 0.019338239669799806, 0.01923686408996582, 0.01937411117553711, 0.019382272720336914, 0.019409887313842772, 0.019176448822021484, 0.01926348876953125, 0.019280895233154297, 0.019129344940185547, 0.01925529670715332, 0.019353599548339845, 0.01925324821472168, 0.019377151489257814, 0.01924505615234375, 0.01926348876953125, 0.01922764778137207, 0.0192542724609375, 0.01923993682861328, 0.019301376342773437, 0.019281919479370118, 0.019317760467529296, 0.019252223968505858, 0.019281919479370118, 0.019277856826782225, 0.01945699119567871, 0.019335168838500977, 0.019362815856933592, 0.019348480224609374, 0.019755008697509766, 0.019346431732177736, 0.019280895233154297, 0.01924505615234375, 0.019579904556274414, 0.019311616897583008, 0.019314687728881837, 0.019389440536499023, 0.019400703430175782, 0.05290086364746094, 0.0194334716796875, 0.019324928283691405, 0.019294208526611328, 0.019224576950073242, 0.019211263656616212, 0.01924403190612793, 0.01922969627380371, 0.01960550308227539, 0.019281919479370118, 0.019346431732177736, 0.019277824401855468, 0.019198976516723632, 0.0193832950592041, 0.018531328201293946, 0.018554880142211915, 0.01850060844421387, 0.018494464874267577, 0.018967552185058592, 0.019288063049316406, 0.01923788833618164, 0.019191808700561523, 0.019322879791259767, 0.019365888595581054, 0.019288063049316406, 0.01928704071044922, 0.018523231506347656, 0.018533279418945312, 0.018509824752807616, 0.019338239669799806, 0.019122175216674805, 0.01920204734802246, 0.019180543899536134, 0.019260416030883788, 0.01939967918395996, 0.01929523277282715, 0.01922355270385742, 0.01926553535461426, 0.019200000762939453, 0.019358720779418945, 0.01939558410644531, 0.019293184280395507, 0.01925529670715332, 0.01927065658569336, 0.01939148712158203, 0.01924710464477539, 0.01946931266784668, 0.019218496322631836, 0.01933305549621582, 0.01924403190612793, 0.019315711975097655, 0.01922150421142578, 0.019259456634521485, 0.019472320556640624, 0.020176895141601564, 0.019950592041015625, 0.019332096099853514, 0.019352575302124024, 0.01924198341369629, 0.01922764778137207, 0.01925939178466797, 0.0192174072265625, 0.01927168083190918, 0.05098086547851562, 0.018482175827026368, 0.018447359085083007, 0.01846681594848633, 0.018469888687133788, 0.01841152000427246, 0.01840640068054199, 0.01843916893005371, 0.01842278480529785, 0.018431999206542968, 0.01845452880859375, 0.018435136795043945, 0.018371519088745118, 0.01845248031616211, 0.018573312759399413, 0.018508800506591795, 0.018512895584106445, 0.018437120437622072, 0.01848320007324219, 0.018536447525024414, 0.01841971206665039, 0.01845248031616211, 0.01838591957092285, 0.01839820861816406, 0.018440256118774413, 0.01850156784057617, 0.018465791702270508, 0.018447359085083007, 0.01900441551208496, 0.01924812889099121, 0.01942425537109375, 0.01942835235595703, 0.01925632095336914, 0.01923276710510254, 0.019327999114990235, 0.0192174072265625, 0.01925119972229004, 0.019169279098510742, 0.018537471771240235, 0.018576383590698242, 0.018455551147460936, 0.018472959518432617, 0.018548736572265623, 0.018532352447509767, 0.01858355140686035, 0.018577407836914063, 0.018527231216430663, 0.018799615859985352, 0.01925836753845215, 0.01924198341369629, 0.019336191177368164, 0.019294208526611328, 0.019362848281860353, 0.019238880157470703, 0.01925836753845215, 0.019308544158935546, 0.01922867202758789, 0.019224576950073242, 0.019401727676391603, 0.01922047996520996, 0.019330047607421876, 0.019339263916015623, 0.01923481559753418, 0.051972095489501956, 0.01927987289428711, 0.019349536895751952, 0.01934332847595215, 0.01925324821472168, 0.019148799896240236, 0.019187711715698243, 0.019207168579101562, 0.01927884864807129, 0.019310592651367187, 0.01927168083190918, 0.019276800155639647, 0.01949286460876465, 0.019309568405151366, 0.019317760467529296, 0.01922047996520996, 0.019288063049316406, 0.019386367797851564, 0.0192542724609375, 0.019336191177368164, 0.019272703170776367, 0.019374080657958984, 0.019524608612060547, 0.01924505615234375, 0.01923583984375, 0.01922969627380371, 0.019198976516723632, 0.01922969627380371, 0.019106815338134766, 0.019327999114990235, 0.01921843147277832, 0.019294208526611328, 0.01922047996520996, 0.019191808700561523, 0.019168256759643554, 0.01927884864807129, 0.01927577590942383, 0.019172351837158205, 0.019216384887695313, 0.01921023941040039, 0.01924198341369629, 0.019174400329589843, 0.019288063049316406, 0.019311616897583008, 0.019303455352783203, 0.019271648406982422, 0.019201120376586913, 0.019274656295776366, 0.019296255111694336, 0.019170303344726563, 0.019309568405151366, 0.019268608093261717, 0.019327999114990235, 0.019316736221313476, 0.01921331214904785, 0.0192491512298584, 0.019273727416992188, 0.0186060791015625, 0.01845043182373047, 0.018533376693725585, 0.018503679275512695, 0.01886310386657715, 0.019301376342773437, 0.05171104049682617, 0.019302335739135743, 0.019169279098510742, 0.01925939178466797, 0.01920614433288574, 0.02011955261230469, 0.019392511367797852, 0.01924403190612793, 0.01926144027709961, 0.01943449592590332, 0.019347455978393553, 0.01919795227050781, 0.01921331214904785, 0.01924403190612793, 0.019215360641479492, 0.019224576950073242, 0.01923174476623535, 0.019376127243041993, 0.019164159774780275, 0.01924198341369629, 0.01922662353515625, 0.019145727157592773, 0.019198976516723632, 0.01919692802429199, 0.01925324821472168, 0.01925836753845215, 0.01927168083190918, 0.0192542724609375, 0.01926553535461426, 0.019405824661254883, 0.019159040451049804, 0.0184453125, 0.01828659248352051, 0.018989055633544923, 0.01840230369567871, 0.018499616622924805, 0.018381792068481444, 0.018274303436279296, 0.018568191528320312, 0.018888704299926756, 0.019260416030883788, 0.01923481559753418, 0.019393535614013673, 0.019179519653320314, 0.01926348876953125, 0.01924403190612793, 0.01925939178466797, 0.01925119972229004, 0.018553855895996094, 0.01882828712463379, 0.019148799896240236, 0.019211263656616212, 0.01925119972229004, 0.019224576950073242, 0.018720767974853517, 0.018320415496826174, 0.018784223556518556, 0.018537471771240235, 0.018315263748168945, 0.018371583938598633, 0.018564096450805666, 0.018747392654418944, 0.019340288162231444]",tokens/s,51.44320643698,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,1,1,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/1/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6693261c-5a247c995982ca14416c97f9;c0000257-cc85-45ff-841d-062c97f05184) Repository Not Found for url: https://huggingface.co/1/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: 1 is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-7b,tiiuae/falcon-7b,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ self.load_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 250, in load_model_with_no_weights self.pretrained_model = self.automodel_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 559, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3704, in from_pretrained config = cls._autoset_attn_implementation( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1481, in _autoset_attn_implementation cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1572, in _check_and_enable_flash_attn_2 raise ValueError( ValueError: FalconForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpm9a8bh80/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,meta-llama/Llama-2-7b-hf,meta-llama/Llama-2-7b-hf,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.1,,,,MB,1235.451904,14563.147776,0.0,13916.700672,13738.740736,s,10,0.9376921615600585,0.09376921615600585,0.0004724126133072065,0.09365096282958985,0.09393276901245118,0.09454305610656738,0.09503128578186035,"[0.0951533432006836, 0.09379714965820313, 0.09371504211425781, 0.09355184173583984, 0.0934661102294922, 0.09347478485107422, 0.09365590667724609, 0.09354041290283203, 0.0936915512084961, 0.09364601898193359]",tokens/s,2730.107070257336,kWh,1.1095883643887124e-06,6.079862246997861e-07,6.014490792897036e-06,7.732065381985534e-06,tokens/kWh,33108876.7816732,MB,1235.451904,14563.147776,0.0,13916.700672,13889.779712,s,10,22.594381347656253,2.259438134765625,0.0034224782668009815,2.25855224609375,2.2615166259765624,2.265331188964844,2.268382839355469,"[2.269145751953125, 2.2606689453125, 2.259339599609375, 2.256840087890625, 2.256826171875, 2.258702880859375, 2.258401611328125, 2.257575927734375, 2.25783837890625, 2.2590419921875]",tokens/s,27.883038278689178,kWh,2.6614463425716262e-05,1.4585633785131862e-05,0.00011539618966890334,0.00015659628687975148,tokens/kWh,402308.38965151826,,s,629,23.400172534942634,0.03720218209052882,0.01118870796955968,0.03584415817260742,0.03596513290405273,0.03607998580932617,0.12996663513183593,"[0.03646054458618164, 0.03655987167358398, 0.03643084716796875, 0.036604927062988284, 0.036539390563964845, 0.03630080032348633, 0.03634175872802734, 0.036560897827148435, 0.036391937255859375, 0.036375553131103515, 0.03626496124267578, 0.036127742767333985, 0.03625267028808594, 0.036103168487548826, 0.03593836975097656, 0.03575392150878906, 0.03577139282226562, 0.035743743896484374, 0.03577446365356445, 0.03573452758789063, 0.035810367584228515, 0.03576723098754883, 0.0358205451965332, 0.03642371368408203, 0.03593417739868164, 0.035817470550537106, 0.03583798217773437, 0.03594543838500976, 0.03581644821166992, 0.03579289627075195, 0.03587583923339844, 0.03586867141723633, 0.03585433578491211, 0.03582156753540039, 0.0358922233581543, 0.0358656005859375, 0.03593523025512695, 0.03596799850463867, 0.03592704010009766, 0.03589120101928711, 0.035914752960205076, 0.03585228729248047, 0.03586969757080078, 0.03585945510864258, 0.03591167831420899, 0.03591990280151367, 0.035928031921386716, 0.03587276840209961, 0.03599564743041992, 0.035931137084960936, 0.035917823791503906, 0.035904510498046875, 0.03593734359741211, 0.03593209457397461, 0.03594956970214844, 0.035917823791503906, 0.035932159423828124, 0.035904510498046875, 0.03596083068847656, 0.035912704467773435, 0.03594035339355469, 0.035914752960205076, 0.13034393310546874, 0.035794944763183595, 0.03573657608032227, 0.035737598419189456, 0.03572633743286133, 0.03578572845458984, 0.035722240447998044, 0.035745792388916016, 0.035757057189941405, 0.035757057189941405, 0.03575193786621094, 0.03574784088134766, 0.0357918701171875, 0.03577446365356445, 0.035778560638427735, 0.03579391860961914, 0.035784702301025394, 0.035827713012695314, 0.03580416107177734, 0.03579289627075195, 0.03581235122680664, 0.035811328887939455, 0.03578572845458984, 0.03577446365356445, 0.03578777694702148, 0.03585126495361328, 0.035855358123779296, 0.0358287353515625, 0.0358389778137207, 0.03583180618286133, 0.03583795166015625, 0.03583795166015625, 0.03586867141723633, 0.03588614273071289, 0.035859390258789064, 0.03586764907836914, 0.035914752960205076, 0.035932159423828124, 0.03597619247436523, 0.03589529418945313, 0.035904510498046875, 0.035955711364746096, 0.03618105697631836, 0.03609183883666992, 0.03602431869506836, 0.0359628791809082, 0.03593625640869141, 0.03593830490112305, 0.035975166320800785, 0.035945472717285154, 0.035939327239990236, 0.035933246612548826, 0.03597100830078125, 0.03594649505615234, 0.035961856842041014, 0.035963905334472655, 0.03598233413696289, 0.03601100921630859, 0.0359628791809082, 0.03595673751831055, 0.03597107315063477, 0.03600896072387695, 0.03596799850463867, 0.1301370849609375, 0.03570687866210937, 0.03569868850708008, 0.035983360290527344, 0.036055038452148434, 0.03576627349853516, 0.03580313491821289, 0.03577139282226562, 0.0357386245727539, 0.035729408264160156, 0.035762176513671876, 0.03574784088134766, 0.03577958297729492, 0.03575500869750976, 0.03575398254394531, 0.035776512145996094, 0.03576319885253906, 0.035746814727783204, 0.035768318176269534, 0.03578675079345703, 0.03578572845458984, 0.03575193786621094, 0.03578879928588867, 0.035765247344970705, 0.03579084777832031, 0.03581542587280274, 0.035789825439453124, 0.035833889007568356, 0.03581436920166016, 0.03582156753540039, 0.03581235122680664, 0.035833854675292966, 0.03589734268188476, 0.03590963363647461, 0.03586457443237305, 0.03583180618286133, 0.03601919937133789, 0.03587891387939453, 0.03587071990966797, 0.03585433578491211, 0.03586355209350586, 0.035860481262207033, 0.035851295471191404, 0.035944416046142576, 0.036034561157226565, 0.035942401885986325, 0.035920894622802735, 0.035896320343017575, 0.03594649505615234, 0.0359332160949707, 0.035931102752685545, 0.03596083068847656, 0.03591884613037109, 0.03590963363647461, 0.03595673751831055, 0.035912704467773435, 0.035914752960205076, 0.03592704010009766, 0.03593523025512695, 0.03594956970214844, 0.03593318557739258, 0.03594137573242188, 0.03593318557739258, 0.12997120666503906, 0.03568537521362305, 0.03565465545654297, 0.03570483016967774, 0.03568025588989258, 0.03570687866210937, 0.035692543029785154, 0.03572326278686523, 0.035724288940429685, 0.03569561767578125, 0.03570073699951172, 0.0357212142944336, 0.03571507263183594, 0.0357314567565918, 0.03571916961669922, 0.035729408264160156, 0.0357283821105957, 0.035722270965576175, 0.03575190353393555, 0.03576627349853516, 0.035743743896484374, 0.03575296020507813, 0.03575500869750976, 0.03577753448486328, 0.035759105682373046, 0.035796993255615236, 0.035775489807128906, 0.035794944763183595, 0.03580928039550781, 0.03579289627075195, 0.035810302734375, 0.03579296112060547, 0.03581024169921875, 0.03581849670410156, 0.03580416107177734, 0.03583488082885742, 0.03582156753540039, 0.035829761505126956, 0.035844097137451174, 0.03583078384399414, 0.0358922233581543, 0.035901439666748046, 0.03584511947631836, 0.03583795166015625, 0.035884033203125, 0.03589017486572266, 0.03586764907836914, 0.035922943115234376, 0.03587788772583008, 0.03591884613037109, 0.035914752960205076, 0.035952640533447267, 0.03589120101928711, 0.035934207916259765, 0.03590655899047852, 0.03592396926879883, 0.03594137573242188, 0.035922943115234376, 0.03595161437988281, 0.03590553665161133, 0.035920894622802735, 0.03598643112182617, 0.035961856842041014, 0.12988108825683595, 0.035672065734863284, 0.03567308807373047, 0.035645439147949216, 0.035699710845947266, 0.03571507263183594, 0.03571814346313477, 0.03570278549194336, 0.035692543029785154, 0.035714046478271484, 0.03567923355102539, 0.03571200180053711, 0.03570175933837891, 0.03573964691162109, 0.03570892715454101, 0.035730430603027344, 0.03581644821166992, 0.03594649505615234, 0.035705856323242184, 0.0357498893737793, 0.03573356628417969, 0.03575084686279297, 0.035746814727783204, 0.035776512145996094, 0.0357314567565918, 0.03576627349853516, 0.035743743896484374, 0.035767295837402346, 0.035760128021240234, 0.03579084777832031, 0.03577446365356445, 0.03579084777832031, 0.03578060913085938, 0.03582259368896484, 0.03580108642578125, 0.03581235122680664, 0.035817470550537106, 0.03585843276977539, 0.03584307098388672, 0.03584716796875, 0.03606220626831055, 0.036261886596679685, 0.0360079345703125, 0.03586150360107422, 0.03582463836669922, 0.035866622924804685, 0.035866622924804685, 0.03587276840209961, 0.03585945510864258, 0.03590041732788086, 0.035866622924804685, 0.03586764907836914, 0.03587788772583008, 0.03589324951171875, 0.03585945510864258, 0.03589017486572266, 0.03587686538696289, 0.035932159423828124, 0.0358809585571289, 0.03589120101928711, 0.0359628791809082, 0.035915775299072264, 0.03589734268188476, 0.1299548797607422, 0.035752895355224606, 0.035730430603027344, 0.03571814346313477, 0.0357212142944336, 0.03569356918334961, 0.03572636795043945, 0.035742687225341796, 0.03572633743286133, 0.03577139282226562, 0.03571814346313477, 0.035727359771728515, 0.035737598419189456, 0.03577241516113281, 0.03574272155761719, 0.03576627349853516, 0.035800064086914066, 0.035767295837402346, 0.035778560638427735, 0.035784702301025394, 0.03578572845458984, 0.03579084777832031, 0.03578572845458984, 0.03579904174804688, 0.035814399719238284, 0.035817470550537106, 0.03577446365356445, 0.03581849670410156, 0.03582463836669922, 0.03582463836669922, 0.03581849670410156, 0.035829761505126956, 0.03584307098388672, 0.03584204864501953, 0.03587891387939453, 0.0358205451965332, 0.03586764907836914, 0.035871742248535156, 0.035885055541992186, 0.03586150360107422, 0.035866622924804685, 0.03589324951171875, 0.03590553665161133, 0.03589427185058594, 0.035860481262207033, 0.035883007049560545, 0.03587686538696289, 0.03589836883544922, 0.03591680145263672, 0.035896320343017575, 0.03589324951171875, 0.03590655899047852, 0.03590553665161133, 0.0359659538269043, 0.035934207916259765, 0.03598745727539063, 0.03592499160766602, 0.035988479614257815, 0.03616972732543945, 0.0361267204284668, 0.0359444465637207, 0.0359659538269043, 0.035937278747558594, 0.13011354064941405, 0.0357212142944336, 0.035722240447998044, 0.035748863220214845, 0.0357314567565918, 0.035716094970703126, 0.035699710845947266, 0.03571507263183594, 0.035716094970703126, 0.03574169540405273, 0.035736640930175784, 0.03573958587646484, 0.03574169540405273, 0.03575193786621094, 0.0357314567565918, 0.035746849060058594, 0.03578467178344726, 0.035768318176269534, 0.03577958297729492, 0.035850273132324216, 0.036018142700195314, 0.03598950576782227, 0.03581542587280274, 0.035810302734375, 0.035762176513671876, 0.03578675079345703, 0.03583590316772461, 0.03582259368896484, 0.03581542587280274, 0.03581235122680664, 0.035833854675292966, 0.03585433578491211, 0.03582156753540039, 0.03581951904296875, 0.03582463836669922, 0.03582156753540039, 0.03583692932128906, 0.03586150360107422, 0.0358737907409668, 0.03585638427734375, 0.03586150360107422, 0.03588608169555664, 0.035884033203125, 0.03586355209350586, 0.03588710403442383, 0.03589017486572266, 0.035912704467773435, 0.03592704010009766, 0.03591987228393555, 0.035929088592529294, 0.03593523025512695, 0.03590860748291016, 0.03590758514404297, 0.035917823791503906, 0.03592704010009766, 0.03591167831420899, 0.03594649505615234, 0.03590963363647461, 0.03591167831420899, 0.03591884613037109, 0.03592396926879883, 0.03592601776123047, 0.035950592041015625, 0.129976318359375, 0.03573350524902344, 0.03577859115600586, 0.03569558334350586, 0.035699710845947266, 0.0356864013671875, 0.035710975646972655, 0.03575296020507813, 0.035844097137451174, 0.0357386245727539, 0.03573555374145508, 0.0357314567565918, 0.035800064086914066, 0.035767295837402346, 0.03573350524902344, 0.0357386245727539, 0.03578777694702148, 0.035767295837402346, 0.03573452758789063, 0.035746814727783204, 0.03576627349853516, 0.03575296020507813, 0.03576319885253906, 0.03575296020507813, 0.03575507354736328, 0.035772350311279295, 0.035776512145996094, 0.03577753448486328, 0.03578777694702148, 0.03582566452026367, 0.03580723190307617, 0.035817470550537106, 0.03579391860961914, 0.035827713012695314, 0.03582463836669922, 0.03584415817260742, 0.035853248596191406, 0.0358656005859375, 0.035850238800048825, 0.03590553665161133, 0.03598950576782227, 0.0358737907409668, 0.035862529754638675, 0.03587788772583008, 0.0358656005859375, 0.03586150360107422, 0.035917823791503906, 0.03586969757080078, 0.03588915252685547, 0.03587689590454102, 0.03588297653198242, 0.03588198471069336, 0.03592806243896484, 0.03593427276611328, 0.03590854263305664, 0.03591167831420899, 0.03592396926879883, 0.03592704010009766, 0.0359659538269043, 0.035929088592529294, 0.035934207916259765, 0.035950592041015625, 0.035914752960205076, 0.13009408569335937, 0.03570892715454101, 0.03565363311767578, 0.03571814346313477, 0.035675136566162106, 0.03566694259643555, 0.03572326278686523, 0.035759105682373046, 0.03570175933837891, 0.035732479095458985, 0.03570380783081055, 0.035722240447998044, 0.03576425552368164, 0.03581846237182617, 0.03573555374145508, 0.0357283821105957, 0.035732513427734376, 0.03576931381225586, 0.035724288940429685, 0.03575193786621094, 0.03574169540405273, 0.035759105682373046, 0.035743743896484374, 0.03575091171264649, 0.03574784088134766, 0.03578879928588867, 0.03575398254394531, 0.03577241516113281, 0.03578265762329102, 0.03580108642578125, 0.03582876968383789, 0.0358583984375, 0.03584000015258789, 0.03598950576782227, 0.03623833465576172, 0.035952640533447267, 0.03584000015258789, 0.035871742248535156, 0.03581951904296875, 0.03580825424194336, 0.03585126495361328, 0.03580518341064453, 0.03580723190307617, 0.0358359375, 0.03583075332641601, 0.03589529418945313, 0.035881023406982425, 0.03593414306640625, 0.03597107315063477, 0.03588608169555664, 0.0358809585571289, 0.03587891387939453, 0.035885055541992186, 0.035888126373291016, 0.03587583923339844, 0.03589120101928711, 0.03594137573242188, 0.03625369644165039, 0.036125694274902344, 0.0359444465637207, 0.03589427185058594, 0.035939327239990236, 0.03590041732788086, 0.13008384704589843, 0.03574272155761719, 0.03569868850708008, 0.03570687866210937, 0.03572531127929687, 0.035714046478271484, 0.03572326278686523, 0.03574784088134766, 0.035762176513671876, 0.03578060913085938, 0.0357283821105957, 0.03574784088134766, 0.03575296020507813, 0.03578169631958008, 0.03576723098754883, 0.035775489807128906, 0.03576319885253906, 0.035767295837402346, 0.03573350524902344, 0.03577446365356445, 0.03577446365356445, 0.03582361602783203, 0.035806209564208984, 0.03585228729248047, 0.03577446365356445, 0.035871742248535156, 0.03584716796875, 0.03581644821166992, 0.03583590316772461, 0.03582259368896484, 0.03582566452026367, 0.03585433578491211, 0.03588608169555664, 0.03589120101928711, 0.03589734268188476, 0.03587481689453125, 0.035866622924804685, 0.03590553665161133, 0.03595161437988281, 0.035966976165771485, 0.035866622924804685, 0.03589836883544922, 0.0358737907409668, 0.03588710403442383, 0.03585740661621094, 0.035884033203125, 0.03589017486572266, 0.03594137573242188, 0.03590655899047852, 0.03594342422485351, 0.035950592041015625, 0.0359813117980957, 0.035937278747558594, 0.035939327239990236, 0.03591987228393555, 0.03595065689086914, 0.03592902374267578, 0.03594854354858398, 0.03595980834960937, 0.03596492767333984, 0.03599155044555664, 0.03596492767333984, 0.03591372680664062]",tokens/s,26.880143685297067,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-rw-1b,tiiuae/falcon-rw-1b,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ self.load_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 250, in load_model_with_no_weights self.pretrained_model = self.automodel_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 559, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3704, in from_pretrained config = cls._autoset_attn_implementation( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1481, in _autoset_attn_implementation cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1572, in _check_and_enable_flash_attn_2 raise ValueError( ValueError: FalconForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpdgesj8w9/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,8,8,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/8/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6693244d-3e8365621e70dcc524737454;20d49f01-19bc-42be-974f-1d1e611960f0) Repository Not Found for url: https://huggingface.co/8/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: 8 is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-4B,,cuda,0,42,,,,,True,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,217063f5c507ed7cc255df7e1f64c4333a0b4dfe,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,,MB,1302.646784,10129.768448,0.0,9483.321344,8972.087296,s,10,0.5423986625671388,0.05423986625671388,0.0016276465308673474,0.05385987281799316,0.05461903495788575,0.056808237648010254,0.058559599800109864,"[0.05899744033813477, 0.05392371368408203, 0.0537050895690918, 0.0539073600769043, 0.05367036819458008, 0.052700031280517576, 0.05413254547119141, 0.05363488006591797, 0.05391484832763672, 0.05381238555908203]",tokens/s,4719.7756496737675,kWh,6.290662461279696e-07,3.4450458291022626e-07,2.9854589486488546e-06,3.959029777687051e-06,tokens/kWh,64662307.27609243,MB,1302.646784,10129.768448,0.0,9483.321344,9040.66816,s,10,23.94006005859375,2.3940060058593744,0.0085533717680495,2.39621923828125,2.40174501953125,2.404019848632813,2.4058397119140627,"[2.406294677734375, 2.395516845703125, 2.397144775390625, 2.396314453125, 2.38508251953125, 2.37354248046875, 2.401239501953125, 2.3961240234375, 2.3963525390625, 2.3924482421875]",tokens/s,26.315723455081702,kWh,2.75197861419361e-05,1.5081964739536566e-05,8.02084409397519e-05,0.00012281019182122454,tokens/kWh,512986.7404792384,,s,629,24.384089099884065,0.03876643736070594,0.00646733166231255,0.038001632690429686,0.03830620803833008,0.03868409652709961,0.09201754364013674,"[0.03917004776000976, 0.039120960235595706, 0.03913004684448242, 0.038708225250244144, 0.03921612930297851, 0.0391014404296875, 0.03893964767456055, 0.03885567855834961, 0.039400447845458986, 0.03895296096801758, 0.038740993499755856, 0.03890486526489258, 0.03876553726196289, 0.03816755294799805, 0.03803647994995117, 0.03800985717773438, 0.03831196975708008, 0.03798729705810547, 0.03793305587768555, 0.03799552154541016, 0.03790028762817383, 0.037972991943359374, 0.037838848114013675, 0.03789004898071289, 0.03801804733276367, 0.03806105422973633, 0.03764019012451172, 0.037833728790283204, 0.03789311981201172, 0.038019073486328124, 0.03794841766357422, 0.03792998504638672, 0.03777433776855469, 0.03779481506347656, 0.03756851196289063, 0.0377149429321289, 0.03807027053833008, 0.037953536987304685, 0.03784806442260742, 0.037926910400390625, 0.03802624130249024, 0.037976062774658204, 0.03796889495849609, 0.03831500625610352, 0.03794124984741211, 0.03783065414428711, 0.037618686676025394, 0.03793100738525391, 0.03792281723022461, 0.03808665466308594, 0.03799039840698242, 0.037996543884277346, 0.038005760192871094, 0.03798425674438476, 0.03798019027709961, 0.037964832305908205, 0.0379749755859375, 0.03798425674438476, 0.03804265594482422, 0.037966815948486325, 0.038012928009033206, 0.03805286407470703, 0.09257062530517578, 0.03809280014038086, 0.03805081558227539, 0.0381102066040039, 0.03806515121459961, 0.03803033447265625, 0.03801599884033203, 0.0379156494140625, 0.03787161636352539, 0.03803340911865234, 0.03809075164794922, 0.0379422721862793, 0.03794636917114258, 0.03805491256713867, 0.03799039840698242, 0.03784499359130859, 0.037868545532226565, 0.0377262077331543, 0.03781228637695312, 0.03780089569091797, 0.03778252792358398, 0.03770470428466797, 0.03799347305297852, 0.037705726623535156, 0.0377784309387207, 0.037937152862548826, 0.03832012939453125, 0.0378152961730957, 0.037874687194824216, 0.03797196960449219, 0.03804876708984375, 0.03801599884033203, 0.03841740798950195, 0.03808563232421875, 0.038056961059570314, 0.03803443145751953, 0.03797094345092773, 0.038005760192871094, 0.03802214431762695, 0.0379607048034668, 0.03810201644897461, 0.038002689361572264, 0.03800166320800781, 0.038035457611083984, 0.03804159927368164, 0.03785318374633789, 0.03801804733276367, 0.038019073486328124, 0.03806208038330078, 0.038042625427246096, 0.03796582412719727, 0.03797401428222656, 0.0379422721862793, 0.038019073486328124, 0.037966846466064456, 0.03815935897827148, 0.03802624130249024, 0.038280193328857424, 0.03945369720458984, 0.0382371826171875, 0.038007808685302735, 0.037969921112060545, 0.0379576301574707, 0.09151692962646485, 0.03800681686401367, 0.037881824493408205, 0.0379791374206543, 0.038004737854003906, 0.03784294509887695, 0.03712307357788086, 0.03794841766357422, 0.037694465637207034, 0.03739033508300781, 0.03882393646240234, 0.038730751037597655, 0.03816960144042969, 0.03791667175292969, 0.03792281723022461, 0.03796889495849609, 0.0377968635559082, 0.03803238296508789, 0.037917697906494144, 0.03686707305908203, 0.037738494873046875, 0.037855232238769534, 0.03799039840698242, 0.03800371170043945, 0.03783679962158203, 0.037910526275634765, 0.037804031372070314, 0.037961727142333986, 0.037645313262939455, 0.039757823944091795, 0.03846553421020508, 0.03799859237670898, 0.038029312133789066, 0.03796889495849609, 0.03799039840698242, 0.038100990295410156, 0.03819007873535156, 0.03791462326049805, 0.037904384613037106, 0.03799961471557617, 0.038084606170654296, 0.03912396621704101, 0.03846758270263672, 0.038204414367675785, 0.03810611343383789, 0.038010879516601564, 0.03836108779907227, 0.03811840057373047, 0.03792486572265625, 0.03820236968994141, 0.038193153381347655, 0.038076416015625, 0.03769036865234375, 0.037897216796875, 0.03802835083007813, 0.03804153442382813, 0.037991424560546876, 0.0389857292175293, 0.03844812774658203, 0.038005760192871094, 0.038076416015625, 0.03815116882324219, 0.03808563232421875, 0.09239347076416016, 0.03808768081665039, 0.038373374938964845, 0.038193153381347655, 0.0381399040222168, 0.03830374526977539, 0.037972991943359374, 0.03798015975952149, 0.03825664138793945, 0.037959678649902344, 0.0377077751159668, 0.03806003189086914, 0.03793203353881836, 0.038141952514648435, 0.03798732757568359, 0.03804876708984375, 0.0379576301574707, 0.037989376068115234, 0.03821363067626953, 0.038351871490478515, 0.0380579833984375, 0.03798425674438476, 0.03798323059082031, 0.03824332809448242, 0.03812761688232422, 0.03716403198242187, 0.03687833786010742, 0.03740364837646484, 0.038145023345947264, 0.037966846466064456, 0.037968959808349606, 0.03816851043701172, 0.03809996795654297, 0.038012928009033206, 0.03830476760864258, 0.03803340911865234, 0.03792588806152344, 0.037953536987304685, 0.03865087890625, 0.03821875381469726, 0.03806924819946289, 0.03794841766357422, 0.03798323059082031, 0.038090782165527345, 0.038021087646484375, 0.037961727142333986, 0.037978111267089845, 0.038076416015625, 0.037976062774658204, 0.03792486572265625, 0.037958656311035156, 0.03809894561767578, 0.03797708892822266, 0.03787059020996094, 0.03784505462646484, 0.03829446411132813, 0.038040576934814455, 0.03802316665649414, 0.03826278305053711, 0.038188030242919925, 0.0380313606262207, 0.03836006546020508, 0.037956607818603515, 0.0922204132080078, 0.037997566223144534, 0.03784912109375, 0.037335006713867185, 0.037594112396240234, 0.03793203353881836, 0.037920768737792966, 0.03808563232421875, 0.037994495391845705, 0.03778867340087891, 0.03788185501098633, 0.03771084976196289, 0.03791667175292969, 0.037964832305908205, 0.03785827255249023, 0.03802214431762695, 0.03770470428466797, 0.037749759674072264, 0.03767091369628906, 0.03772927856445313, 0.0380682258605957, 0.03794841766357422, 0.037969921112060545, 0.03842559814453125, 0.037945343017578126, 0.03797708892822266, 0.03775283050537109, 0.037956607818603515, 0.03791257476806641, 0.03680255889892578, 0.036822017669677735, 0.0369889907836914, 0.03798931121826172, 0.037920768737792966, 0.03791462326049805, 0.03794636917114258, 0.0377149429321289, 0.03784294509887695, 0.0379576301574707, 0.03797708892822266, 0.03815321731567383, 0.03793407821655274, 0.037988414764404295, 0.03811219024658203, 0.03804774475097656, 0.03799347305297852, 0.03818188858032227, 0.03827609634399414, 0.03803033447265625, 0.038386688232421876, 0.03695206451416016, 0.03681894302368164, 0.038043647766113284, 0.03807129669189453, 0.0380682258605957, 0.03799859237670898, 0.038204414367675785, 0.038012928009033206, 0.03801190567016602, 0.036847614288330076, 0.03787776184082031, 0.037975040435791016, 0.03801200103759766, 0.0912475814819336, 0.0380302734375, 0.03681587219238281, 0.03684864044189453, 0.03804467010498047, 0.037956607818603515, 0.037967872619628903, 0.03793920135498047, 0.03818700790405274, 0.038076416015625, 0.038039615631103516, 0.03832928085327148, 0.03799763107299805, 0.03798316955566406, 0.03796275329589844, 0.03686809539794922, 0.036977664947509765, 0.036961280822753906, 0.03693062210083008, 0.036783039093017576, 0.03691212844848633, 0.03689164733886719, 0.03672576141357422, 0.03688447952270508, 0.03689267349243164, 0.03685283279418945, 0.03714863967895508, 0.03684652709960937, 0.03688857650756836, 0.03696537780761719, 0.036923393249511716, 0.03676979064941406, 0.037746688842773435, 0.03803033447265625, 0.037997566223144534, 0.03792588806152344, 0.03794636917114258, 0.03796275329589844, 0.03832729721069336, 0.037994495391845705, 0.03803033447265625, 0.03804876708984375, 0.037917697906494144, 0.037877792358398436, 0.03796271896362305, 0.03808665466308594, 0.03779481506347656, 0.03805491256713867, 0.03784806442260742, 0.03777433776855469, 0.03776409530639648, 0.03798220825195313, 0.03799039840698242, 0.03802828979492188, 0.038214656829833986, 0.03784089660644531, 0.037967872619628903, 0.038035457611083984, 0.0381317138671875, 0.03808051300048828, 0.038073345184326174, 0.038095870971679685, 0.03809280014038086, 0.09410457611083985, 0.038338558197021484, 0.03813478469848633, 0.038158336639404294, 0.037996543884277346, 0.038421504974365236, 0.03823616027832031, 0.03811532974243164, 0.037991424560546876, 0.03808665466308594, 0.03807027053833008, 0.03798425674438476, 0.03827711868286133, 0.037988353729248046, 0.037988353729248046, 0.03809382247924804, 0.038362110137939456, 0.039109630584716795, 0.03829862213134766, 0.0379791374206543, 0.03800166320800781, 0.038042625427246096, 0.038073345184326174, 0.038182910919189454, 0.03806003189086914, 0.0380324478149414, 0.03852896118164063, 0.03809075164794922, 0.0380579833984375, 0.0381328010559082, 0.038096832275390624, 0.03703705596923828, 0.03692854309082031, 0.03707183837890625, 0.03697663879394531, 0.036999168395996096, 0.039241729736328126, 0.038391807556152346, 0.03807129669189453, 0.03802521514892578, 0.03807027053833008, 0.03834265518188477, 0.038401023864746094, 0.038095870971679685, 0.038079486846923825, 0.03811942291259766, 0.03794537734985352, 0.038862815856933595, 0.03844403076171875, 0.03829452896118164, 0.038161407470703124, 0.03806003189086914, 0.03870624160766602, 0.03798009490966797, 0.03802316665649414, 0.03815935897827148, 0.03817062377929688, 0.037872638702392575, 0.038145023345947264, 0.0380313606262207, 0.03792486572265625, 0.03787366485595703, 0.037891071319580076, 0.09221222686767579, 0.03798633575439453, 0.038026206970214846, 0.038063102722167966, 0.037967872619628903, 0.037787647247314454, 0.03809075164794922, 0.0380682258605957, 0.03805081558227539, 0.03799859237670898, 0.038206527709960934, 0.03777734375, 0.038095870971679685, 0.037969921112060545, 0.0383631362915039, 0.038040576934814455, 0.038042625427246096, 0.03799552154541016, 0.03804774475097656, 0.037917697906494144, 0.03784703826904297, 0.037919807434082034, 0.0380302734375, 0.03802009582519531, 0.038013950347900394, 0.03801497650146484, 0.03809382247924804, 0.03780915069580078, 0.03797196960449219, 0.03794432067871094, 0.03792588806152344, 0.03803852844238281, 0.038056961059570314, 0.03794124984741211, 0.037852161407470705, 0.03799552154541016, 0.03777951812744141, 0.038080448150634764, 0.03819007873535156, 0.03813683319091797, 0.03856793594360351, 0.03801804733276367, 0.038019073486328124, 0.03800678253173828, 0.038215679168701173, 0.0382371826171875, 0.03800064086914062, 0.038095870971679685, 0.03796377563476563, 0.037926910400390625, 0.037746688842773435, 0.03802521514892578, 0.03806719970703125, 0.03804569625854492, 0.03816960144042969, 0.03786444854736328, 0.03796889495849609, 0.03820544052124023, 0.038147071838378906, 0.037956607818603515, 0.037978111267089845, 0.03818092727661133, 0.037976001739501955, 0.09265766143798829, 0.03806719970703125, 0.03846963119506836, 0.037991424560546876, 0.03800883102416992, 0.03814918518066406, 0.03816543960571289, 0.038091777801513675, 0.03806617736816406, 0.03812966537475586, 0.03824435043334961, 0.03780915069580078, 0.03803750228881836, 0.037956607818603515, 0.03808153533935547, 0.03808768081665039, 0.038012928009033206, 0.037792766571044925, 0.037996543884277346, 0.03797708892822266, 0.038002689361572264, 0.03810201644897461, 0.03808768081665039, 0.038079486846923825, 0.03798527908325195, 0.03779072189331055, 0.03844812774658203, 0.038217727661132815, 0.038403072357177735, 0.03815321731567383, 0.03809689712524414, 0.03803750228881836, 0.038027263641357424, 0.03800883102416992, 0.037684223175048825, 0.037748737335205076, 0.037833728790283204, 0.038029312133789066, 0.03784703826904297, 0.03796480178833008, 0.03783475112915039, 0.037907455444335936, 0.03796275329589844, 0.0375838737487793, 0.03783987045288086, 0.03785625457763672, 0.037928958892822266, 0.0380497932434082, 0.038110240936279294, 0.03816243362426758, 0.03812041473388672, 0.038063102722167966, 0.03808051300048828, 0.038250495910644534, 0.0381102066040039, 0.03839385604858398, 0.038075393676757815, 0.03803443145751953, 0.03795558547973633, 0.03803955078125, 0.0379504623413086, 0.037823486328125, 0.037838848114013675, 0.09228697967529297, 0.03798227310180664, 0.038029247283935544, 0.03792486572265625, 0.03779379272460937, 0.0380497932434082, 0.038042625427246096, 0.03804774475097656, 0.03796889495849609, 0.03798323059082031, 0.037907455444335936, 0.03797401428222656, 0.03784806442260742, 0.03794432067871094, 0.03806208038330078, 0.0379617919921875, 0.03791763305664062, 0.03819724655151367, 0.03810815811157227, 0.038059009552001956, 0.037966846466064456, 0.0379504623413086, 0.0380313606262207, 0.037997566223144534, 0.0376627197265625, 0.03794124984741211, 0.03798220825195313, 0.03788083267211914, 0.037967872619628903, 0.03808256149291992, 0.03803955078125, 0.037948448181152346, 0.038009822845458986, 0.03793203353881836, 0.038046718597412106, 0.03794841766357422, 0.038004737854003906, 0.037981182098388674, 0.038145023345947264, 0.03796889495849609, 0.03782963180541992, 0.0379607048034668, 0.03803033447265625, 0.03834579086303711, 0.03797087860107422, 0.03778252792358398, 0.03802009582519531, 0.03799244689941406, 0.038024192810058595, 0.037969921112060545, 0.03806105422973633, 0.03794636917114258, 0.03798323059082031, 0.037978111267089845, 0.03779379272460937, 0.03774771118164062, 0.037795841217041014, 0.03800271987915039, 0.038001632690429686, 0.037716991424560545, 0.03775795364379883, 0.037975040435791016, 0.03803340911865234]",tokens/s,25.795509416957938,,,main,False,False,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,internlm/internlm2-20b,internlm/internlm2-20b,cuda,0,42,,,True,,,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",internlm2,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,i,i,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/i/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669474d4-2c191d8d1c3d1dae66c79b4c;505bb106-6a63-453a-8b3b-0d0f5037a99c) Repository Not Found for url: https://huggingface.co/i/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: i is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Deci/DeciLM-7B,Deci/DeciLM-7B,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1271.107584,14990.966784,0.0,14344.51968,14290.684928,s,10,0.8109073486328124,0.08109073486328125,0.00033108307087312787,0.08096873474121094,0.08144343719482422,0.08171106414794922,0.08192516571044922,"[0.08197869110107422, 0.08092742156982421, 0.08138396453857422, 0.08095558166503906, 0.080795166015625, 0.08105903625488281, 0.08102355194091797, 0.08089532470703124, 0.08098188781738282, 0.08090672302246094]",tokens/s,3156.957455517147,kWh,9.597559502902424e-07,5.258984752242302e-07,6.139534481516101e-06,7.625188907030573e-06,tokens/kWh,33572938.73256871,MB,1271.107584,14990.966784,0.0,14344.51968,14290.687488,s,10,22.71733959960937,2.2717339599609376,0.002440544220494791,2.2711645507812497,2.2729692626953124,2.2757928588867187,2.2780517358398438,"[2.278616455078125, 2.2720029296875, 2.271243896484375, 2.272341796875, 2.271085205078125, 2.27039990234375, 2.270092041015625, 2.271480712890625, 2.269405517578125, 2.270671142578125]",tokens/s,27.732120534520373,kWh,2.676503750372104e-05,1.4668148241254391e-05,0.00011490731235588376,0.00015634049810085915,tokens/kWh,402966.60663929273,,s,629,23.409343505859386,0.037216762330460056,0.009680149138883303,0.03602534484863281,0.036102809143066406,0.03627806625366211,0.11743506317138673,"[0.03676883316040039, 0.03676972961425781, 0.03669913482666016, 0.03677494430541992, 0.03670524978637695, 0.03685887908935547, 0.03666534423828125, 0.036743167877197266, 0.036743167877197266, 0.03661619186401367, 0.03632844924926758, 0.036364288330078126, 0.03649740982055664, 0.036176895141601564, 0.03603968048095703, 0.03602841567993164, 0.03602022552490235, 0.036013057708740234, 0.036036609649658206, 0.03599257659912109, 0.03599155044555664, 0.03606630325317383, 0.03599359893798828, 0.035953662872314454, 0.03600691223144531, 0.03602022552490235, 0.03597721481323242, 0.03598745727539063, 0.0360079345703125, 0.03599359893798828, 0.03599155044555664, 0.035975166320800785, 0.03598643112182617, 0.036034561157226565, 0.03597619247436523, 0.035996673583984375, 0.03600691223144531, 0.03599257659912109, 0.035996673583984375, 0.036055038452148434, 0.03599871826171875, 0.036021247863769534, 0.03600998306274414, 0.0360079345703125, 0.03602022552490235, 0.03600384140014649, 0.03604787063598633, 0.03607551956176758, 0.036040702819824216, 0.036020286560058595, 0.035992511749267576, 0.03600281524658203, 0.03604991912841797, 0.036013057708740234, 0.03601001739501953, 0.03605295944213867, 0.03600076675415039, 0.036034561157226565, 0.03603353500366211, 0.03600998306274414, 0.03602534484863281, 0.036070400238037106, 0.1174476776123047, 0.03598950576782227, 0.03600896072387695, 0.035983360290527344, 0.0359741439819336, 0.036001792907714845, 0.03601203155517578, 0.03602739334106445, 0.036001792907714845, 0.03602431869506836, 0.03618304061889648, 0.03604585647583008, 0.0360212173461914, 0.03599564743041992, 0.03690188980102539, 0.03660902404785156, 0.0360263671875, 0.036001792907714845, 0.03600896072387695, 0.03599871826171875, 0.03596799850463867, 0.03597312164306641, 0.03599769592285156, 0.036013057708740234, 0.0360079345703125, 0.03601715087890625, 0.03602329635620117, 0.03600998306274414, 0.035983360290527344, 0.03599257659912109, 0.03600998306274414, 0.036016128540039063, 0.03600076675415039, 0.03600896072387695, 0.03598438262939453, 0.036055038452148434, 0.03600384140014649, 0.03605299377441406, 0.036078590393066406, 0.0360079345703125, 0.03603046417236328, 0.0360263671875, 0.036037631988525394, 0.03601408004760742, 0.035985408782958986, 0.03601919937133789, 0.03604991912841797, 0.036036609649658206, 0.03609292984008789, 0.03602534484863281, 0.036018177032470705, 0.03605196762084961, 0.0360263671875, 0.03603251266479492, 0.03606016159057617, 0.03603353500366211, 0.036073471069335936, 0.03604582214355469, 0.03604377746582031, 0.03602131271362305, 0.03607238388061523, 0.0361082878112793, 0.03652403259277344, 0.11806822204589844, 0.036004863739013675, 0.03597721481323242, 0.0359741439819336, 0.03607551956176758, 0.0359741439819336, 0.036004863739013675, 0.036004863739013675, 0.03599769592285156, 0.036004863739013675, 0.03599462509155273, 0.03600076675415039, 0.03602431869506836, 0.03597312164306641, 0.03599257659912109, 0.03603353500366211, 0.03607759857177734, 0.0360294075012207, 0.036050945281982424, 0.03603968048095703, 0.035996673583984375, 0.03599769592285156, 0.03597721481323242, 0.036057086944580076, 0.0359813117980957, 0.036029441833496094, 0.03603046417236328, 0.03604787063598633, 0.036013057708740234, 0.03605606460571289, 0.036016128540039063, 0.03602227020263672, 0.03605913543701172, 0.036018177032470705, 0.03599769592285156, 0.03604377746582031, 0.03601408004760742, 0.0360447998046875, 0.03604278564453125, 0.03600585556030273, 0.035963905334472655, 0.03602022552490235, 0.03613491058349609, 0.036432895660400394, 0.036212734222412106, 0.03605913543701172, 0.03601510238647461, 0.03607654571533203, 0.036034561157226565, 0.03603558349609375, 0.03606630325317383, 0.03603046417236328, 0.036004863739013675, 0.03602739334106445, 0.03610214233398437, 0.03604991912841797, 0.036038654327392575, 0.03606630325317383, 0.03612160110473633, 0.03612876892089844, 0.03620454406738281, 0.03606937789916992, 0.03606835174560547, 0.11754700469970703, 0.03606016159057617, 0.03599257659912109, 0.035970046997070314, 0.03598950576782227, 0.03600896072387695, 0.036329471588134765, 0.03661929702758789, 0.03622396850585938, 0.03605196762084961, 0.03602739334106445, 0.036021247863769534, 0.03605196762084961, 0.03600691223144531, 0.036021247863769534, 0.03601408004760742, 0.03601715087890625, 0.036018177032470705, 0.0360263671875, 0.03603353500366211, 0.03611545562744141, 0.03606016159057617, 0.036050945281982424, 0.03602227020263672, 0.036029441833496094, 0.036168704986572264, 0.0360447998046875, 0.03604582214355469, 0.03607449722290039, 0.03604991912841797, 0.036018177032470705, 0.03608575820922852, 0.0361451530456543, 0.03604889678955078, 0.03606425476074219, 0.036037631988525394, 0.036029441833496094, 0.03602329635620117, 0.03602739334106445, 0.036034561157226565, 0.03600384140014649, 0.03603046417236328, 0.036071422576904294, 0.03604377746582031, 0.0360816650390625, 0.036057086944580076, 0.03603148651123047, 0.03604991912841797, 0.03614720153808594, 0.036101119995117184, 0.03607756805419922, 0.03608473587036133, 0.03602534484863281, 0.036040702819824216, 0.03603046417236328, 0.03602841567993164, 0.03601203155517578, 0.03602227020263672, 0.03602329635620117, 0.036040702819824216, 0.03605401611328125, 0.036038654327392575, 0.0360263671875, 0.11732685089111328, 0.03600998306274414, 0.0359854736328125, 0.03597817611694336, 0.035988479614257815, 0.03599155044555664, 0.036018177032470705, 0.036015167236328124, 0.03601913452148438, 0.03599155044555664, 0.03602841567993164, 0.03604787063598633, 0.03616563034057617, 0.03606425476074219, 0.03602534484863281, 0.03602329635620117, 0.03602534484863281, 0.03604889678955078, 0.03602022552490235, 0.03611545562744141, 0.03605302429199219, 0.03605807876586914, 0.03602739334106445, 0.03610009765625, 0.03602227020263672, 0.036057086944580076, 0.03602841567993164, 0.036094974517822266, 0.036116481781005856, 0.03604582214355469, 0.03605299377441406, 0.03605606460571289, 0.036055072784423825, 0.03603247833251953, 0.03604275131225586, 0.036078590393066406, 0.03613183975219727, 0.03604995346069336, 0.03600790405273437, 0.03604172897338867, 0.036102207183837894, 0.03601811218261719, 0.03603148651123047, 0.036018177032470705, 0.036013057708740234, 0.03604275131225586, 0.03599462509155273, 0.03599871826171875, 0.03602534484863281, 0.03603558349609375, 0.03609804916381836, 0.036036609649658206, 0.0360263671875, 0.03604172897338867, 0.036055038452148434, 0.036089855194091795, 0.03602742385864258, 0.03608163070678711, 0.03601203155517578, 0.036029441833496094, 0.03603558349609375, 0.036034561157226565, 0.03602227020263672, 0.11761254119873046, 0.03600288009643555, 0.03601606369018555, 0.03601107025146484, 0.03601401519775391, 0.036050945281982424, 0.036021247863769534, 0.03601100921630859, 0.03599359893798828, 0.036037631988525394, 0.03601919937133789, 0.03599564743041992, 0.036025375366210935, 0.036012001037597656, 0.03600588989257812, 0.036021247863769534, 0.03598950576782227, 0.036037631988525394, 0.036021247863769534, 0.03603353500366211, 0.036029441833496094, 0.03601510238647461, 0.03602841567993164, 0.036016128540039063, 0.03601100921630859, 0.03604684829711914, 0.03606425476074219, 0.035996673583984375, 0.03602329635620117, 0.03605404663085938, 0.03611542510986328, 0.03619942474365234, 0.03612876892089844, 0.0360263671875, 0.036079681396484375, 0.036023231506347654, 0.03601715087890625, 0.03603046417236328, 0.03607756805419922, 0.036029441833496094, 0.03602841567993164, 0.036021247863769534, 0.03603558349609375, 0.0360079345703125, 0.03600588989257812, 0.03600076675415039, 0.03604172897338867, 0.036018177032470705, 0.03601510238647461, 0.03600998306274414, 0.03602534484863281, 0.03601919937133789, 0.03599871826171875, 0.0360263671875, 0.03604684829711914, 0.03603353500366211, 0.03603251266479492, 0.036013057708740234, 0.036038654327392575, 0.036050945281982424, 0.03601203155517578, 0.03604582214355469, 0.03601203155517578, 0.1176258544921875, 0.03606630325317383, 0.03597824096679687, 0.03596492767333984, 0.036013057708740234, 0.03600998306274414, 0.03599052810668945, 0.035980289459228515, 0.0359813117980957, 0.03599052810668945, 0.03600076675415039, 0.03600896072387695, 0.03601408004760742, 0.03599564743041992, 0.036037631988525394, 0.035988479614257815, 0.036016128540039063, 0.03598950576782227, 0.035975166320800785, 0.03600691223144531, 0.0360079345703125, 0.036038654327392575, 0.035961856842041014, 0.03603148651123047, 0.03595980834960937, 0.035985408782958986, 0.03601510238647461, 0.03598950576782227, 0.03601715087890625, 0.03602739334106445, 0.03602329635620117, 0.0360263671875, 0.03606835174560547, 0.03601408004760742, 0.03599875259399414, 0.03599459075927734, 0.03601408004760742, 0.0360263671875, 0.03600691223144531, 0.0360263671875, 0.035983360290527344, 0.03601919937133789, 0.036073471069335936, 0.03604275131225586, 0.036063232421875, 0.036070400238037106, 0.036029441833496094, 0.03602841567993164, 0.03604275131225586, 0.03606528091430664, 0.03603353500366211, 0.03599782562255859, 0.036001663208007816, 0.036050945281982424, 0.03603251266479492, 0.036036609649658206, 0.03598233413696289, 0.035999744415283204, 0.036087806701660154, 0.03621887969970703, 0.03623628616333008, 0.0360447998046875, 0.03612979125976563, 0.11740262603759766, 0.03594342422485351, 0.035942401885986325, 0.03592499160766602, 0.03600896072387695, 0.03598950576782227, 0.03599257659912109, 0.03605811309814453, 0.03600998306274414, 0.03599462509155273, 0.03599052810668945, 0.03601100921630859, 0.03602022552490235, 0.03600588989257812, 0.03598438262939453, 0.03599980926513672, 0.03606419372558594, 0.03598745727539063, 0.03598950576782227, 0.03599052810668945, 0.03599564743041992, 0.03601408004760742, 0.03609702301025391, 0.036318206787109376, 0.03642060852050781, 0.03644825744628906, 0.03609600067138672, 0.03599769592285156, 0.03597926330566406, 0.036037631988525394, 0.03598745727539063, 0.03598233413696289, 0.03597107315063477, 0.036028446197509764, 0.03613385772705078, 0.03614310455322266, 0.03600998306274414, 0.03604377746582031, 0.03603148651123047, 0.036083713531494144, 0.036040702819824216, 0.03602329635620117, 0.036040702819824216, 0.036018177032470705, 0.03606937789916992, 0.0360816650390625, 0.03602534484863281, 0.03606835174560547, 0.036038654327392575, 0.03604787063598633, 0.036040702819824216, 0.03605606460571289, 0.03603046417236328, 0.03605401611328125, 0.03606534576416016, 0.036054977416992186, 0.036089855194091795, 0.03605097579956055, 0.036082656860351565, 0.03605606460571289, 0.03605913543701172, 0.03609088134765625, 0.036106273651123046, 0.11750089263916015, 0.03600185775756836, 0.03600479888916015, 0.03597721481323242, 0.035961856842041014, 0.035958816528320316, 0.036012001037597656, 0.0359813117980957, 0.03597721481323242, 0.03600896072387695, 0.03599359893798828, 0.03599052810668945, 0.03601203155517578, 0.0359741439819336, 0.0360263671875, 0.03600588989257812, 0.035985408782958986, 0.03610521697998047, 0.03600588989257812, 0.03603558349609375, 0.035985408782958986, 0.03600588989257812, 0.035975166320800785, 0.03600384140014649, 0.03598233413696289, 0.03602022552490235, 0.036013057708740234, 0.035985408782958986, 0.03604172897338867, 0.03608575820922852, 0.03604787063598633, 0.03601203155517578, 0.03602534484863281, 0.03599980926513672, 0.036058048248291015, 0.03606630325317383, 0.036034561157226565, 0.035996673583984375, 0.03600076675415039, 0.0360263671875, 0.035999744415283204, 0.03599052810668945, 0.03599564743041992, 0.03602739334106445, 0.03600998306274414, 0.036034561157226565, 0.03604377746582031, 0.03599257659912109, 0.036018177032470705, 0.036004863739013675, 0.036001792907714845, 0.03602534484863281, 0.03599257659912109, 0.036018177032470705, 0.03600691223144531, 0.035996673583984375, 0.03606528091430664, 0.03617792129516602, 0.03602534484863281, 0.036021247863769534, 0.03602841567993164, 0.036038654327392575, 0.03601919937133789, 0.117465087890625, 0.036001792907714845, 0.036013057708740234, 0.0360079345703125, 0.036001792907714845, 0.0360447998046875, 0.03605913543701172, 0.03606528091430664, 0.03603968048095703, 0.03600076675415039, 0.035980289459228515, 0.03599462509155273, 0.03600076675415039, 0.03598745727539063, 0.03600076675415039, 0.03602431869506836, 0.03600588989257812, 0.035980289459228515, 0.03599871826171875, 0.03602640151977539, 0.03605193710327149, 0.036004863739013675, 0.03599564743041992, 0.03599564743041992, 0.03605606460571289, 0.03619942474365234, 0.03606118392944336, 0.03599462509155273, 0.036004863739013675, 0.03604582214355469, 0.03599564743041992, 0.03601408004760742, 0.03599462509155273, 0.035992607116699216, 0.036042720794677734, 0.03609702301025391, 0.036051998138427736, 0.036004833221435543, 0.03600998306274414, 0.03600281524658203, 0.03599052810668945, 0.036008991241455075, 0.036061153411865235, 0.03609190368652344, 0.03604377746582031, 0.03599359893798828, 0.036149246215820316, 0.0363059196472168, 0.03622809600830078, 0.03607961654663086, 0.036089855194091795, 0.03602227020263672, 0.03602329635620117, 0.03601919937133789, 0.0360079345703125, 0.03602329635620117, 0.03603148651123047, 0.0360079345703125, 0.03603353500366211, 0.03603353500366211, 0.03602227020263672, 0.03602534484863281, 0.03601408004760742]",tokens/s,26.869612974945703,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,stabilityai/stablelm-3b-4e1t,stabilityai/stablelm-3b-4e1t,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1275.564032,6715.604992,0.0,6069.157888,5763.863552,s,10,0.30143148612976073,0.030143148612976078,0.0002693544417139532,0.030052528381347657,0.030259852027893067,0.030593701648712158,0.03086078134536743,"[0.03092755126953125, 0.029989919662475585, 0.030001663208007814, 0.029964672088623048, 0.0301856632232666, 0.03008675193786621, 0.03006937599182129, 0.030035680770874023, 0.030029600143432617, 0.030140607833862305]",tokens/s,8492.808872985373,kWh,3.571821008404056e-07,1.9571900876306465e-07,2.174576261311172e-06,2.7274773709146423e-06,tokens/kWh,93859623.81574298,MB,1275.564032,6736.576512,0.0,6090.129408,5863.102976,s,10,15.549432983398438,1.5549432983398437,0.01955409553439953,1.5539713745117187,1.5737054565429687,1.5871450866699217,1.5978967907714843,"[1.556048095703125, 1.5428585205078125, 1.559728759765625, 1.52816845703125, 1.5333924560546874, 1.5707188720703125, 1.5518946533203124, 1.560875, 1.5451634521484374, 1.600584716796875]",tokens/s,40.51594683051324,kWh,1.855575921592316e-05,1.0168665647714278e-05,5.9869667818489635e-05,8.859409268212708e-05,tokens/kWh,711108.3605319159,,s,629,15.793867799758907,0.025109487757963293,0.0036701201584507974,0.024382463455200197,0.025477402114868163,0.025684992599487304,0.05475516510009766,"[0.02547302436828613, 0.025034751892089844, 0.025161792755126953, 0.02518726348876953, 0.025349119186401366, 0.025007104873657225, 0.025176063537597656, 0.024830976486206056, 0.025829376220703124, 0.026022911071777344, 0.026002431869506838, 0.025477216720581054, 0.0248636474609375, 0.024754175186157225, 0.02475539207458496, 0.025433919906616212, 0.024584320068359374, 0.02454924774169922, 0.024483840942382814, 0.025398271560668945, 0.025673728942871094, 0.025447423934936524, 0.025438304901123046, 0.024879007339477538, 0.024748031616210937, 0.024761344909667967, 0.02440809631347656, 0.02394927978515625, 0.023871488571166992, 0.02389401626586914, 0.023967903137207033, 0.024062816619873046, 0.02387353515625, 0.0242739200592041, 0.024213504791259766, 0.024343551635742186, 0.02542598342895508, 0.02523027229309082, 0.02505523109436035, 0.025100288391113282, 0.024952831268310546, 0.024209407806396483, 0.02430156707763672, 0.024247295379638673, 0.024206335067749024, 0.02426572799682617, 0.02426470375061035, 0.024066144943237305, 0.024175519943237304, 0.02391347122192383, 0.023928831100463867, 0.024015871047973633, 0.024199167251586915, 0.024329216003417968, 0.024465408325195313, 0.024382463455200197, 0.02426470375061035, 0.024345600128173828, 0.02426572799682617, 0.02431488037109375, 0.02430771255493164, 0.024308736801147462, 0.054626304626464846, 0.024159231185913087, 0.023970815658569337, 0.02410393524169922, 0.02394316864013672, 0.023998464584350586, 0.02428927993774414, 0.02427903938293457, 0.024227840423583984, 0.024221696853637696, 0.024266752243041992, 0.024273984909057616, 0.024296480178833006, 0.024367008209228515, 0.025418752670288085, 0.025181184768676756, 0.024779775619506835, 0.024591360092163086, 0.02434662437438965, 0.02428927993774414, 0.024237056732177735, 0.024275968551635742, 0.024345600128173828, 0.024253440856933595, 0.024213504791259766, 0.024220672607421875, 0.02430463981628418, 0.02426470375061035, 0.024343551635742186, 0.024391679763793944, 0.026005504608154296, 0.025540607452392578, 0.0252938232421875, 0.025181184768676756, 0.025034751892089844, 0.024297567367553712, 0.02430454444885254, 0.02429747200012207, 0.024656896591186524, 0.024352767944335937, 0.0243056640625, 0.024224800109863283, 0.02415203285217285, 0.023983104705810547, 0.023992319107055664, 0.024212480545043946, 0.02432512092590332, 0.024270847320556642, 0.024263679504394533, 0.024276992797851563, 0.024429567337036134, 0.024361984252929687, 0.02444697570800781, 0.024378368377685547, 0.02448076820373535, 0.02609971237182617, 0.025423871994018556, 0.024905920028686523, 0.024601535797119142, 0.024997760772705078, 0.024549375534057616, 0.02434048080444336, 0.024413183212280275, 0.05488127899169922, 0.024435712814331056, 0.024194047927856444, 0.024231935501098634, 0.024253440856933595, 0.024321023941040038, 0.02430771255493164, 0.024359935760498046, 0.024382463455200197, 0.02429132843017578, 0.024226816177368164, 0.025118719100952147, 0.025198591232299804, 0.025309183120727538, 0.025219072341918947, 0.025176063537597656, 0.02525388717651367, 0.025199615478515625, 0.02430771255493164, 0.024389631271362306, 0.024368127822875976, 0.024197120666503907, 0.024263679504394533, 0.024244224548339844, 0.02531942367553711, 0.02532147216796875, 0.02527743911743164, 0.025174016952514647, 0.02515456008911133, 0.025222143173217772, 0.02612633514404297, 0.026534912109375, 0.025338880538940428, 0.025076736450195314, 0.024799232482910157, 0.024724479675292968, 0.024804351806640625, 0.025076736450195314, 0.024855552673339845, 0.025191423416137695, 0.02512179183959961, 0.024952831268310546, 0.02510233688354492, 0.02512179183959961, 0.025116672515869142, 0.024999967575073244, 0.02496099281311035, 0.025004032135009766, 0.02493132781982422, 0.02503987121582031, 0.024938495635986328, 0.024976383209228514, 0.023965696334838867, 0.02392268753051758, 0.023940095901489256, 0.02411724853515625, 0.02426470375061035, 0.02406809616088867, 0.02405068778991699, 0.024285184860229493, 0.02429849624633789, 0.024171520233154296, 0.024268800735473633, 0.054790145874023435, 0.024285184860229493, 0.024391679763793944, 0.024396896362304688, 0.02419804763793945, 0.02429132843017578, 0.024206335067749024, 0.0243507194519043, 0.024360960006713867, 0.024344575881958007, 0.024190975189208985, 0.024303615570068358, 0.024185855865478514, 0.024204288482666016, 0.024240127563476564, 0.024210432052612304, 0.024250368118286132, 0.02424831962585449, 0.024227840423583984, 0.024217599868774413, 0.024235008239746093, 0.024262655258178712, 0.024161279678344725, 0.024252416610717774, 0.024187904357910156, 0.02428313636779785, 0.024122367858886717, 0.02415001678466797, 0.024243200302124023, 0.023976959228515626, 0.024046592712402344, 0.02389299201965332, 0.02409071922302246, 0.02406595230102539, 0.024054784774780274, 0.024007680892944337, 0.024032255172729493, 0.023949312210083007, 0.024015871047973633, 0.023993343353271485, 0.023929983139038085, 0.024066944122314454, 0.023832576751708984, 0.02389206314086914, 0.023932832717895508, 0.025021568298339843, 0.02512985610961914, 0.023987199783325194, 0.024139776229858398, 0.024039487838745117, 0.023937023162841797, 0.02398918342590332, 0.023947328567504884, 0.02400351905822754, 0.02389504051208496, 0.024054784774780274, 0.02401900863647461, 0.02403424072265625, 0.025596927642822266, 0.02556723213195801, 0.025143295288085937, 0.025037824630737306, 0.02503167915344238, 0.05501235198974609, 0.02405887985229492, 0.02388787269592285, 0.023789567947387694, 0.024002559661865236, 0.02424019241333008, 0.023903167724609375, 0.02410188865661621, 0.02415001678466797, 0.024185888290405272, 0.02417558479309082, 0.02432512092590332, 0.024177663803100585, 0.02428211212158203, 0.02446233558654785, 0.02434867286682129, 0.02429952049255371, 0.024428543090820314, 0.024156255722045897, 0.024154016494750977, 0.024194047927856444, 0.024357887268066408, 0.024217599868774413, 0.024239103317260743, 0.024182783126831055, 0.024254463195800782, 0.02427801513671875, 0.02416640090942383, 0.024409088134765625, 0.024227840423583984, 0.024195072174072265, 0.024207359313964845, 0.024254463195800782, 0.02446950340270996, 0.024259584426879883, 0.02424115180969238, 0.024154111862182616, 0.02435481643676758, 0.024424448013305664, 0.024005632400512695, 0.024229888916015626, 0.024254463195800782, 0.02430463981628418, 0.024373247146606446, 0.023998464584350586, 0.024179712295532226, 0.02423200035095215, 0.024391616821289062, 0.024216575622558592, 0.024226816177368164, 0.024224767684936522, 0.024172544479370117, 0.02426470375061035, 0.024222719192504884, 0.024154111862182616, 0.02432307243347168, 0.024209407806396483, 0.024228864669799805, 0.02429644775390625, 0.02616524887084961, 0.026343423843383788, 0.025742336273193358, 0.02557439994812012, 0.05578035354614258, 0.025776128768920898, 0.02529484748840332, 0.025563135147094726, 0.025809919357299805, 0.02551296043395996, 0.025604095458984375, 0.025845760345458983, 0.025825279235839844, 0.02551603126525879, 0.02551705551147461, 0.025387008666992186, 0.025591808319091795, 0.025436159133911132, 0.025655296325683592, 0.025466880798339843, 0.025423871994018556, 0.025453567504882812, 0.025540607452392578, 0.025422847747802735, 0.024232959747314452, 0.024222719192504884, 0.02427289581298828, 0.024182783126831055, 0.02429952049255371, 0.024238079071044923, 0.025062400817871092, 0.02534809684753418, 0.025424896240234376, 0.025439231872558594, 0.02552217674255371, 0.02433945655822754, 0.02424934387207031, 0.02415718460083008, 0.024284160614013672, 0.02509823989868164, 0.025257984161376954, 0.025523199081420898, 0.02533478355407715, 0.025284671783447267, 0.02517190361022949, 0.02517094421386719, 0.025174144744873048, 0.025056127548217774, 0.02413670349121094, 0.023971935272216797, 0.024264736175537108, 0.02388979148864746, 0.023985151290893555, 0.02414080047607422, 0.023949312210083007, 0.0251474552154541, 0.025529279708862304, 0.025182207107543944, 0.024253440856933595, 0.024075424194335938, 0.02395939254760742, 0.024111167907714844, 0.023986207962036134, 0.02422262382507324, 0.02408038330078125, 0.024975360870361327, 0.025261056900024413, 0.05466521453857422, 0.023949312210083007, 0.02400774383544922, 0.023903167724609375, 0.02391244888305664, 0.02390937614440918, 0.023877664566040038, 0.02409881591796875, 0.02388476753234863, 0.023799808502197265, 0.02389504051208496, 0.023981056213378905, 0.023847936630249023, 0.02385408020019531, 0.024145919799804686, 0.024590335845947265, 0.025190528869628907, 0.025098175048828126, 0.025044927597045897, 0.025151519775390624, 0.02551190376281738, 0.024482816696166993, 0.025582719802856445, 0.024899456024169923, 0.02432614326477051, 0.02411724853515625, 0.02415718460083008, 0.02431590461730957, 0.024185855865478514, 0.024358911514282225, 0.024030208587646484, 0.0243507194519043, 0.024363008499145508, 0.02429030418395996, 0.02428108787536621, 0.024245248794555665, 0.025030656814575194, 0.025243648529052733, 0.025176063537597656, 0.025069568634033205, 0.025211904525756838, 0.02494156837463379, 0.024715263366699217, 0.02486783981323242, 0.02503167915344238, 0.024822784423828126, 0.02487500762939453, 0.02488012886047363, 0.02475632095336914, 0.02486467170715332, 0.025127935409545898, 0.02497331237792969, 0.025264127731323242, 0.025401344299316408, 0.02515456008911133, 0.02530303955078125, 0.025366527557373047, 0.024967168807983397, 0.024843263626098632, 0.024989696502685548, 0.02490985679626465, 0.024901599884033204, 0.025008127212524413, 0.05577011108398437, 0.024879104614257814, 0.024985599517822265, 0.02474496078491211, 0.025030656814575194, 0.025223167419433593, 0.02518134307861328, 0.025044832229614258, 0.024952831268310546, 0.02487398338317871, 0.024765439987182617, 0.024990720748901366, 0.024808448791503908, 0.024910848617553712, 0.025207807540893554, 0.024878080368041993, 0.02488422393798828, 0.02391551971435547, 0.023982080459594726, 0.02407731246948242, 0.023833599090576172, 0.024012800216674804, 0.02411212730407715, 0.024006656646728516, 0.02392268753051758, 0.02411724853515625, 0.024049663543701173, 0.02388275146484375, 0.024178688049316405, 0.024025087356567384, 0.023838720321655273, 0.024054784774780274, 0.024081407546997072, 0.023991296768188477, 0.023948287963867186, 0.024236032485961914, 0.02432614326477051, 0.024311807632446288, 0.024460287094116212, 0.025202688217163087, 0.025363487243652345, 0.025236448287963866, 0.02511769676208496, 0.02529587173461914, 0.02488934326171875, 0.024975360870361327, 0.025252864837646483, 0.02550169563293457, 0.02484223937988281, 0.02532659149169922, 0.02535424041748047, 0.02590835189819336, 0.025867136001586914, 0.025394176483154295, 0.025220096588134764, 0.0252938232421875, 0.02505628776550293, 0.02521801567077637, 0.02529996871948242, 0.025167871475219726, 0.025239551544189453, 0.02526310348510742, 0.025184255599975586, 0.055070720672607425, 0.024242176055908202, 0.0243558406829834, 0.024230911254882814, 0.024152063369750978, 0.024203264236450195, 0.02426576042175293, 0.02452783966064453, 0.02427801513671875, 0.024225791931152343, 0.024245248794555665, 0.024188928604125977, 0.024244224548339844, 0.024260608673095704, 0.0241080322265625, 0.024292352676391602, 0.0243240966796875, 0.02438860893249512, 0.023994367599487306, 0.024223775863647462, 0.024266719818115234, 0.024048639297485352, 0.024381439208984376, 0.024046592712402344, 0.02412031936645508, 0.024201215744018553, 0.024387584686279298, 0.02410905647277832, 0.02431385612487793, 0.024263679504394533, 0.024053760528564453, 0.024199199676513673, 0.024463327407836914, 0.0242739200592041, 0.0241408634185791, 0.024315839767456056, 0.024245248794555665, 0.023964672088623046, 0.024129535675048826, 0.024245248794555665, 0.024224767684936522, 0.024247295379638673, 0.024268800735473633, 0.024168447494506837, 0.024239103317260743, 0.02429439926147461, 0.024106048583984376, 0.024260608673095704, 0.02447148895263672, 0.024352767944335937, 0.024247360229492188, 0.024399808883666992, 0.02629631996154785, 0.026839040756225587, 0.02569011116027832, 0.025689088821411132, 0.025321535110473633, 0.02579859161376953, 0.025475072860717773, 0.02555904006958008, 0.026072063446044923, 0.025590784072875978, 0.02554572868347168, 0.055911422729492184, 0.0253439998626709, 0.025390079498291016, 0.025397247314453125, 0.02530303955078125, 0.025351167678833008, 0.025111551284790038, 0.02574745559692383, 0.02550169563293457, 0.025362432479858397, 0.025392127990722657, 0.025444351196289062, 0.025266176223754884, 0.02530713653564453, 0.02546073532104492, 0.02545462417602539, 0.02535215950012207, 0.02567884826660156, 0.025469024658203124, 0.02551696014404297, 0.025472000122070314, 0.025397247314453125, 0.02546892738342285, 0.025667583465576172, 0.025302175521850587, 0.025448287963867187, 0.025437183380126953, 0.025351167678833008, 0.025233407974243165, 0.025372671127319335, 0.02531123161315918, 0.02529484748840332, 0.025402368545532225, 0.02536038398742676, 0.0252620792388916, 0.025412607192993163, 0.025300991058349608, 0.025359359741210938, 0.025483264923095703, 0.025390079498291016, 0.025355264663696288, 0.025401344299316408, 0.025359359741210938, 0.025403392791748046, 0.02547711944580078, 0.02527027130126953, 0.025414655685424805, 0.025437183380126953, 0.02530816078186035, 0.0253439998626709, 0.025391103744506836, 0.025478143692016602, 0.025367551803588868, 0.02527129554748535, 0.025436159133911132, 0.02548428726196289, 0.0255098876953125, 0.025453567504882812, 0.025436159133911132, 0.025435136795043944, 0.025349119186401366, 0.025479167938232423, 0.025426975250244142]",tokens/s,39.82558344635514,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,huggyllama/llama-13b,huggyllama/llama-13b,cuda,0,42,,,True,,,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",llama,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-2.7b,,cuda,0,42,,,,,True,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,d35829e539df8480b726c647eeabf91e41eae047,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,,MB,1166.340096,6415.712256,0.0,5769.265152,5440.254976,s,11,0.28550201797485353,0.025954728906804867,0.00026299352270797233,0.025875776290893555,0.0259453125,0.026361760139465333,0.0266949182510376,"[0.026778207778930665, 0.025898784637451173, 0.025914527893066405, 0.0259453125, 0.025887231826782226, 0.025857696533203123, 0.025820928573608397, 0.025857791900634766, 0.025851200103759766, 0.025875776290893555, 0.025814559936523436]",tokens/s,9863.327832057663,kWh,3.0742872716028456e-07,1.6839201611998888e-07,1.679481141563561e-06,2.1553018848438345e-06,tokens/kWh,118776864.53122962,MB,1166.340096,6415.712256,0.0,5769.265152,5540.548096,s,11,10.171441650390625,0.924676513671875,0.0053055555117638336,0.9260877075195313,0.930017578125,0.9331093444824219,0.9355827575683593,"[0.9362011108398437, 0.9210723266601563, 0.926318359375, 0.927143798828125, 0.930017578125, 0.9260877075195313, 0.9211717529296874, 0.91734423828125, 0.926868408203125, 0.9185348510742187, 0.9206815185546875]",tokens/s,68.13193486425654,kWh,1.0813989020443238e-05,5.925677318238222e-06,4.533351101425638e-05,6.207317735293784e-05,tokens/kWh,1014931.1294601918,,s,692,10.415091710090643,0.015050710563714792,0.0031378477517591775,0.014643199920654297,0.014824345302581787,0.015105945253372189,0.040826952590942385,"[0.014900223731994629, 0.014862336158752442, 0.01490227222442627, 0.014836735725402832, 0.014833663940429688, 0.01490227222442627, 0.014883839607238769, 0.014950400352478027, 0.014932991981506348, 0.014847999572753906, 0.014823424339294433, 0.014787584304809571, 0.015187968254089355, 0.015266816139221191, 0.015326208114624023, 0.015528960227966309, 0.01539072036743164, 0.01539891242980957, 0.015129599571228027, 0.015158271789550782, 0.015278079986572265, 0.015148032188415527, 0.015149056434631348, 0.015060992240905761, 0.015357952117919921, 0.014962688446044922, 0.014846976280212403, 0.014803968429565429, 0.014806015968322754, 0.014807040214538575, 0.014744576454162597, 0.014632960319519044, 0.015027199745178223, 0.015145983695983887, 0.015498240470886231, 0.014850048065185547, 0.014635007858276367, 0.014708736419677734, 0.014680064201354981, 0.014631936073303223, 0.014621696472167968, 0.01457254409790039, 0.014551039695739745, 0.014618623733520507, 0.01458073616027832, 0.014569472312927247, 0.014562303543090821, 0.014559231758117675, 0.014622719764709472, 0.01457151985168457, 0.01457151985168457, 0.014611455917358398, 0.014625791549682618, 0.014614527702331542, 0.014590975761413574, 0.014576640129089356, 0.014592000007629394, 0.014583807945251465, 0.014629887580871583, 0.014622719764709472, 0.014643199920654297, 0.014671872138977051, 0.040888320922851565, 0.014537728309631348, 0.014690303802490234, 0.014558272361755371, 0.014569408416748047, 0.014561280250549317, 0.014539775848388671, 0.014561280250549317, 0.01457151985168457, 0.01458073616027832, 0.014540800094604492, 0.014551039695739745, 0.014507007598876954, 0.014554112434387208, 0.014547967910766601, 0.014517248153686524, 0.014550016403198243, 0.01455513572692871, 0.014791680335998534, 0.015131648063659669, 0.015449088096618652, 0.014748703956604005, 0.014609439849853516, 0.014649279594421387, 0.014590975761413574, 0.014630911827087402, 0.014601216316223145, 0.014707712173461914, 0.014592000007629394, 0.014607359886169433, 0.014558208465576173, 0.014550016403198243, 0.014561280250549317, 0.014554112434387208, 0.014582783699035644, 0.014564352035522461, 0.01457049560546875, 0.014578687667846679, 0.014560256004333496, 0.01459615993499756, 0.014581695556640625, 0.014593024253845215, 0.014576640129089356, 0.014606335639953612, 0.014574591636657714, 0.01460428810119629, 0.014595071792602539, 0.014583807945251465, 0.014583807945251465, 0.014583807945251465, 0.014558272361755371, 0.01460319995880127, 0.014601216316223145, 0.014578687667846679, 0.014607359886169433, 0.014606335639953612, 0.01458073616027832, 0.014600192070007324, 0.014592000007629394, 0.014608384132385254, 0.014614527702331542, 0.014652416229248047, 0.014749759674072266, 0.04090156936645508, 0.014516223907470703, 0.014724096298217774, 0.014608384132385254, 0.014599167823791503, 0.014614527702331542, 0.014585856437683106, 0.014562303543090821, 0.015447104454040527, 0.015940544128417968, 0.014824447631835937, 0.014671872138977051, 0.014697471618652343, 0.01466163158416748, 0.014668800354003907, 0.014650367736816406, 0.014630911827087402, 0.014650367736816406, 0.014651391983032227, 0.014655488014221191, 0.014625791549682618, 0.014694399833679199, 0.014729215621948242, 0.014600192070007324, 0.014723072052001953, 0.014627840042114258, 0.014642175674438476, 0.01464627170562744, 0.014631936073303223, 0.01457151985168457, 0.014697471618652343, 0.014612480163574219, 0.014657535552978516, 0.014708736419677734, 0.014684160232543946, 0.014691328048706055, 0.014665727615356445, 0.014663680076599121, 0.014628864288330079, 0.014689279556274413, 0.01476915168762207, 0.014648320198059082, 0.014745599746704101, 0.014682111740112304, 0.014676992416381835, 0.01464627170562744, 0.01467903995513916, 0.014652416229248047, 0.014773247718811035, 0.014676992416381835, 0.014626815795898437, 0.014714879989624024, 0.014668800354003907, 0.01470464038848877, 0.01466476821899414, 0.014650303840637207, 0.01465446376800537, 0.014700544357299805, 0.014665727615356445, 0.014709759712219238, 0.014723072052001953, 0.014667776107788086, 0.014692352294921876, 0.041388031005859374, 0.014621696472167968, 0.014672896385192872, 0.014637056350708008, 0.014639103889465332, 0.014624768257141114, 0.014812224388122559, 0.01459500789642334, 0.014625791549682618, 0.014651391983032227, 0.015435775756835938, 0.014898176193237305, 0.014767104148864747, 0.014650367736816406, 0.014630911827087402, 0.014647295951843262, 0.0145797119140625, 0.014600192070007324, 0.014694399833679199, 0.014652416229248047, 0.014638079643249511, 0.014656512260437012, 0.014693375587463378, 0.01458892822265625, 0.01469542407989502, 0.014708736419677734, 0.014660608291625977, 0.014686207771301269, 0.014620672225952149, 0.01464627170562744, 0.014625791549682618, 0.014665727615356445, 0.01460428810119629, 0.014642175674438476, 0.014643199920654297, 0.014638079643249511, 0.014707712173461914, 0.01467084789276123, 0.014629887580871583, 0.014650367736816406, 0.014657535552978516, 0.014624768257141114, 0.014657535552978516, 0.014699520111083985, 0.014660608291625977, 0.014740480422973632, 0.014827520370483398, 0.014905344009399414, 0.014788640022277833, 0.014729215621948242, 0.014631903648376465, 0.014745599746704101, 0.01475276756286621, 0.014628864288330079, 0.015086591720581055, 0.015075360298156738, 0.014832608222961426, 0.014714879989624024, 0.014716927528381347, 0.014675968170166016, 0.01469644832611084, 0.014667776107788086, 0.014712832450866698, 0.04105215835571289, 0.01457049560546875, 0.014645248413085938, 0.014645312309265137, 0.01468716812133789, 0.014705663681030273, 0.014690303802490234, 0.014628864288330079, 0.014729215621948242, 0.014813183784484863, 0.015070207595825195, 0.014800895690917968, 0.014771200180053711, 0.014779392242431641, 0.01466163158416748, 0.014651391983032227, 0.01468723201751709, 0.014628864288330079, 0.01476915168762207, 0.014689279556274413, 0.014657535552978516, 0.014686207771301269, 0.014648320198059082, 0.014709759712219238, 0.014701567649841308, 0.014791744232177734, 0.014702560424804688, 0.014689248085021972, 0.014700544357299805, 0.014733311653137206, 0.014887935638427734, 0.014716927528381347, 0.014788607597351074, 0.014703680038452148, 0.01471788787841797, 0.014744576454162597, 0.014791680335998534, 0.014742527961730957, 0.014676992416381835, 0.014622719764709472, 0.014737407684326171, 0.014707712173461914, 0.014718976020812988, 0.014657535552978516, 0.01467801570892334, 0.014665727615356445, 0.014692352294921876, 0.014658559799194336, 0.014706720352172851, 0.014643168449401856, 0.014896127700805664, 0.015568896293640137, 0.015171584129333495, 0.014748671531677245, 0.014798848152160645, 0.01469542407989502, 0.014780415534973144, 0.014745599746704101, 0.014753791809082031, 0.014833663940429688, 0.014781439781188965, 0.015465472221374512, 0.014741503715515136, 0.04084838485717773, 0.014510080337524415, 0.014665727615356445, 0.014560256004333496, 0.014532608032226562, 0.014741503715515136, 0.014856191635131836, 0.01480294418334961, 0.014700544357299805, 0.014723072052001953, 0.014665792465209961, 0.01476806354522705, 0.014734335899353027, 0.014643199920654297, 0.014758912086486817, 0.01469644832611084, 0.014560256004333496, 0.014567423820495605, 0.01458892822265625, 0.01470361614227295, 0.014637056350708008, 0.01467801570892334, 0.014635007858276367, 0.014629887580871583, 0.014723072052001953, 0.014690303802490234, 0.014650367736816406, 0.01476095962524414, 0.014663680076599121, 0.01466982364654541, 0.01469644832611084, 0.01477017593383789, 0.014643199920654297, 0.014864383697509765, 0.014653440475463866, 0.014653440475463866, 0.014671872138977051, 0.014651391983032227, 0.014660608291625977, 0.014640128135681153, 0.014693375587463378, 0.014686207771301269, 0.01477734375, 0.014742527961730957, 0.014842880249023438, 0.014721023559570312, 0.014686207771301269, 0.014664704322814942, 0.01472000026702881, 0.01467903995513916, 0.014663680076599121, 0.014723072052001953, 0.014655488014221191, 0.01487564754486084, 0.014820351600646972, 0.014729215621948242, 0.014717951774597168, 0.01468825626373291, 0.014699520111083985, 0.01469542407989502, 0.014726143836975097, 0.014784511566162109, 0.01470464038848877, 0.04082483291625977, 0.01448960018157959, 0.014536704063415527, 0.014513152122497559, 0.01450598430633545, 0.014540800094604492, 0.014511103630065919, 0.01451417636871338, 0.014517248153686524, 0.014550016403198243, 0.014682111740112304, 0.014616576194763184, 0.01451417636871338, 0.014554112434387208, 0.014520319938659668, 0.014530559539794922, 0.014537728309631348, 0.014525440216064453, 0.014702591896057129, 0.014636032104492188, 0.014639103889465332, 0.014633983612060546, 0.014618623733520507, 0.014627840042114258, 0.014583807945251465, 0.014663680076599121, 0.014614527702331542, 0.014584832191467285, 0.014620672225952149, 0.014686207771301269, 0.014648320198059082, 0.014617600440979005, 0.014838784217834473, 0.01466163158416748, 0.014652416229248047, 0.014727168083190918, 0.014635007858276367, 0.014750720024108887, 0.014592000007629394, 0.014628864288330079, 0.014647295951843262, 0.014656512260437012, 0.014650367736816406, 0.01467903995513916, 0.014660608291625977, 0.014637056350708008, 0.014702591896057129, 0.014663680076599121, 0.014614527702331542, 0.014638079643249511, 0.014610431671142577, 0.014615551948547363, 0.014626815795898437, 0.014608384132385254, 0.014657567977905274, 0.014622688293457031, 0.014610431671142577, 0.014650431632995606, 0.014618559837341309, 0.01461350440979004, 0.01465446376800537, 0.014633983612060546, 0.014617600440979005, 0.040822784423828126, 0.014499839782714843, 0.014517248153686524, 0.014500864028930664, 0.014545920372009278, 0.014510080337524415, 0.014507007598876954, 0.01454694366455078, 0.014514240264892578, 0.014516160011291504, 0.014527487754821777, 0.014532608032226562, 0.014526464462280274, 0.014534655570983887, 0.014523424148559571, 0.01453052806854248, 0.014565376281738282, 0.014517248153686524, 0.01460223960876465, 0.014527487754821777, 0.014513152122497559, 0.014524415969848633, 0.014550016403198243, 0.014520319938659668, 0.014525440216064453, 0.014530559539794922, 0.014529536247253418, 0.014511103630065919, 0.014525440216064453, 0.014534655570983887, 0.014538847923278808, 0.014516127586364747, 0.01452236843109131, 0.01452236843109131, 0.0145797758102417, 0.014552000045776368, 0.014537728309631348, 0.0145797119140625, 0.014545920372009278, 0.014536704063415527, 0.01460428810119629, 0.014551039695739745, 0.014586879730224609, 0.014593024253845215, 0.014578687667846679, 0.014566399574279786, 0.014611455917358398, 0.014560256004333496, 0.01460531234741211, 0.01458892822265625, 0.014562303543090821, 0.014592000007629394, 0.014600192070007324, 0.01455513572692871, 0.014585856437683106, 0.014581760406494141, 0.014573568344116212, 0.014573568344116212, 0.01460223960876465, 0.014574591636657714, 0.01458790397644043, 0.014850048065185547, 0.014609408378601075, 0.04080230331420898, 0.014523391723632812, 0.014509056091308594, 0.014520319938659668, 0.01456332778930664, 0.014512127876281738, 0.014537728309631348, 0.014512127876281738, 0.014510080337524415, 0.014553088188171387, 0.01450598430633545, 0.014519295692443847, 0.014655488014221191, 0.014757887840270996, 0.014680064201354981, 0.014601216316223145, 0.014706687927246094, 0.014813183784484863, 0.014841856002807617, 0.014595071792602539, 0.01460428810119629, 0.014655488014221191, 0.014640128135681153, 0.014672896385192872, 0.01477017593383789, 0.0146463041305542, 0.014545887947082519, 0.0146626558303833, 0.014629887580871583, 0.014558208465576173, 0.01459712028503418, 0.01475276756286621, 0.014731264114379883, 0.014658559799194336, 0.014692352294921876, 0.01468825626373291, 0.01468825626373291, 0.015613951683044434, 0.014807040214538575, 0.014659584045410156, 0.014676992416381835, 0.0146627197265625, 0.01469331169128418, 0.01464627170562744, 0.014810111999511719, 0.014722047805786133, 0.014675968170166016, 0.014664704322814942, 0.014690303802490234, 0.014693375587463378, 0.01468723201751709, 0.014647295951843262, 0.014676992416381835, 0.014756863594055175, 0.015331328392028809, 0.01571225643157959, 0.01496985626220703, 0.014738431930541992, 0.01478553581237793, 0.01476915168762207, 0.014751744270324708, 0.0147640323638916, 0.014689279556274413, 0.0408616943359375, 0.014503935813903808, 0.014544896125793457, 0.014497856140136718, 0.014543840408325195, 0.014519264221191406, 0.014504960060119629, 0.014536704063415527, 0.014521344184875488, 0.01451417636871338, 0.014576704025268554, 0.014738368034362794, 0.01472000026702881, 0.014573568344116212, 0.014566399574279786, 0.014548992156982422, 0.01455513572692871, 0.014498847961425781, 0.014533599853515626, 0.014516223907470703, 0.014501888275146485, 0.014567423820495605, 0.014538816452026367, 0.014541760444641113, 0.014545920372009278, 0.014554112434387208, 0.014521344184875488, 0.014510080337524415, 0.014526464462280274, 0.014556159973144531, 0.014544896125793457, 0.014521344184875488, 0.014539775848388671, 0.014527487754821777, 0.014559231758117675, 0.014551039695739745, 0.01456332778930664, 0.014562303543090821, 0.014533632278442383, 0.014532608032226562, 0.014574591636657714, 0.01457254409790039, 0.014561280250549317, 0.01458790397644043, 0.014576640129089356, 0.014569472312927247, 0.01458796787261963, 0.014562239646911621, 0.014722047805786133, 0.014576640129089356, 0.014624768257141114, 0.014583807945251465, 0.014598143577575684, 0.01460428810119629, 0.014600192070007324, 0.014610431671142577, 0.01458073616027832, 0.014632960319519044, 0.014655488014221191, 0.014806015968322754, 0.01472214412689209, 0.014699423789978027, 0.01467084789276123, 0.04086272048950195, 0.014504960060119629, 0.01467903995513916, 0.014650367736816406, 0.01460223960876465, 0.01458073616027832, 0.01459609603881836, 0.01457254409790039, 0.01465446376800537, 0.014775296211242676, 0.014552063941955566, 0.014631936073303223, 0.014542847633361817, 0.014561280250549317, 0.01454694366455078, 0.01449779224395752, 0.014525440216064453, 0.014542847633361817, 0.01452236843109131, 0.014518303871154785, 0.014516192436218262, 0.014512127876281738, 0.014516223907470703, 0.014493696212768555, 0.014535679817199706, 0.014516223907470703, 0.014491647720336913, 0.014525440216064453, 0.014528512001037597, 0.014544896125793457, 0.014947327613830566, 0.014631936073303223, 0.014583807945251465, 0.01459712028503418, 0.01458892822265625, 0.01460428810119629, 0.014612480163574219, 0.01457151985168457, 0.014551039695739745, 0.014554112434387208, 0.014529536247253418, 0.014575615882873535, 0.014554112434387208, 0.014568448066711426, 0.014625791549682618, 0.014615551948547363, 0.014625887870788574, 0.01466870403289795, 0.014697471618652343, 0.014655488014221191, 0.014673919677734374, 0.014805055618286133, 0.014645183563232422, 0.014743552207946778, 0.014726143836975097, 0.014705663681030273, 0.014687295913696288, 0.014664640426635742, 0.01463811206817627, 0.014724063873291015, 0.014709759712219238, 0.014684160232543946, 0.01467903995513916]",tokens/s,66.44204576034194,,,main,False,False,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,microsoft/phi-1_5,microsoft/phi-1_5,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1258.684416,3855.089664,0.0,3208.64256,2982.448128,s,10,0.2073017292022705,0.02073017292022705,0.0002540584937076844,0.020653135299682615,0.02079423770904541,0.021139391231536863,0.02141551404953003,"[0.02148454475402832, 0.02071753692626953, 0.020625600814819334, 0.020592479705810546, 0.02067411231994629, 0.020590656280517577, 0.020671648025512697, 0.020638879776000978, 0.020646463394165038, 0.020659807205200196]",tokens/s,12349.149280381214,kWh,2.4616023249648235e-07,1.3485836867289434e-07,1.18494365628748e-06,1.5659622574568567e-06,tokens/kWh,163477758.6630647,MB,1258.684416,3855.089664,0.0,3208.64256,3022.612992,s,10,12.214529541015626,1.2214529541015626,0.01696061576248178,1.2289693603515626,1.2328647705078124,1.2332307250976562,1.2335234887695312,"[1.229090087890625, 1.2298055419921874, 1.1763951416015626, 1.2245936279296874, 1.2288486328125, 1.22539501953125, 1.2335966796875, 1.20430419921875, 1.2297171630859376, 1.232783447265625]",tokens/s,51.577917748243955,kWh,1.4198180897230352e-05,7.779600564446128e-06,3.7808837538713236e-05,5.978661900038973e-05,tokens/kWh,1053747.494896631,,s,629,12.3801599788666,0.01968228931457325,0.002501032530214098,0.019481599807739256,0.01963007926940918,0.01973166046142578,0.03972911087036133,"[0.019680255889892577, 0.019377151489257814, 0.019293184280395507, 0.01937919998168945, 0.0195020809173584, 0.01943449592590332, 0.01940275192260742, 0.019308544158935546, 0.01941094398498535, 0.019252223968505858, 0.019265567779541016, 0.019406816482543946, 0.01944780731201172, 0.019384319305419923, 0.019389440536499023, 0.019369983673095705, 0.019520511627197267, 0.01927987289428711, 0.01939558410644531, 0.019581951141357423, 0.019504159927368165, 0.01933820724487305, 0.019212287902832033, 0.02002022361755371, 0.019982336044311523, 0.020198400497436524, 0.020398080825805662, 0.019870719909667968, 0.019579904556274414, 0.019570688247680663, 0.019534847259521485, 0.01944576072692871, 0.01943449592590332, 0.019538944244384765, 0.01946009635925293, 0.019506175994873046, 0.01947750473022461, 0.01960140800476074, 0.019475456237792968, 0.01945088005065918, 0.019401727676391603, 0.01947648048400879, 0.019434560775756837, 0.019422143936157227, 0.01946112060546875, 0.019535903930664063, 0.019522527694702148, 0.0194969596862793, 0.01948467254638672, 0.019449920654296876, 0.019474367141723632, 0.019501056671142578, 0.019514368057250975, 0.019495935440063478, 0.01948057556152344, 0.01945497512817383, 0.019458047866821288, 0.01941913604736328, 0.01920204734802246, 0.01922969627380371, 0.01942118453979492, 0.01960550308227539, 0.04069683074951172, 0.019508224487304687, 0.01948467254638672, 0.019494911193847657, 0.01942937660217285, 0.019474431991577147, 0.0196495361328125, 0.019562496185302734, 0.019478527069091797, 0.019495935440063478, 0.01952463912963867, 0.019461088180541992, 0.019519487380981446, 0.019478527069091797, 0.019491840362548828, 0.019511295318603517, 0.019586048126220702, 0.01947750473022461, 0.019552255630493166, 0.01964134407043457, 0.0194703369140625, 0.019523584365844726, 0.019564544677734375, 0.019475456237792968, 0.019513343811035155, 0.019548160552978516, 0.01942732810974121, 0.0194969596862793, 0.01946214485168457, 0.019559423446655275, 0.019525632858276368, 0.019515392303466796, 0.019561471939086913, 0.01946316719055176, 0.019530752182006835, 0.019569664001464843, 0.019491840362548828, 0.01961676788330078, 0.019595264434814453, 0.01966182327270508, 0.01968332862854004, 0.0194703369140625, 0.01960140800476074, 0.019487743377685548, 0.019483648300170898, 0.019483648300170898, 0.019549184799194336, 0.019582975387573243, 0.019471359252929688, 0.0194969596862793, 0.019514368057250975, 0.019543039321899415, 0.019494911193847657, 0.01941606330871582, 0.01943552017211914, 0.019406848907470704, 0.01943244743347168, 0.0195020809173584, 0.019490816116333007, 0.01946214485168457, 0.01943244743347168, 0.019514368057250975, 0.019499008178710937, 0.03973427200317383, 0.018657312393188477, 0.018718687057495118, 0.01866854476928711, 0.018715648651123046, 0.018619392395019533, 0.01863270378112793, 0.01861529541015625, 0.018637823104858398, 0.018637823104858398, 0.018733055114746093, 0.01857535934448242, 0.01864806365966797, 0.018752511978149415, 0.01863884735107422, 0.01863270378112793, 0.018677759170532226, 0.01862246322631836, 0.018715648651123046, 0.018694143295288086, 0.01862451171875, 0.01863680076599121, 0.018731008529663085, 0.018647039413452148, 0.01865318489074707, 0.018711551666259766, 0.018676736831665038, 0.018637823104858398, 0.018708480834960937, 0.01864499282836914, 0.018520063400268554, 0.018488319396972656, 0.018520063400268554, 0.018480127334594726, 0.018766847610473633, 0.018603008270263673, 0.01866547203063965, 0.018722816467285155, 0.018581504821777343, 0.018524160385131837, 0.018554880142211915, 0.018461696624755858, 0.018766847610473633, 0.018669567108154296, 0.018659328460693358, 0.018720767974853517, 0.018716672897338867, 0.01866035270690918, 0.01866649627685547, 0.01862758445739746, 0.01861324882507324, 0.01875660705566406, 0.018663423538208008, 0.01862348747253418, 0.018663423538208008, 0.018685951232910156, 0.01860095977783203, 0.018946048736572265, 0.01915292739868164, 0.01888252830505371, 0.018710527420043945, 0.018733055114746093, 0.01864806365966797, 0.039678974151611326, 0.018707456588745116, 0.01864908790588379, 0.01943142318725586, 0.019130367279052735, 0.01924300765991211, 0.01946112060546875, 0.019540992736816407, 0.01941094398498535, 0.019553279876708983, 0.019514368057250975, 0.01942835235595703, 0.01943142318725586, 0.019509248733520508, 0.019381248474121093, 0.019192832946777344, 0.019141632080078123, 0.019118080139160155, 0.019283967971801756, 0.0194703369140625, 0.01944268798828125, 0.01941913604736328, 0.01948057556152344, 0.019478527069091797, 0.019414016723632813, 0.019430400848388672, 0.019572736740112305, 0.01925939178466797, 0.019304447174072266, 0.019148799896240236, 0.019208192825317383, 0.019475456237792968, 0.019441728591918946, 0.019463104248046877, 0.0194652156829834, 0.0194467830657959, 0.019400703430175782, 0.01942425537109375, 0.01942835235595703, 0.019483648300170898, 0.019508224487304687, 0.019479551315307618, 0.019467327117919923, 0.01939142417907715, 0.0194150390625, 0.01947238349914551, 0.019495935440063478, 0.019523584365844726, 0.01945088005065918, 0.019491840362548828, 0.01943449592590332, 0.02085171127319336, 0.02035916709899902, 0.01965977668762207, 0.019555328369140625, 0.01961369514465332, 0.019534847259521485, 0.01945907211303711, 0.019510271072387696, 0.019481599807739256, 0.019510271072387696, 0.019568639755249022, 0.019515392303466796, 0.04063436889648438, 0.019519487380981446, 0.019539968490600586, 0.01944883155822754, 0.0194652156829834, 0.0194969596862793, 0.019487743377685548, 0.019447839736938477, 0.0195184326171875, 0.01946828842163086, 0.019474431991577147, 0.019529727935791014, 0.01948057556152344, 0.019509248733520508, 0.019644416809082032, 0.01922764778137207, 0.01920614433288574, 0.019190784454345702, 0.01923891258239746, 0.019494911193847657, 0.019561471939086913, 0.019542015075683594, 0.019488832473754884, 0.019561407089233398, 0.019491840362548828, 0.019466239929199217, 0.019398687362670898, 0.01951638412475586, 0.01944166374206543, 0.01946828842163086, 0.019843072891235353, 0.01948467254638672, 0.019546112060546874, 0.01945292854309082, 0.01944780731201172, 0.01945702362060547, 0.019478527069091797, 0.020287488937377928, 0.019804159164428712, 0.01947340774536133, 0.019679231643676756, 0.01957683181762695, 0.019515392303466796, 0.01947340774536133, 0.01943244743347168, 0.0194703369140625, 0.01963212776184082, 0.01942835235595703, 0.01944883155822754, 0.01944576072692871, 0.019482624053955077, 0.01942835235595703, 0.01943961524963379, 0.019535871505737306, 0.0194467830657959, 0.019471359252929688, 0.019553279876708983, 0.019572736740112305, 0.0194334716796875, 0.01945702362060547, 0.01942630386352539, 0.019500032424926757, 0.01946419143676758, 0.03971583938598633, 0.018735103607177735, 0.01884569549560547, 0.019610624313354492, 0.019438592910766602, 0.01946316719055176, 0.019491872787475585, 0.01947235107421875, 0.01943654441833496, 0.01943756866455078, 0.01963315200805664, 0.019412992477416992, 0.01943142318725586, 0.019546112060546874, 0.019516416549682617, 0.019261472702026366, 0.019381216049194336, 0.019542015075683594, 0.01943961524963379, 0.019471359252929688, 0.019430400848388672, 0.019451904296875, 0.019479551315307618, 0.019483648300170898, 0.019478527069091797, 0.019887104034423828, 0.019396608352661132, 0.01942118453979492, 0.019532800674438477, 0.019583999633789064, 0.019544063568115236, 0.01945702362060547, 0.01946828842163086, 0.019471359252929688, 0.019430400848388672, 0.01944063949584961, 0.019481599807739256, 0.019586048126220702, 0.01944371223449707, 0.01945702362060547, 0.01945395278930664, 0.019562496185302734, 0.019515392303466796, 0.019545087814331053, 0.019558399200439454, 0.019500032424926757, 0.019338239669799806, 0.019283967971801756, 0.019314687728881837, 0.019174400329589843, 0.019169279098510742, 0.019185663223266602, 0.019990528106689453, 0.019527679443359376, 0.019490816116333007, 0.019723264694213868, 0.019587072372436523, 0.019514368057250975, 0.019471359252929688, 0.019479551315307618, 0.019525632858276368, 0.01946112060546875, 0.01944780731201172, 0.04066918563842774, 0.019561471939086913, 0.01957683181762695, 0.019479551315307618, 0.01946828842163086, 0.019550207138061524, 0.019525632858276368, 0.019478527069091797, 0.019489791870117186, 0.019490816116333007, 0.019534847259521485, 0.019540992736816407, 0.019738624572753907, 0.019574783325195313, 0.019727359771728514, 0.019612672805786133, 0.019583999633789064, 0.01946419143676758, 0.019504192352294922, 0.019428287506103516, 0.020251680374145507, 0.01968841552734375, 0.01962393569946289, 0.01944166374206543, 0.019514368057250975, 0.019561471939086913, 0.019497983932495116, 0.019519487380981446, 0.0194467830657959, 0.01946009635925293, 0.019494911193847657, 0.019532800674438477, 0.01942732810974121, 0.019563520431518554, 0.0195020809173584, 0.01967411231994629, 0.02039910316467285, 0.01987379264831543, 0.019542015075683594, 0.01966592025756836, 0.01963929557800293, 0.019730432510375977, 0.01962393569946289, 0.019504127502441407, 0.019629056930541993, 0.01964646339416504, 0.019732479095458985, 0.019547136306762695, 0.019578880310058593, 0.019595264434814453, 0.019560480117797853, 0.019506143569946288, 0.019346431732177736, 0.019296255111694336, 0.019525632858276368, 0.01949388885498047, 0.019497983932495116, 0.01962393569946289, 0.019544063568115236, 0.019527679443359376, 0.01948057556152344, 0.019481599807739256, 0.01948876762390137, 0.039806015014648435, 0.01863264083862305, 0.018713600158691408, 0.018742271423339844, 0.018689023971557618, 0.01864908790588379, 0.018662399291992187, 0.01866649627685547, 0.01865011215209961, 0.01863065528869629, 0.01866035270690918, 0.018760704040527345, 0.01843609619140625, 0.01868185615539551, 0.018634752273559572, 0.018677759170532226, 0.018670591354370117, 0.018663423538208008, 0.01866649627685547, 0.018770944595336913, 0.018800640106201173, 0.018610176086425782, 0.01868083190917969, 0.018663423538208008, 0.01866854476928711, 0.018683904647827147, 0.01866854476928711, 0.018720767974853517, 0.018718719482421875, 0.018491392135620118, 0.01946931266784668, 0.019409919738769533, 0.01927577590942383, 0.0195020809173584, 0.01965260887145996, 0.01944473648071289, 0.019294208526611328, 0.019392511367797852, 0.019466239929199217, 0.0194334716796875, 0.019489791870117186, 0.019531776428222656, 0.019538944244384765, 0.01927475166320801, 0.019167232513427734, 0.019110912322998046, 0.019311616897583008, 0.019558399200439454, 0.01963007926940918, 0.019481599807739256, 0.01963007926940918, 0.019566591262817384, 0.019581951141357423, 0.019537919998168944, 0.019490816116333007, 0.0198287353515625, 0.020250688552856444, 0.019817407608032227, 0.01963315200805664, 0.019522560119628905, 0.019581951141357423, 0.019451904296875, 0.019491840362548828, 0.04060671997070313, 0.019556352615356445, 0.019581951141357423, 0.019500032424926757, 0.019533824920654298, 0.019497983932495116, 0.01942527961730957, 0.01949388885498047, 0.01947750473022461, 0.019535871505737306, 0.01945497512817383, 0.0194703369140625, 0.0194969596862793, 0.019523584365844726, 0.019516416549682617, 0.019536895751953123, 0.019559423446655275, 0.0198604793548584, 0.01957683181762695, 0.019481599807739256, 0.019471359252929688, 0.019449855804443358, 0.019260416030883788, 0.019533824920654298, 0.019539968490600586, 0.019536895751953123, 0.019511295318603517, 0.01920614433288574, 0.01924300765991211, 0.019397632598876953, 0.019490816116333007, 0.01946009635925293, 0.019579904556274414, 0.01944371223449707, 0.019204095840454103, 0.019336191177368164, 0.019525632858276368, 0.019511295318603517, 0.0193832950592041, 0.01943654441833496, 0.01963212776184082, 0.019539968490600586, 0.01963007926940918, 0.019532800674438477, 0.019557376861572266, 0.019526655197143555, 0.019577856063842772, 0.019663871765136717, 0.019583999633789064, 0.019523584365844726, 0.01956255912780762, 0.019493824005126954, 0.019532800674438477, 0.019571712493896484, 0.01965977668762207, 0.01965363121032715, 0.019586048126220702, 0.019564544677734375, 0.019574783325195313, 0.019542015075683594, 0.019566591262817384, 0.019531776428222656, 0.019569664001464843, 0.04064255905151367, 0.01961881637573242, 0.019597312927246095, 0.01944883155822754, 0.01957683181762695, 0.019695615768432616, 0.019554304122924804, 0.02061414337158203, 0.019511295318603517, 0.019538944244384765, 0.019534847259521485, 0.019500032424926757, 0.01945088005065918, 0.019512319564819337, 0.019548160552978516, 0.019504127502441407, 0.019550207138061524, 0.019568639755249022, 0.019589120864868165, 0.01959219169616699, 0.01944371223449707, 0.019560447692871095, 0.01964851188659668, 0.019511295318603517, 0.01945497512817383, 0.019556352615356445, 0.019552255630493166, 0.019534847259521485, 0.019524608612060547, 0.01978982353210449, 0.0196997127532959, 0.01958505630493164, 0.01968639945983887, 0.019530719757080078, 0.01949388885498047, 0.019491840362548828, 0.01948876762390137, 0.019548160552978516, 0.019602432250976562, 0.0195020809173584, 0.019544063568115236, 0.019554304122924804, 0.01946316719055176, 0.019578880310058593, 0.019537919998168944, 0.019620864868164063, 0.01945702362060547, 0.019583999633789064, 0.019549184799194336, 0.019521535873413084, 0.019524608612060547, 0.019722240447998047, 0.019475456237792968, 0.019568639755249022, 0.01947750473022461, 0.019559423446655275, 0.0194703369140625, 0.019495935440063478, 0.01944268798828125, 0.01946214485168457, 0.019526655197143555, 0.019537919998168944, 0.01944063949584961]",tokens/s,50.80709789483561,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,google/recurrentgemma-2b,google/recurrentgemma-2b,cuda,0,42,,,True,,,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",recurrent_gemma,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/gpt-j-6b,EleutherAI/gpt-j-6b,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1273.294848,13248.233472,0.0,12601.786368,12440.741376,s,10,0.8421385269165039,0.0842138526916504,0.0026801695709132464,0.08331206512451172,0.08456477661132812,0.08840313873291014,0.09147382843017578,"[0.09224150085449219, 0.08340889739990234, 0.08336707305908203, 0.08371180725097656, 0.08317648315429688, 0.08333411407470703, 0.08322550201416015, 0.0832900161743164, 0.08315699005126953, 0.08322614288330078]",tokens/s,3039.8799225745647,kWh,9.865773483007043e-07,5.404877841283451e-07,4.9880965830666434e-06,6.515161715495692e-06,tokens/kWh,39292961.737408355,MB,1273.618432,13248.233472,0.0,12601.786368,12548.118016,s,10,28.160445800781247,2.816044580078125,0.0033447196805785283,2.8155970458984374,2.820051318359375,2.8210337158203123,2.821819633789062,"[2.82201611328125, 2.812439697265625, 2.8151025390625, 2.816051513671875, 2.8119130859375, 2.812067626953125, 2.8198330078125, 2.815142578125, 2.8198232421875, 2.816056396484375]",tokens/s,22.371804923007364,kWh,3.328508268489882e-05,1.82398965548964e-05,0.00012315778371133358,0.0001746827629511288,tokens/kWh,360653.78710334224,,s,629,28.864495620727542,0.045889500191935675,0.009936046686600265,0.044663806915283204,0.04497141799926758,0.04524605484008789,0.12810968994140626,"[0.04559564971923828, 0.04529683303833008, 0.04530873489379883, 0.04604620742797851, 0.04544102478027344, 0.045608959197998046, 0.046292991638183595, 0.04553113555908203, 0.04501913452148437, 0.04538675308227539, 0.04499456024169922, 0.044655616760253904, 0.044355583190917966, 0.044243968963623044, 0.04440063858032227, 0.04428595352172852, 0.044232704162597655, 0.04431872177124024, 0.04426956939697266, 0.04436991882324219, 0.04437299346923828, 0.04426243209838867, 0.04443849563598633, 0.04434739303588867, 0.0444323844909668, 0.04453171157836914, 0.04494028854370117, 0.0445665283203125, 0.04465356826782227, 0.0447375373840332, 0.04475596618652344, 0.04464230346679687, 0.04458905410766602, 0.044663806915283204, 0.04465868759155273, 0.04472934341430664, 0.044665855407714845, 0.04461056137084961, 0.044657726287841794, 0.04470163345336914, 0.04469145584106445, 0.044657665252685545, 0.04463513565063477, 0.0445747184753418, 0.044629024505615233, 0.04470985412597656, 0.04475904083251953, 0.04481747055053711, 0.044692447662353516, 0.04495561599731445, 0.04472217559814453, 0.04477849578857422, 0.04469247817993164, 0.044677120208740234, 0.04472012710571289, 0.04475699234008789, 0.044846080780029295, 0.044731391906738284, 0.04488294219970703, 0.044715007781982424, 0.04481536102294922, 0.044777473449707034, 0.1284433898925781, 0.04455936050415039, 0.04444876861572265, 0.04452864074707031, 0.044375038146972655, 0.044461055755615236, 0.044477439880371096, 0.04475494384765625, 0.04443340682983398, 0.04458905410766602, 0.04485529708862305, 0.04463411331176758, 0.04443852615356445, 0.04450611114501953, 0.04447436904907227, 0.04453478240966797, 0.04437094497680664, 0.04451225662231445, 0.044437503814697264, 0.044543998718261715, 0.04466790390014649, 0.04469452667236328, 0.04455936050415039, 0.04450611114501953, 0.04453478240966797, 0.04456345748901367, 0.04453580856323242, 0.04457369613647461, 0.04459014511108399, 0.04491667175292969, 0.04501094436645508, 0.044595199584960936, 0.044616703033447266, 0.04466995239257812, 0.04470579147338867, 0.04459622573852539, 0.044644351959228515, 0.04461875152587891, 0.0445665283203125, 0.044668926239013675, 0.04463206481933594, 0.044698623657226565, 0.04460748672485351, 0.0446668815612793, 0.04479283142089844, 0.044677120208740234, 0.04469145584106445, 0.04454604721069336, 0.04459110260009766, 0.04465971374511719, 0.04467814254760742, 0.04470377731323242, 0.0449463996887207, 0.044763137817382816, 0.04466080093383789, 0.0449054069519043, 0.044780544281005856, 0.04473855972290039, 0.04478464126586914, 0.04483071899414062, 0.0447457275390625, 0.04478668975830078, 0.044772350311279296, 0.12841574096679687, 0.04443033599853516, 0.04443545532226562, 0.04441907119750976, 0.044365825653076174, 0.04442214584350586, 0.044418048858642575, 0.04505500793457031, 0.04459107208251953, 0.04444979095458984, 0.044880897521972656, 0.04460748672485351, 0.044472320556640625, 0.044477439880371096, 0.044523521423339846, 0.044507137298583986, 0.044459007263183595, 0.04453068923950195, 0.04446310424804688, 0.04453887939453125, 0.044543998718261715, 0.044609535217285154, 0.04471603012084961, 0.044556289672851565, 0.04499967956542969, 0.04465974426269531, 0.04458390426635742, 0.04460748672485351, 0.04458803176879883, 0.044614654541015625, 0.04452454376220703, 0.044611583709716796, 0.04496895980834961, 0.04478668975830078, 0.044637214660644534, 0.044848094940185546, 0.044746753692626956, 0.044660736083984375, 0.044662784576416016, 0.04495257568359375, 0.045227008819580077, 0.044799999237060545, 0.04462694549560547, 0.04481126403808594, 0.04466175842285156, 0.044763137817382816, 0.04471603012084961, 0.044727294921875, 0.044677120208740234, 0.04476006317138672, 0.04483174514770508, 0.04479897689819336, 0.04477337646484375, 0.04480614471435547, 0.045052928924560545, 0.044802047729492187, 0.044731391906738284, 0.04477030563354492, 0.044717056274414066, 0.04481024169921875, 0.04475904083251953, 0.04479180908203125, 0.044761089324951174, 0.12878746032714844, 0.044480510711669925, 0.04440576171875, 0.04437401580810547, 0.044426239013671875, 0.044447742462158206, 0.04436070251464844, 0.044336128234863284, 0.0444846076965332, 0.04449792098999023, 0.044488704681396485, 0.04442931365966797, 0.04469968032836914, 0.0445266227722168, 0.04491155242919922, 0.04485836791992188, 0.044542976379394535, 0.044631038665771484, 0.0444846076965332, 0.04449894332885742, 0.04453887939453125, 0.04450406265258789, 0.04447948837280274, 0.044523521423339846, 0.04467302322387695, 0.044614654541015625, 0.04459110260009766, 0.04457984161376953, 0.046296062469482424, 0.044967937469482425, 0.04461260986328125, 0.0446382064819336, 0.044622848510742184, 0.04499660873413086, 0.045059070587158204, 0.044878849029541014, 0.044614654541015625, 0.045011966705322266, 0.044781566619873044, 0.04469145584106445, 0.04466175842285156, 0.0446484489440918, 0.04462694549560547, 0.044706817626953124, 0.044744705200195314, 0.04468121719360352, 0.04469452667236328, 0.0446668815612793, 0.04487475204467774, 0.04480819320678711, 0.04474367904663086, 0.044767230987548826, 0.04470988845825195, 0.04472217559814453, 0.04470988845825195, 0.04476416015625, 0.04478668975830078, 0.04507340621948242, 0.04480716705322266, 0.044886016845703126, 0.04482559967041016, 0.04477337646484375, 0.04477644729614258, 0.12807168579101563, 0.0443873291015625, 0.04430847930908203, 0.04426649475097656, 0.04438016128540039, 0.04437299346923828, 0.04466995239257812, 0.04454707336425781, 0.04435353469848633, 0.04440063858032227, 0.044409854888916016, 0.0444139518737793, 0.04444979095458984, 0.04446003341674805, 0.04493926239013672, 0.044461055755615236, 0.04439859390258789, 0.04448767852783203, 0.044461055755615236, 0.044440574645996093, 0.04458291244506836, 0.044595199584960936, 0.04449587249755859, 0.04451123046875, 0.04445798492431641, 0.04453273773193359, 0.04453887939453125, 0.0446300163269043, 0.04457574462890625, 0.044614654541015625, 0.044543998718261715, 0.04458803176879883, 0.04458803176879883, 0.04465971374511719, 0.04454912185668945, 0.04465049743652344, 0.0448276481628418, 0.0446453742980957, 0.04460543823242188, 0.044695552825927735, 0.044750846862792966, 0.04494643020629883, 0.04467200088500976, 0.044608512878417966, 0.04456345748901367, 0.04480716705322266, 0.04470988845825195, 0.0447375373840332, 0.044665855407714845, 0.044693504333496094, 0.04472115325927734, 0.04502732849121094, 0.044902400970458986, 0.044825664520263674, 0.04478252792358398, 0.044700672149658206, 0.044834815979003906, 0.04486860656738281, 0.045143039703369144, 0.04518707275390625, 0.04477439880371094, 0.04481536102294922, 0.04481228637695313, 0.12827545166015625, 0.04438118362426758, 0.044462078094482424, 0.044424190521240234, 0.04442009735107422, 0.04459110260009766, 0.04449587249755859, 0.04456447982788086, 0.04456755065917969, 0.04446822357177734, 0.04445491027832031, 0.04443340682983398, 0.04448153686523437, 0.04448665618896484, 0.04453068923950195, 0.04442521667480469, 0.04445491027832031, 0.044843006134033206, 0.04452556610107422, 0.04450406265258789, 0.04556697463989258, 0.04503244781494141, 0.04482457733154297, 0.044539905548095705, 0.04455731201171875, 0.04455936050415039, 0.04448665618896484, 0.04460031890869141, 0.04452556610107422, 0.04462694549560547, 0.04447641754150391, 0.04460441589355469, 0.044624897003173826, 0.04460031890869141, 0.04468326568603516, 0.0445747184753418, 0.044641342163085934, 0.044732353210449216, 0.04462182235717774, 0.044984321594238284, 0.04460748672485351, 0.04461363220214844, 0.04460031890869141, 0.044614654541015625, 0.04455936050415039, 0.04468326568603516, 0.04467507171630859, 0.04479385757446289, 0.0447457275390625, 0.04477951812744141, 0.04456857681274414, 0.0446668815612793, 0.04462694549560547, 0.0447375373840332, 0.04467814254760742, 0.04460441589355469, 0.044616703033447266, 0.04468838500976562, 0.04472627258300781, 0.04472326278686523, 0.044656574249267576, 0.045134849548339843, 0.04467609786987305, 0.12843930053710936, 0.04417843246459961, 0.044336128234863284, 0.04473651123046875, 0.04531612777709961, 0.044659679412841796, 0.044393470764160156, 0.044437503814697264, 0.0445030403137207, 0.04441190338134766, 0.044440574645996093, 0.04455430221557617, 0.04452755355834961, 0.04442009735107422, 0.044447742462158206, 0.04445798492431641, 0.044439552307128906, 0.044563488006591795, 0.04444873428344726, 0.04489324951171875, 0.04460332870483399, 0.044662784576416016, 0.0446033935546875, 0.04463718414306641, 0.04453273773193359, 0.04456243133544922, 0.044595199584960936, 0.04490444946289063, 0.045314048767089846, 0.04468838500976562, 0.044590110778808596, 0.04464944076538086, 0.04475801467895508, 0.04492287826538086, 0.04480614471435547, 0.044665855407714845, 0.044631038665771484, 0.04484710311889648, 0.044728321075439455, 0.04463616180419922, 0.04459417724609375, 0.04521779251098633, 0.044744705200195314, 0.0447088623046875, 0.04518912124633789, 0.044799999237060545, 0.044717056274414066, 0.04482662582397461, 0.044864513397216796, 0.044744705200195314, 0.044819454193115234, 0.04501708984375, 0.04488806533813477, 0.04479283142089844, 0.0448276481628418, 0.045284351348876956, 0.045233150482177735, 0.04494956970214844, 0.04626015853881836, 0.04518809509277344, 0.0448614387512207, 0.04477542495727539, 0.04483689498901367, 0.1283542785644531, 0.04442009735107422, 0.04436991882324219, 0.04442828750610352, 0.04442521667480469, 0.04444569778442383, 0.04447948837280274, 0.044388351440429685, 0.0445747184753418, 0.04485126495361328, 0.04460947036743164, 0.04441292953491211, 0.04447641754150391, 0.044679168701171876, 0.04453171157836914, 0.04450918579101563, 0.04455936050415039, 0.04451327896118164, 0.044595199584960936, 0.044493824005126956, 0.04455116653442383, 0.044746753692626956, 0.04491987228393555, 0.04459206390380859, 0.04486963272094727, 0.044723201751708984, 0.044548095703125, 0.044576766967773435, 0.04464025497436523, 0.04460748672485351, 0.04463411331176758, 0.044655616760253904, 0.044862464904785154, 0.04468326568603516, 0.0446382064819336, 0.04497305679321289, 0.04468735885620117, 0.044625919342041014, 0.04463513565063477, 0.04470272064208984, 0.04467302322387695, 0.04466995239257812, 0.044734462738037106, 0.04464230346679687, 0.04494540786743164, 0.04466790390014649, 0.044611583709716796, 0.044639232635498044, 0.04476927947998047, 0.04468940734863281, 0.04484198379516602, 0.04470783996582031, 0.044685310363769534, 0.04463616180419922, 0.04497100830078125, 0.044947486877441406, 0.04478665542602539, 0.04482870483398437, 0.0453416633605957, 0.04503244781494141, 0.04513689422607422, 0.0447979507446289, 0.044750846862792966, 0.12807577514648438, 0.04472934341430664, 0.044442623138427735, 0.0444661750793457, 0.04465356826782227, 0.044490814208984375, 0.04449273681640625, 0.04439961624145508, 0.0445030403137207, 0.04451327896118164, 0.04455936050415039, 0.04456447982788086, 0.04484403228759765, 0.04458291244506836, 0.044508159637451174, 0.04476416015625, 0.044878849029541014, 0.04456345748901367, 0.044639232635498044, 0.044693504333496094, 0.04468428802490235, 0.04458803176879883, 0.04545024108886719, 0.045497344970703124, 0.04480819320678711, 0.044679168701171876, 0.04482457733154297, 0.04472524642944336, 0.044674049377441405, 0.04471091079711914, 0.0446607666015625, 0.04459926223754883, 0.04522905731201172, 0.04542566299438477, 0.044711936950683595, 0.04475596618652344, 0.044657665252685545, 0.044801025390625, 0.04480307388305664, 0.04471603012084961, 0.04465459060668946, 0.044731391906738284, 0.04469452667236328, 0.04483078384399414, 0.04469446563720703, 0.04530483245849609, 0.04476620864868164, 0.044865535736083983, 0.044679168701171876, 0.04487475204467774, 0.04493414306640625, 0.044717056274414066, 0.04486963272094727, 0.04476518249511719, 0.04465663909912109, 0.044902400970458986, 0.04499558258056641, 0.044777473449707034, 0.04475801467895508, 0.044717056274414066, 0.044816383361816405, 0.04481126403808594, 0.044821502685546875, 0.1281228790283203, 0.04504883193969727, 0.0444846076965332, 0.0447457275390625, 0.04445798492431641, 0.04440678405761719, 0.04440883255004883, 0.04439244842529297, 0.04448153686523437, 0.04443340682983398, 0.04441292953491211, 0.04463616180419922, 0.0445030403137207, 0.04451430511474609, 0.04451737594604492, 0.04449587249755859, 0.044467201232910154, 0.04455731201171875, 0.045254657745361325, 0.04457984161376953, 0.044461055755615236, 0.04451123046875, 0.044488704681396485, 0.04448153686523437, 0.04454604721069336, 0.044652542114257815, 0.04498227310180664, 0.04451532745361328, 0.04456755065917969, 0.04471398544311524, 0.04462080001831055, 0.04477337646484375, 0.04501299285888672, 0.04531097412109375, 0.04491059112548828, 0.04460441589355469, 0.04465459060668946, 0.04474367904663086, 0.04455936050415039, 0.044611583709716796, 0.04483379364013672, 0.044701694488525394, 0.04472115325927734, 0.04463513565063477, 0.0447825927734375, 0.044711936950683595, 0.04466175842285156, 0.044690433502197265, 0.0449986572265625, 0.044701694488525394, 0.04474879837036133, 0.044723201751708984, 0.04473344039916992, 0.04476006317138672, 0.044870655059814454, 0.04469964981079102, 0.04488499069213867, 0.045707263946533204, 0.04484096145629883, 0.04481228637695313, 0.044813312530517575, 0.0448092155456543, 0.04481228637695313]",tokens/s,21.79147726206296,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,stabilityai/stablelm-2-1_6b,stabilityai/stablelm-2-1_6b,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1262.53056,4276.617216,0.0,3632.267264,3408.132096,s,10,0.20172547531127932,0.020172547531127932,0.0006042842284314442,0.019867584228515624,0.020415465545654297,0.021160420989990235,0.021756385345458987,"[0.021905376434326173, 0.020249919891357424, 0.019857440948486328, 0.020231647491455076, 0.019795808792114258, 0.020247552871704103, 0.019860736846923827, 0.01987443161010742, 0.019846559524536133, 0.019856000900268556]",tokens/s,12690.514155684627,kWh,2.3836587425521627e-07,1.3061257829554786e-07,1.1820781186007993e-06,1.5510565711515636e-06,tokens/kWh,165048783.36573878,MB,1262.53056,4278.714368,0.0,3632.267264,3474.14784,s,10,11.84439318847656,1.1844393188476563,0.017149637998338907,1.1800570068359375,1.2016350952148436,1.2148355529785155,1.225395919189453,"[1.2280360107421875, 1.1738792724609375, 1.17937646484375, 1.1902847900390625, 1.1834696044921875, 1.19870166015625, 1.167730224609375, 1.1738626708984374, 1.180737548828125, 1.16831494140625]",tokens/s,53.189723608038314,kWh,1.4160408100587281e-05,7.759553078422177e-06,3.803503658839977e-05,5.995499776740923e-05,tokens/kWh,1050788.1301973127,,s,629,12.003487743377686,0.019083446332873906,0.002451252994608814,0.01860403251647949,0.019179769897460938,0.019804773712158205,0.038735421295166016,"[0.020542463302612304, 0.0214466552734375, 0.02125209617614746, 0.02061311912536621, 0.020566015243530272, 0.020291584014892578, 0.019958816528320312, 0.019805183410644533, 0.0193023681640625, 0.02062131118774414, 0.0198287353515625, 0.020178943634033202, 0.020082687377929686, 0.02007142448425293, 0.02020249557495117, 0.019955711364746095, 0.019833856582641602, 0.019764223098754884, 0.019976192474365235, 0.02011238479614258, 0.019804159164428712, 0.019768320083618163, 0.019677183151245118, 0.01968230438232422, 0.01901670455932617, 0.019979263305664064, 0.02011238479614258, 0.019560447692871095, 0.01923379135131836, 0.019111936569213867, 0.01924608039855957, 0.019086336135864256, 0.019124223709106446, 0.019074047088623047, 0.01901670455932617, 0.01906892776489258, 0.019103744506835937, 0.019129344940185547, 0.019164159774780275, 0.019133440017700197, 0.018712575912475587, 0.0190382080078125, 0.019080223083496092, 0.019180511474609373, 0.019127296447753905, 0.018946048736572265, 0.018535423278808593, 0.019070976257324217, 0.019120128631591796, 0.019128320693969726, 0.019109888076782225, 0.019177471160888672, 0.01882111930847168, 0.01920204734802246, 0.018685951232910156, 0.018507776260375978, 0.018527231216430663, 0.018521087646484375, 0.018503679275512695, 0.018556928634643553, 0.018671615600585938, 0.01925632095336914, 0.03919974517822265, 0.01861631965637207, 0.019177471160888672, 0.019099647521972657, 0.018858015060424806, 0.019159008026123046, 0.01917958450317383, 0.019191743850708008, 0.019079168319702147, 0.018308095932006836, 0.01856716728210449, 0.018927711486816406, 0.018954143524169922, 0.019115007400512696, 0.018861055374145508, 0.018431999206542968, 0.018561023712158203, 0.018520063400268554, 0.018456575393676757, 0.018655231475830078, 0.01902694320678711, 0.018495487213134765, 0.01842483139038086, 0.018486272811889647, 0.01847500801086426, 0.018513919830322266, 0.01843507194519043, 0.018495487213134765, 0.018380800247192384, 0.01845452880859375, 0.018482175827026368, 0.01843609619140625, 0.01846067237854004, 0.018529279708862305, 0.018449407577514648, 0.018412544250488282, 0.018395135879516602, 0.018659328460693358, 0.019100671768188478, 0.01879654312133789, 0.019056640625, 0.01904025650024414, 0.018675712585449217, 0.018489343643188477, 0.01820364761352539, 0.018440191268920898, 0.018295808792114256, 0.018214912414550782, 0.018335744857788085, 0.018545663833618165, 0.018498559951782227, 0.018548736572265623, 0.018561023712158203, 0.01850060844421387, 0.018526208877563476, 0.01884774398803711, 0.01880268859863281, 0.018512895584106445, 0.01844121551513672, 0.01847500801086426, 0.018453504562377928, 0.0184453125, 0.01846272087097168, 0.03910041427612305, 0.019066879272460938, 0.018532352447509767, 0.01863680076599121, 0.018544639587402344, 0.01841049575805664, 0.018513919830322266, 0.01845248031616211, 0.018541568756103514, 0.018669567108154296, 0.01903615951538086, 0.01903308868408203, 0.01908531188964844, 0.019106815338134766, 0.018551807403564453, 0.018509824752807616, 0.01843814468383789, 0.018505727767944336, 0.018486272811889647, 0.018520063400268554, 0.0184770565032959, 0.018503679275512695, 0.01846272087097168, 0.0186060791015625, 0.018523136138916017, 0.018553855895996094, 0.01862758445739746, 0.019130367279052735, 0.019154943466186524, 0.019125247955322267, 0.019168256759643554, 0.019204095840454103, 0.01881292724609375, 0.018530303955078126, 0.018546688079833985, 0.018530303955078126, 0.018527231216430663, 0.018471935272216796, 0.018535423278808593, 0.018525184631347655, 0.018541568756103514, 0.01859071922302246, 0.018577407836914063, 0.018478080749511717, 0.018478080749511717, 0.018824192047119142, 0.01904332733154297, 0.019099647521972657, 0.019079168319702147, 0.019099647521972657, 0.01864806365966797, 0.018506752014160157, 0.018541568756103514, 0.018465791702270508, 0.018586624145507814, 0.018512895584106445, 0.018491392135620118, 0.018508800506591795, 0.019112960815429687, 0.01902694320678711, 0.019100671768188478, 0.01906073570251465, 0.019109888076782225, 0.03967488098144531, 0.018980863571166993, 0.018959360122680666, 0.01905356788635254, 0.019014656066894533, 0.019109888076782225, 0.01842585563659668, 0.018521087646484375, 0.018512895584106445, 0.01840947151184082, 0.018479103088378905, 0.018529279708862305, 0.018541568756103514, 0.018449407577514648, 0.018489343643188477, 0.018473983764648438, 0.018508800506591795, 0.018973695755004884, 0.01903104019165039, 0.01884364891052246, 0.01903001594543457, 0.01906380844116211, 0.019094528198242186, 0.019133440017700197, 0.018968576431274413, 0.01906380844116211, 0.019105791091918945, 0.019078144073486326, 0.01905971145629883, 0.019108863830566408, 0.019158016204833983, 0.018573312759399413, 0.01858252716064453, 0.018556928634643553, 0.01900851249694824, 0.019284992218017577, 0.0185743350982666, 0.018363391876220703, 0.018496511459350586, 0.018543615341186523, 0.018938880920410156, 0.01904332733154297, 0.01881088066101074, 0.01902899169921875, 0.018911231994628908, 0.018923519134521484, 0.018784351348876953, 0.019183647155761718, 0.018815872192382812, 0.019014656066894533, 0.01904025650024414, 0.01876479911804199, 0.020634624481201173, 0.019451904296875, 0.019106815338134766, 0.019137535095214844, 0.018933792114257813, 0.018851808547973633, 0.018938880920410156, 0.018910207748413087, 0.018840576171875, 0.018852863311767578, 0.01882316780090332, 0.03844300842285156, 0.018851839065551757, 0.01883340835571289, 0.01881088066101074, 0.018976768493652343, 0.01887129592895508, 0.019036256790161132, 0.018843616485595703, 0.01892140769958496, 0.018868223190307617, 0.018856960296630858, 0.018538496017456055, 0.018280448913574218, 0.01825484848022461, 0.018266111373901366, 0.01823142433166504, 0.01822604751586914, 0.0182476806640625, 0.01845964813232422, 0.018541568756103514, 0.018534400939941405, 0.018465791702270508, 0.018488319396972656, 0.018486272811889647, 0.01839411163330078, 0.018455551147460936, 0.018525184631347655, 0.018516992568969725, 0.018450464248657227, 0.018618335723876955, 0.018513919830322266, 0.018513919830322266, 0.018619392395019533, 0.018453504562377928, 0.01861222457885742, 0.01840025520324707, 0.018787328720092773, 0.019181631088256837, 0.019106752395629884, 0.01908531188964844, 0.019148799896240236, 0.019058719635009765, 0.019175392150878906, 0.019174400329589843, 0.01904435157775879, 0.018679807662963867, 0.019136512756347656, 0.019075071334838867, 0.019108863830566408, 0.01903923225402832, 0.019118080139160155, 0.01923174476623535, 0.019125247955322267, 0.01907302474975586, 0.019110944747924803, 0.019160032272338867, 0.019133440017700197, 0.019054592132568358, 0.019722240447998047, 0.018531328201293946, 0.018663423538208008, 0.019140607833862306, 0.018979839324951172, 0.03991961669921875, 0.019216384887695313, 0.019150848388671874, 0.019146751403808594, 0.018997247695922852, 0.019105791091918945, 0.018881536483764647, 0.0184586238861084, 0.018907136917114258, 0.020163583755493163, 0.019698688507080078, 0.019131391525268555, 0.018559999465942383, 0.018554880142211915, 0.019185663223266602, 0.019090431213378906, 0.019098623275756836, 0.01905971145629883, 0.018977792739868164, 0.01920512008666992, 0.019120128631591796, 0.01923276710510254, 0.019084287643432618, 0.01907711982727051, 0.019142656326293944, 0.019140607833862306, 0.019058687210083008, 0.019076095581054688, 0.019161088943481445, 0.019091455459594727, 0.01904947280883789, 0.01905971145629883, 0.019100671768188478, 0.019079168319702147, 0.019120128631591796, 0.01906585693359375, 0.018800640106201173, 0.0191016960144043, 0.019168256759643554, 0.021114879608154297, 0.019520511627197267, 0.019134464263916014, 0.01889695930480957, 0.01913337516784668, 0.019062784194946288, 0.019168256759643554, 0.019506175994873046, 0.0189040641784668, 0.018456575393676757, 0.018568256378173827, 0.01847596740722656, 0.018541568756103514, 0.018523136138916017, 0.018531328201293946, 0.01897881507873535, 0.01904742431640625, 0.018505727767944336, 0.018555904388427736, 0.018508800506591795, 0.018526208877563476, 0.018465791702270508, 0.019215360641479492, 0.01844326400756836, 0.03929702377319336, 0.018728960037231446, 0.018301952362060548, 0.01826918411254883, 0.018327552795410155, 0.018184255599975586, 0.018796480178833008, 0.018395135879516602, 0.018572288513183592, 0.01848320007324219, 0.018505727767944336, 0.01849241638183594, 0.018431999206542968, 0.01860403251647949, 0.018431999206542968, 0.018523136138916017, 0.018487295150756835, 0.018504703521728515, 0.018507776260375978, 0.018421760559082033, 0.018283519744873047, 0.018692096710205077, 0.01823232078552246, 0.01826521682739258, 0.018296703338623047, 0.018242559432983398, 0.01857535934448242, 0.01842995262145996, 0.018525184631347655, 0.018471935272216796, 0.018486272811889647, 0.018513919830322266, 0.018495487213134765, 0.018522111892700196, 0.018548736572265623, 0.018480127334594726, 0.018588672637939452, 0.018471935272216796, 0.018526208877563476, 0.018544639587402344, 0.018559999465942383, 0.018527231216430663, 0.018585599899291993, 0.018502656936645507, 0.01850060844421387, 0.018550783157348632, 0.018537471771240235, 0.018516992568969725, 0.018516992568969725, 0.018473983764648438, 0.018694143295288086, 0.019397632598876953, 0.019135488510131835, 0.0184770565032959, 0.01862246322631836, 0.018573312759399413, 0.018544639587402344, 0.018519039154052733, 0.018516992568969725, 0.018408447265625, 0.018513919830322266, 0.018778112411499022, 0.018700288772583007, 0.03871846389770508, 0.018515968322753908, 0.01841971206665039, 0.01843404769897461, 0.018421760559082033, 0.018549760818481444, 0.018811904907226562, 0.018521087646484375, 0.018496511459350586, 0.018473983764648438, 0.018569215774536133, 0.018568191528320312, 0.0184268798828125, 0.01855897521972656, 0.019104768753051758, 0.01906073570251465, 0.01902387237548828, 0.019129344940185547, 0.018748416900634765, 0.01847500801086426, 0.018535423278808593, 0.018514944076538087, 0.018431999206542968, 0.0185743350982666, 0.01899622344970703, 0.01903104019165039, 0.019083263397216797, 0.01902079963684082, 0.018520063400268554, 0.0186060791015625, 0.01848422431945801, 0.018485248565673826, 0.018537471771240235, 0.018910207748413087, 0.018465791702270508, 0.018509855270385744, 0.018459615707397462, 0.01848320007324219, 0.018551807403564453, 0.018479103088378905, 0.018778112411499022, 0.018624576568603515, 0.018471872329711914, 0.018502656936645507, 0.018510847091674804, 0.018465791702270508, 0.018528255462646484, 0.018777120590209962, 0.01847702407836914, 0.018780223846435545, 0.019100608825683593, 0.019056640625, 0.018651136398315428, 0.01859993553161621, 0.018498559951782227, 0.018539520263671876, 0.018534400939941405, 0.018488319396972656, 0.018525184631347655, 0.018588672637939452, 0.018491392135620118, 0.018496511459350586, 0.018549760818481444, 0.038742015838623044, 0.018594816207885743, 0.018518016815185546, 0.018554880142211915, 0.018515968322753908, 0.018482175827026368, 0.018509824752807616, 0.018501632690429686, 0.018498559951782227, 0.018556928634643553, 0.018472959518432617, 0.018512895584106445, 0.01846784019470215, 0.018527231216430663, 0.018957311630249024, 0.018927616119384767, 0.01883443260192871, 0.018907136917114258, 0.01888876724243164, 0.018984895706176758, 0.019397632598876953, 0.018735103607177735, 0.019118112564086916, 0.019149856567382814, 0.018949056625366213, 0.01899622344970703, 0.018898944854736328, 0.018889728546142577, 0.01884876823425293, 0.018915327072143554, 0.018873344421386717, 0.01883852767944336, 0.01836031913757324, 0.01820364761352539, 0.018283519744873047, 0.018251775741577148, 0.018232416152954102, 0.018818080902099608, 0.018825088500976563, 0.01881395149230957, 0.01887129592895508, 0.01886617660522461, 0.0188570556640625, 0.01900124740600586, 0.019188735961914064, 0.019133440017700197, 0.018939903259277344, 0.018927616119384767, 0.018935808181762694, 0.01884160041809082, 0.01886310386657715, 0.018912479400634764, 0.01886297607421875, 0.018691999435424805, 0.018296831130981444, 0.01836031913757324, 0.018362464904785155, 0.018191263198852538, 0.018739200592041014, 0.018753536224365236, 0.018817024230957033, 0.019144704818725586, 0.018720767974853517, 0.0388853759765625, 0.018555904388427736, 0.018421823501586915, 0.018602943420410155, 0.018553855895996094, 0.018581504821777343, 0.01853753662109375, 0.01853331184387207, 0.018557952880859374, 0.018505727767944336, 0.018494464874267577, 0.018523136138916017, 0.018592767715454102, 0.018498559951782227, 0.018589696884155273, 0.018584575653076172, 0.018554912567138673, 0.01848214340209961, 0.018537471771240235, 0.018555904388427736, 0.018501632690429686, 0.01858371162414551, 0.01849839973449707, 0.018569215774536133, 0.018603008270263673, 0.0185948486328125, 0.018607072830200196, 0.01859891128540039, 0.018527231216430663, 0.018573312759399413, 0.018562047958374024, 0.018562047958374024, 0.018478080749511717, 0.018533376693725585, 0.018543615341186523, 0.018505727767944336, 0.018522111892700196, 0.018504703521728515, 0.018525184631347655, 0.018526208877563476, 0.018485279083251954, 0.018565088272094726, 0.01856716728210449, 0.01856716728210449, 0.018518016815185546, 0.018661376953125, 0.01846784019470215, 0.018555904388427736, 0.01861734390258789, 0.01846272087097168, 0.018493440628051756, 0.0185262393951416, 0.01848931121826172, 0.018480127334594726, 0.018537471771240235, 0.018504703521728515, 0.018534400939941405, 0.018581504821777343, 0.018494464874267577, 0.018588672637939452, 0.018421760559082033, 0.018534400939941405, 0.018485248565673826]",tokens/s,52.401436436423964,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,togethercomputer/RedPajama-INCITE-Base-3B-v1,togethercomputer/RedPajama-INCITE-Base-3B-v1,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1270.267904,6833.045504,0.0,6186.5984,6020.355072,s,10,0.3213976917266846,0.03213976917266846,0.00036806584513412007,0.03201321601867676,0.0321806095123291,0.0327108320236206,0.03313501003265381,"[0.03324105453491211, 0.03205043029785156, 0.03199142456054688, 0.032021217346191407, 0.031999135971069334, 0.031981311798095706, 0.03200521469116211, 0.03199555206298828, 0.032049568176269534, 0.03206278228759766]",tokens/s,7965.20966359962,kWh,3.804204537639366e-07,2.0845340897599228e-07,2.0093245799486035e-06,2.5981984426885324e-06,tokens/kWh,98529810.42321749,MB,1270.267904,6833.045504,0.0,6186.5984,6150.076416,s,10,14.384355590820311,1.4384355590820312,0.02662642684824122,1.4396050415039063,1.4704298217773437,1.4776764587402342,1.4834737683105468,"[1.484923095703125, 1.4688194580078124, 1.4525489501953126, 1.4121748046875, 1.4084876708984375, 1.4079842529296875, 1.4099942626953126, 1.44141064453125, 1.4602130126953126, 1.4377994384765624]",tokens/s,43.79758245145498,kWh,1.6617015781584444e-05,9.106055178637102e-06,5.852164051145294e-05,8.424471147167449e-05,tokens/kWh,747821.4228460195,,s,629,14.64821042633056,0.02328809288764796,0.0038780692554575094,0.022871999740600585,0.02342133674621582,0.024104550552368164,0.054898360290527345,"[0.024374271392822267, 0.024207359313964845, 0.024366079330444337, 0.024183807373046876, 0.024327167510986326, 0.024180736541748047, 0.024491008758544923, 0.02410700798034668, 0.024161279678344725, 0.02447871971130371, 0.024222719192504884, 0.023268384933471678, 0.024062944412231446, 0.02410086441040039, 0.024192031860351564, 0.024439775466918945, 0.024195135116577147, 0.02318022346496582, 0.024639488220214844, 0.023451648712158202, 0.024154111862182616, 0.02433228874206543, 0.024045568466186523, 0.023796735763549806, 0.023248895645141602, 0.02353561592102051, 0.023292959213256834, 0.02330006408691406, 0.023423999786376954, 0.023216127395629883, 0.02349772834777832, 0.023311359405517578, 0.023134208679199218, 0.02311065673828125, 0.023119871139526366, 0.0231844482421875, 0.02304915237426758, 0.023045120239257814, 0.023418880462646483, 0.02312294387817383, 0.022880319595336915, 0.022871999740600585, 0.023043071746826172, 0.023772159576416017, 0.02446335983276367, 0.0236298885345459, 0.023316415786743164, 0.023207935333251953, 0.023149568557739256, 0.02309017562866211, 0.022320127487182616, 0.022221824645996095, 0.023546880722045898, 0.023133184432983397, 0.022952959060668944, 0.02292531204223633, 0.02289254379272461, 0.023043071746826172, 0.0230830078125, 0.023371776580810546, 0.02326937675476074, 0.023151615142822265, 0.05610393524169922, 0.023630847930908205, 0.023274496078491212, 0.023163904190063478, 0.023244800567626952, 0.023198720932006835, 0.023384063720703126, 0.02323865509033203, 0.023167999267578124, 0.023385087966918947, 0.023422975540161133, 0.023143423080444335, 0.02308710479736328, 0.02326835250854492, 0.023241727828979493, 0.02330624008178711, 0.02330112075805664, 0.02352230453491211, 0.023266304016113282, 0.023143423080444335, 0.023262239456176757, 0.023076831817626955, 0.02328883171081543, 0.023189504623413085, 0.02326425552368164, 0.023202816009521485, 0.023218175888061524, 0.023237632751464843, 0.02327961540222168, 0.023237632751464843, 0.023258111953735353, 0.023307264328002928, 0.02328985595703125, 0.023129087448120117, 0.023185407638549805, 0.023200767517089844, 0.023386112213134767, 0.023215103149414062, 0.02325503921508789, 0.023348224639892577, 0.023408639907836915, 0.023215103149414062, 0.023198720932006835, 0.02325196838378906, 0.023225343704223633, 0.023380992889404296, 0.023137279510498047, 0.023146495819091797, 0.023323648452758788, 0.023992319107055664, 0.02430771255493164, 0.023427072525024413, 0.023596031188964844, 0.023347200393676756, 0.023249919891357423, 0.023185407638549805, 0.02345062446594238, 0.02323865509033203, 0.023203840255737306, 0.023153663635253906, 0.02369740867614746, 0.023154687881469727, 0.02329804801940918, 0.055006206512451174, 0.02348748779296875, 0.023326719284057617, 0.02326016044616699, 0.023147520065307618, 0.022994943618774414, 0.02306559944152832, 0.023174144744873046, 0.023447551727294923, 0.023389184951782226, 0.023565311431884766, 0.023165952682495116, 0.023085056304931642, 0.023145471572875977, 0.023149568557739256, 0.023155712127685548, 0.023034879684448242, 0.023188480377197264, 0.02331340789794922, 0.023370752334594725, 0.023353343963623048, 0.023044095993041993, 0.022890495300292968, 0.023018495559692383, 0.02308403205871582, 0.02289664077758789, 0.023003135681152344, 0.02308710479736328, 0.02304921531677246, 0.022914047241210937, 0.023018495559692383, 0.023031808853149413, 0.022821887969970703, 0.02298067283630371, 0.022905855178833007, 0.02285152053833008, 0.022922367095947267, 0.022827903747558595, 0.023295135498046876, 0.023107423782348632, 0.022951936721801756, 0.022935552597045897, 0.022916095733642578, 0.022816768646240236, 0.02286796760559082, 0.02301241683959961, 0.022929344177246094, 0.022952959060668944, 0.022854656219482423, 0.022815744400024415, 0.02323148727416992, 0.022874111175537108, 0.022898687362670898, 0.022929407119750975, 0.023088319778442383, 0.022900543212890624, 0.022822912216186524, 0.022861824035644532, 0.02287820816040039, 0.022920192718505858, 0.022947839736938477, 0.023216127395629883, 0.023380992889404296, 0.054978591918945316, 0.022420448303222658, 0.022261760711669923, 0.022445056915283205, 0.02230681610107422, 0.02225971221923828, 0.022347776412963868, 0.02240716743469238, 0.022294559478759766, 0.022413280487060545, 0.02229350471496582, 0.02228326416015625, 0.02229964828491211, 0.022344703674316405, 0.022305791854858398, 0.02227609634399414, 0.02230271911621094, 0.022305791854858398, 0.022409215927124023, 0.02269900894165039, 0.02227609634399414, 0.02230886459350586, 0.022331392288208008, 0.022313983917236328, 0.022390783309936522, 0.022305791854858398, 0.022381568908691408, 0.022410240173339844, 0.022396928787231447, 0.022297599792480468, 0.022150144577026368, 0.02231603240966797, 0.022344703674316405, 0.022328319549560546, 0.022337535858154296, 0.022336511611938475, 0.022351903915405272, 0.02230473518371582, 0.022351871490478514, 0.02229452705383301, 0.022379520416259766, 0.022337535858154296, 0.02231808090209961, 0.022429695129394533, 0.02230784034729004, 0.02240716743469238, 0.02230988883972168, 0.02225564765930176, 0.022439903259277343, 0.022329343795776366, 0.02230681610107422, 0.022268928527832032, 0.022403072357177735, 0.0223242244720459, 0.022365184783935548, 0.022380544662475587, 0.0222740478515625, 0.02226585578918457, 0.02309529685974121, 0.023193599700927735, 0.023163904190063478, 0.023185407638549805, 0.02330521583557129, 0.05482905578613281, 0.02245427131652832, 0.02236115264892578, 0.022313919067382813, 0.02228223991394043, 0.022371328353881836, 0.0223191032409668, 0.022286336898803712, 0.022301759719848633, 0.023443391799926758, 0.02293350410461426, 0.022311935424804686, 0.022399999618530272, 0.02228428840637207, 0.02236832046508789, 0.022343616485595703, 0.022319168090820313, 0.022335424423217773, 0.022378496170043945, 0.022296575546264647, 0.022331392288208008, 0.022312959671020507, 0.02233241653442383, 0.022323232650756836, 0.023557088851928712, 0.023173120498657225, 0.022648895263671875, 0.02297235107421875, 0.022961151123046874, 0.022391807556152343, 0.02242355155944824, 0.022566911697387695, 0.02236419105529785, 0.02232009506225586, 0.022374399185180666, 0.021836799621582033, 0.022156288146972656, 0.02208665657043457, 0.022064128875732423, 0.022032384872436524, 0.02206003189086914, 0.02205286407470703, 0.022112255096435548, 0.02201190376281738, 0.02205183982849121, 0.022923263549804687, 0.02270310401916504, 0.02267852783203125, 0.02229964828491211, 0.022033407211303712, 0.022090751647949217, 0.02209791946411133, 0.022072351455688477, 0.022026208877563475, 0.022060127258300782, 0.02209270477294922, 0.022041599273681642, 0.022029312133789062, 0.022191104888916017, 0.02206208038330078, 0.022204416275024414, 0.022189056396484375, 0.02229350471496582, 0.05492531204223633, 0.022343679428100584, 0.02228428840637207, 0.02227916717529297, 0.022185983657836913, 0.022288383483886717, 0.02228223991394043, 0.022210559844970702, 0.02231603240966797, 0.022432767868041992, 0.022215679168701173, 0.022350847244262697, 0.022230016708374024, 0.022297599792480468, 0.02233241653442383, 0.022194175720214843, 0.022393856048583984, 0.02233344078063965, 0.022311935424804686, 0.022337535858154296, 0.02229862403869629, 0.022334463119506837, 0.02227712059020996, 0.022296575546264647, 0.022363136291503907, 0.02230271911621094, 0.022361087799072265, 0.022368255615234374, 0.02226483154296875, 0.02281062316894531, 0.02223411178588867, 0.022297599792480468, 0.022492160797119142, 0.02226688003540039, 0.022296575546264647, 0.02225663948059082, 0.022658048629760744, 0.022839296340942384, 0.022271999359130858, 0.023152639389038086, 0.022804479598999023, 0.022354944229125977, 0.02231500816345215, 0.022258752822875975, 0.022264768600463867, 0.022363136291503907, 0.022227008819580077, 0.02221766471862793, 0.022206464767456056, 0.02224947166442871, 0.022252544403076172, 0.02220134353637695, 0.022303743362426756, 0.02230681610107422, 0.022221824645996095, 0.022328319549560546, 0.0224849910736084, 0.02227302360534668, 0.02227609634399414, 0.02224844741821289, 0.0223242244720459, 0.022181983947753905, 0.022335391998291015, 0.05477478408813476, 0.022334463119506837, 0.022175743103027345, 0.022268928527832032, 0.022262847900390625, 0.02219411277770996, 0.022338560104370117, 0.02213580894470215, 0.022252544403076172, 0.022255615234375, 0.022323200225830078, 0.022255615234375, 0.022323200225830078, 0.022238208770751954, 0.02227712059020996, 0.022320127487182616, 0.022331392288208008, 0.022271999359130858, 0.022347776412963868, 0.022512704849243163, 0.022311872482299804, 0.02227302360534668, 0.022271999359130858, 0.02229452705383301, 0.022231039047241212, 0.02229555130004883, 0.022358015060424806, 0.02229145622253418, 0.022254592895507814, 0.02225049591064453, 0.022368255615234374, 0.022252544403076172, 0.022336511611938475, 0.022330368041992187, 0.02228223991394043, 0.02227302360534668, 0.022361087799072265, 0.022373376846313478, 0.02244095993041992, 0.02249830436706543, 0.022293535232543946, 0.022239200592041014, 0.02234880065917969, 0.02230067253112793, 0.022412288665771486, 0.022352895736694335, 0.022262815475463868, 0.02237539291381836, 0.02224844741821289, 0.022271999359130858, 0.02245631980895996, 0.02468556785583496, 0.023274496078491212, 0.02234880065917969, 0.02233241653442383, 0.02228326416015625, 0.022328319549560546, 0.022331392288208008, 0.022355968475341798, 0.022785024642944338, 0.022534143447875975, 0.022313983917236328, 0.022358015060424806, 0.05495500946044922, 0.022394880294799805, 0.022388736724853517, 0.022261760711669923, 0.02227712059020996, 0.02220953559875488, 0.02226585578918457, 0.022288383483886717, 0.022326271057128907, 0.022501375198364256, 0.02230681610107422, 0.022296575546264647, 0.02229145622253418, 0.022257696151733397, 0.023346143722534178, 0.022845439910888672, 0.02240716743469238, 0.02243174362182617, 0.022399999618530272, 0.022125568389892578, 0.022190080642700196, 0.02239897537231445, 0.022405120849609376, 0.022321151733398437, 0.022328319549560546, 0.022372352600097657, 0.022429695129394533, 0.02243071937561035, 0.022985727310180663, 0.023193599700927735, 0.02323148727416992, 0.023157760620117186, 0.022347776412963868, 0.022377471923828125, 0.022624256134033204, 0.024639488220214844, 0.02385411262512207, 0.023233503341674806, 0.023357440948486328, 0.02325503921508789, 0.02332467269897461, 0.023221248626708983, 0.023212032318115236, 0.023362560272216795, 0.02342092704772949, 0.023341056823730468, 0.02329497528076172, 0.023375871658325196, 0.023243839263916016, 0.023259071350097655, 0.023259136199951173, 0.02325196838378906, 0.023240703582763672, 0.023149568557739256, 0.02332876777648926, 0.02327859115600586, 0.023368703842163087, 0.023267328262329103, 0.02326527976989746, 0.02323967933654785, 0.023372800827026367, 0.023365631103515624, 0.02326937675476074, 0.05528575897216797, 0.023348224639892577, 0.023189504623413085, 0.023409664154052736, 0.02348646354675293, 0.023257087707519532, 0.023149568557739256, 0.02330009651184082, 0.022226943969726562, 0.022347776412963868, 0.023008384704589845, 0.02307468795776367, 0.023234560012817384, 0.023200767517089844, 0.02330112075805664, 0.02325503921508789, 0.023205888748168944, 0.023311359405517578, 0.023451648712158202, 0.02327552032470703, 0.023208959579467774, 0.022408191680908202, 0.02262118339538574, 0.02304614448547363, 0.023221248626708983, 0.023258111953735353, 0.023235584259033205, 0.023236608505249022, 0.023235584259033205, 0.022334463119506837, 0.022381568908691408, 0.022387712478637696, 0.02251263999938965, 0.023274496078491212, 0.02308710479736328, 0.023072959899902344, 0.022961088180541992, 0.022991743087768555, 0.02309529685974121, 0.02432307243347168, 0.02454630470275879, 0.023562240600585937, 0.02330112075805664, 0.023259136199951173, 0.023229440689086913, 0.02329804801940918, 0.02327859115600586, 0.02330316734313965, 0.02322329521179199, 0.023451648712158202, 0.023347200393676756, 0.02330316734313965, 0.023218175888061524, 0.023214080810546874, 0.02334617614746094, 0.023282688140869142, 0.023197696685791015, 0.023258111953735353, 0.023222272872924804, 0.023178239822387696, 0.023195648193359376, 0.023611455917358398, 0.02323961639404297, 0.05541686248779297, 0.02276041603088379, 0.022387712478637696, 0.02230988883972168, 0.022323200225830078, 0.022363136291503907, 0.022377471923828125, 0.02245734405517578, 0.022372352600097657, 0.022353919982910156, 0.02231603240966797, 0.022331392288208008, 0.02234060859680176, 0.022339584350585938, 0.02230067253112793, 0.022322175979614257, 0.022346752166748047, 0.022386688232421875, 0.02245631980895996, 0.023244800567626952, 0.023359487533569336, 0.02327654457092285, 0.023224319458007812, 0.023212032318115236, 0.023193599700927735, 0.023214080810546874, 0.02327859115600586, 0.02330112075805664, 0.02330112075805664, 0.023209983825683594, 0.023391231536865235, 0.023365631103515624, 0.023362560272216795, 0.023244800567626952, 0.02329804801940918, 0.023152639389038086, 0.023310335159301757, 0.023232511520385742, 0.02327654457092285, 0.02311680030822754, 0.023367679595947266, 0.022374399185180666, 0.022413312911987306, 0.022374399185180666, 0.022373376846313478, 0.02270412826538086, 0.02244607925415039, 0.022389759063720704, 0.022420480728149415, 0.022487039566040038, 0.022367231369018553, 0.022314048767089843, 0.022409151077270508, 0.02242252731323242, 0.022409215927124023, 0.02255564880371094, 0.023267328262329103, 0.023315456390380858, 0.02327244758605957, 0.02329702377319336, 0.023222272872924804, 0.02346188735961914, 0.023400447845458985]",tokens/s,42.940398976611846,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,meta-llama/Meta-Llama-3-8B,meta-llama/Meta-Llama-3-8B,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.1,,,,MB,1272.36096,17201.364992,0.0,16554.917888,16436.726784,s,10,0.8342723770141601,0.08342723770141601,0.0003580420483213533,0.08333169555664063,0.08359898300170898,0.08401704368591308,0.08435149223327636,"[0.08443510437011718, 0.08306082916259766, 0.08335958099365234, 0.08332473754882813, 0.08333158111572266, 0.08333145904541016, 0.0835060806274414, 0.0831398696899414, 0.0833318099975586, 0.08345132446289062]",tokens/s,3068.5422057987535,kWh,9.875290799472067e-07,5.41013351166839e-07,6.494095473049896e-06,8.022637904163942e-06,tokens/kWh,31909703.897658132,MB,1272.36096,17201.364992,0.0,16554.917888,16436.729344,s,10,24.22832763671875,2.4228327636718747,0.0029210468648619924,2.4218022460937503,2.4242604736328124,2.427747424316406,2.430536984863281,"[2.431234375, 2.421704345703125, 2.421802978515625, 2.420526123046875, 2.422926025390625, 2.423485595703125, 2.421801513671875, 2.42078857421875, 2.42178955078125, 2.4222685546875]",tokens/s,26.00262013318725,kWh,2.8523219207094775e-05,1.5631873421980343e-05,0.00012452356489655212,0.00016867865752562725,tokens/kWh,373491.23430406983,,s,629,24.938764312744137,0.039648273947129,0.009960137955968645,0.03842047882080078,0.03852062835693359,0.038696978759765625,0.12226396057128906,"[0.03924582290649414, 0.03914547348022461, 0.03921612930297851, 0.03920281600952148, 0.039142398834228515, 0.03914956665039063, 0.039136257171630856, 0.03920387268066406, 0.03906864166259766, 0.03888947296142578, 0.03880448150634765, 0.039018497467041016, 0.038621185302734375, 0.03839897537231445, 0.038401023864746094, 0.03838259124755859, 0.038386688232421876, 0.03842867279052734, 0.03844607925415039, 0.038421504974365236, 0.0384450569152832, 0.03837849426269531, 0.03841740798950195, 0.03840204620361328, 0.03841228866577148, 0.038421504974365236, 0.038408191680908206, 0.038419456481933595, 0.038389759063720705, 0.038373374938964845, 0.03841126251220703, 0.038437889099121096, 0.038441982269287106, 0.03846041488647461, 0.03889766311645508, 0.038435840606689455, 0.038438911437988284, 0.03845734405517578, 0.03839078521728516, 0.038356990814208985, 0.0384716796875, 0.03853619384765625, 0.03869910430908203, 0.03869379043579101, 0.038454273223876956, 0.038430721282958984, 0.03843686294555664, 0.038419456481933595, 0.038400001525878906, 0.03840716934204102, 0.038419456481933595, 0.03837644958496094, 0.038384639739990234, 0.03839385604858398, 0.03841126251220703, 0.03844607925415039, 0.03862220764160156, 0.03838259124755859, 0.03845119857788086, 0.03841846466064453, 0.03840201568603516, 0.03835801696777344, 0.12203724670410156, 0.03843379211425781, 0.03842559814453125, 0.03847782516479492, 0.03833446502685547, 0.03839487838745117, 0.03840716934204102, 0.03851980972290039, 0.03882291030883789, 0.03846246337890625, 0.03836415863037109, 0.03840409469604492, 0.038386688232421876, 0.0383744010925293, 0.038368255615234374, 0.03837849426269531, 0.03839385604858398, 0.038437889099121096, 0.03835395050048828, 0.03838768005371094, 0.0383559684753418, 0.0384266242980957, 0.038406143188476564, 0.038384639739990234, 0.038362110137939456, 0.038391807556152346, 0.03841535949707031, 0.03847990417480469, 0.03837744140625, 0.038365184783935545, 0.0383559684753418, 0.03845939254760742, 0.03846963119506836, 0.038401023864746094, 0.03839078521728516, 0.038430721282958984, 0.03840921783447265, 0.03846041488647461, 0.03837747192382813, 0.038432769775390625, 0.038424575805664066, 0.03848191833496094, 0.03842559814453125, 0.038403072357177735, 0.03857305526733398, 0.038612991333007815, 0.03847372817993164, 0.03844607925415039, 0.03837644958496094, 0.03843379211425781, 0.038406143188476564, 0.038403072357177735, 0.03836723327636719, 0.03838566589355469, 0.03841331100463867, 0.038577152252197267, 0.038623233795166016, 0.03851059341430664, 0.038441982269287106, 0.03844095993041992, 0.03842867279052734, 0.03842764663696289, 0.038406143188476564, 0.12233830261230469, 0.0384450569152832, 0.038362110137939456, 0.03840512084960938, 0.03836723327636719, 0.03840921783447265, 0.03838054275512695, 0.038430721282958984, 0.03841843032836914, 0.03841126251220703, 0.038389759063720705, 0.03840921783447265, 0.03837542343139649, 0.038424575805664066, 0.03841228866577148, 0.0384983024597168, 0.038711296081542966, 0.03861913681030273, 0.03836006546020508, 0.03843174362182617, 0.038416385650634766, 0.03841535949707031, 0.03837235260009766, 0.038413375854492185, 0.03838560104370117, 0.03843174362182617, 0.03840204620361328, 0.03841024017333984, 0.038361152648925784, 0.03843782424926758, 0.038400001525878906, 0.038406143188476564, 0.038508544921875, 0.03841535949707031, 0.03838259124755859, 0.03849318313598633, 0.038418495178222656, 0.038470592498779294, 0.03846656036376953, 0.03841843032836914, 0.03846451187133789, 0.03849523162841797, 0.03853619384765625, 0.0384532470703125, 0.038365184783935545, 0.03850239944458008, 0.03845017623901367, 0.0384716796875, 0.03845119857788086, 0.03840204620361328, 0.03838566589355469, 0.03840921783447265, 0.03841535949707031, 0.03840921783447265, 0.0384389762878418, 0.038440895080566403, 0.03843174362182617, 0.038430721282958984, 0.03836928176879883, 0.03843379211425781, 0.03852799987792969, 0.038470657348632815, 0.03839692687988281, 0.12233010864257812, 0.038397953033447264, 0.038437889099121096, 0.03836928176879883, 0.03833958435058594, 0.038400001525878906, 0.038432769775390625, 0.038414337158203124, 0.03843174362182617, 0.038459457397460935, 0.03837843322753906, 0.03836726379394531, 0.038388702392578126, 0.038381568908691405, 0.03835599899291992, 0.03841331100463867, 0.038362079620361325, 0.03839078521728516, 0.03837849426269531, 0.038397953033447264, 0.03837747192382813, 0.038419456481933595, 0.03840921783447265, 0.038414337158203124, 0.03841126251220703, 0.0384983024597168, 0.03838873672485352, 0.03846656036376953, 0.03837952041625976, 0.038389759063720705, 0.038408191680908206, 0.03842972946166992, 0.03846652984619141, 0.0384634895324707, 0.03839897537231445, 0.038470657348632815, 0.03841024017333984, 0.03844710540771484, 0.03836723327636719, 0.038401023864746094, 0.03837952041625976, 0.03841331100463867, 0.03844095993041992, 0.038437889099121096, 0.03839897537231445, 0.038435840606689455, 0.03841126251220703, 0.03846451187133789, 0.03842969512939453, 0.038424575805664066, 0.03839897537231445, 0.03849420928955078, 0.038397953033447264, 0.03841331100463867, 0.03838771057128906, 0.03842047882080078, 0.03844710540771484, 0.03838159942626953, 0.03843376159667969, 0.03846451187133789, 0.038401023864746094, 0.03840512084960938, 0.0383744010925293, 0.12232704162597656, 0.03843686294555664, 0.03840512084960938, 0.038404129028320313, 0.03837436676025391, 0.038414337158203124, 0.038449153900146485, 0.038416385650634766, 0.03846041488647461, 0.038434814453125, 0.03839692687988281, 0.03842355346679688, 0.0384716796875, 0.03840716934204102, 0.03837849426269531, 0.038422527313232424, 0.03836723327636719, 0.03842867279052734, 0.03837747192382813, 0.03836620712280273, 0.03846144104003906, 0.038449153900146485, 0.03844607925415039, 0.03844403076171875, 0.03837644958496094, 0.038432769775390625, 0.038434814453125, 0.03851878356933594, 0.03847577667236328, 0.03840921783447265, 0.03851468658447266, 0.03854643249511719, 0.03841843032836914, 0.03851366424560547, 0.03845632171630859, 0.03846246337890625, 0.038468608856201174, 0.03840409469604492, 0.03843174362182617, 0.03841331100463867, 0.038434814453125, 0.03844710540771484, 0.03887820816040039, 0.03907174301147461, 0.03854848098754883, 0.038480926513671875, 0.03844707107543945, 0.03842047882080078, 0.038414337158203124, 0.038507518768310545, 0.038365184783935545, 0.03844403076171875, 0.038383617401123046, 0.03846144104003906, 0.03839078521728516, 0.038441982269287106, 0.03837644958496094, 0.038389759063720705, 0.03837644958496094, 0.0384983024597168, 0.03839487838745117, 0.03841228866577148, 0.038458366394042966, 0.12231986999511718, 0.038422527313232424, 0.03839692687988281, 0.03837952041625976, 0.03833446502685547, 0.038437889099121096, 0.038474750518798825, 0.038586368560791014, 0.03925094223022461, 0.038659072875976565, 0.03842355346679688, 0.038400001525878906, 0.03844710540771484, 0.03836928176879883, 0.038391807556152346, 0.038422527313232424, 0.038345729827880856, 0.0384266242980957, 0.03836006546020508, 0.03839385604858398, 0.038400001525878906, 0.038454273223876956, 0.038408191680908206, 0.038400001525878906, 0.03839385604858398, 0.03850035095214844, 0.03847884750366211, 0.03866726303100586, 0.0384450569152832, 0.038421504974365236, 0.03845017623901367, 0.038454273223876956, 0.03846041488647461, 0.0384983024597168, 0.038430721282958984, 0.0384450569152832, 0.03844095993041992, 0.03843993759155274, 0.038373374938964845, 0.03841024017333984, 0.03837952041625976, 0.03840409469604492, 0.03844403076171875, 0.038419456481933595, 0.038370304107666016, 0.03846656036376953, 0.03841740798950195, 0.038508544921875, 0.038371326446533204, 0.03850137710571289, 0.03844812774658203, 0.03841641616821289, 0.038383583068847656, 0.03846041488647461, 0.03847372817993164, 0.038790145874023435, 0.03861814498901367, 0.038424545288085935, 0.0384716796875, 0.03844607925415039, 0.03841331100463867, 0.038561790466308594, 0.03843174362182617, 0.12253593444824219, 0.038441982269287106, 0.038454273223876956, 0.03847884750366211, 0.038371326446533204, 0.038441982269287106, 0.03834777450561523, 0.03841740798950195, 0.03848806381225586, 0.03842559814453125, 0.038365184783935545, 0.03836928176879883, 0.03837235260009766, 0.038371326446533204, 0.03851776123046875, 0.038811649322509766, 0.03867136001586914, 0.03850649642944336, 0.038384639739990234, 0.03842559814453125, 0.038416385650634766, 0.03847788619995117, 0.03843884658813477, 0.03839385604858398, 0.03841740798950195, 0.038454273223876956, 0.03844607925415039, 0.038419456481933595, 0.03837644958496094, 0.03841024017333984, 0.03844403076171875, 0.0384532470703125, 0.03841843032836914, 0.03845017623901367, 0.038401023864746094, 0.03843587112426758, 0.038429664611816405, 0.038434814453125, 0.03839078521728516, 0.038421504974365236, 0.038408191680908206, 0.03851059341430664, 0.03846041488647461, 0.03839078521728516, 0.038408191680908206, 0.03842969512939453, 0.03840921783447265, 0.038392833709716793, 0.038384639739990234, 0.038419456481933595, 0.03843174362182617, 0.03839487838745117, 0.03839487838745117, 0.03842047882080078, 0.038438911437988284, 0.03844512176513672, 0.03839174270629883, 0.038400001525878906, 0.03842559814453125, 0.03844403076171875, 0.038416385650634766, 0.03839078521728516, 0.03842764663696289, 0.12212019348144532, 0.038368255615234374, 0.03839590454101562, 0.03835289764404297, 0.038332416534423826, 0.03838566589355469, 0.03840822219848633, 0.03838256072998047, 0.038385726928710937, 0.03839276885986328, 0.03836723327636719, 0.038416385650634766, 0.03837849426269531, 0.03842047882080078, 0.038400001525878906, 0.03841126251220703, 0.038431774139404296, 0.038392833709716793, 0.03839587020874023, 0.03841331100463867, 0.03842355346679688, 0.03855567932128906, 0.03841430282592773, 0.038384639739990234, 0.03835500717163086, 0.03841222381591797, 0.03840512084960938, 0.038458366394042966, 0.038392833709716793, 0.03839487838745117, 0.03837747192382813, 0.03845225524902344, 0.03843683242797852, 0.03842867279052734, 0.038381568908691405, 0.03843686294555664, 0.03841126251220703, 0.038470657348632815, 0.038400001525878906, 0.038419456481933595, 0.038406143188476564, 0.03853004837036133, 0.038416385650634766, 0.03839385604858398, 0.038386688232421876, 0.03841228866577148, 0.03839590454101562, 0.038474750518798825, 0.03834982299804687, 0.03844812774658203, 0.03837235260009766, 0.038441982269287106, 0.03836419296264648, 0.03845833587646484, 0.038708225250244144, 0.03848191833496094, 0.038386688232421876, 0.038424575805664066, 0.0383744010925293, 0.03851264190673828, 0.038458366394042966, 0.038373374938964845, 0.03842867279052734, 0.12233216094970703, 0.03846144104003906, 0.03838566589355469, 0.03840921783447265, 0.038356990814208985, 0.03843686294555664, 0.03837855911254883, 0.03842758560180664, 0.038392833709716793, 0.03847782516479492, 0.038389759063720705, 0.038422527313232424, 0.03846656036376953, 0.03839078521728516, 0.038391807556152346, 0.03838259124755859, 0.03840409469604492, 0.038386688232421876, 0.0383631362915039, 0.038434814453125, 0.0384716796875, 0.038452224731445314, 0.03838771057128906, 0.03837542343139649, 0.038389759063720705, 0.03841843032836914, 0.03841535949707031, 0.03842969512939453, 0.038392833709716793, 0.038454273223876956, 0.03838771057128906, 0.03848908615112305, 0.0385269775390625, 0.03861094284057617, 0.03842559814453125, 0.0385167350769043, 0.0383744010925293, 0.03843993759155274, 0.038414337158203124, 0.038386688232421876, 0.03838771057128906, 0.038523902893066404, 0.03877280044555664, 0.03860166549682617, 0.038401023864746094, 0.03846553421020508, 0.0384634895324707, 0.03844300842285156, 0.03840204620361328, 0.03844607925415039, 0.038373374938964845, 0.038441982269287106, 0.038397953033447264, 0.0384634895324707, 0.03836723327636719, 0.038435840606689455, 0.03838873672485352, 0.038422527313232424, 0.038383617401123046, 0.038421504974365236, 0.03840716934204102, 0.03838982391357422, 0.03848287963867188, 0.12245299530029297, 0.03836723327636719, 0.03839078521728516, 0.03840921783447265, 0.03841740798950195, 0.038422527313232424, 0.03837747192382813, 0.03849932861328125, 0.038504447937011715, 0.03860377502441406, 0.038422527313232424, 0.0384266242980957, 0.03848806381225586, 0.038383617401123046, 0.0383631362915039, 0.03845734405517578, 0.03839590454101562, 0.038421504974365236, 0.03836620712280273, 0.03841331100463867, 0.03847788619995117, 0.038469566345214846, 0.03844204711914063, 0.038459327697753905, 0.03838873672485352, 0.03851161575317383, 0.038523902893066404, 0.038434814453125, 0.03843993759155274, 0.03841126251220703, 0.03839590454101562, 0.03847987365722656, 0.03841542434692383, 0.038507457733154296, 0.038389759063720705, 0.03842969512939453, 0.03841331100463867, 0.03851571273803711, 0.038416385650634766, 0.038386688232421876, 0.03838873672485352, 0.038437889099121096, 0.03841843032836914, 0.03846144104003906, 0.0384532470703125, 0.03854848098754883, 0.038424575805664066, 0.038474750518798825, 0.03837644958496094, 0.038432769775390625, 0.038370304107666016, 0.03841024017333984, 0.03836723327636719, 0.03842764663696289, 0.038458366394042966, 0.03857612609863281, 0.038484992980957033, 0.03848601531982422, 0.03838259124755859, 0.03853823852539062, 0.0385269775390625, 0.038435840606689455, 0.03839897537231445]",tokens/s,25.22177891863592,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen2-beta-14B,Qwen/Qwen2-beta-14B,cuda,0,42,,,True,,,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",qwen2,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-32B,Qwen/Qwen1.5-32B,cuda,0,42,,,True,,,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",qwen2,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,0,0,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/0/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669325b5-33c200db10426a3f56278bfa;90257684-4aff-4d74-9e78-4716557553a0) Repository Not Found for url: https://huggingface.co/0/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: 0 is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/gpt-neo-125m,EleutherAI/gpt-neo-125m,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1183.854592,1084.751872,0.0,438.304768,353.979904,s,21,0.1660441279411316,0.00790686323529198,0.0002589561832836144,0.007907328128814697,0.008063327789306641,0.008089504241943359,0.008688697814941406,"[0.008838496208190917, 0.00787996816635132, 0.007734047889709473, 0.0076570878028869626, 0.007693984031677246, 0.007907328128814697, 0.00797654390335083, 0.007654399871826172, 0.007729983806610107, 0.007938560009002685, 0.007644959926605225, 0.008015487670898437, 0.00804032039642334, 0.007693632125854492, 0.008063327789306641, 0.008060128211975097, 0.008089504241943359, 0.008010687828063965, 0.007741087913513183, 0.007955967903137207, 0.007718624114990235]",tokens/s,32376.93537651617,kWh,9.464088646944434e-08,5.185899493041618e-08,2.3065478744391884e-07,3.7715466884377934e-07,tokens/kWh,678766620.5612779,MB,1183.854592,1084.751872,0.0,438.304768,363.871232,s,21,10.155196563720704,0.4835807887486049,0.009186378924115106,0.4881736755371094,0.490158203125,0.4901843872070312,0.5018353637695312,"[0.5047481079101562, 0.48843408203125, 0.47019036865234376, 0.46861312866210936, 0.4736457824707031, 0.4897067260742187, 0.4783285522460938, 0.468169921875, 0.4789617614746094, 0.4705062255859375, 0.4776983032226563, 0.4891998291015625, 0.4894041748046875, 0.490158203125, 0.489054931640625, 0.4901843872070312, 0.48867437744140624, 0.4862210388183594, 0.4868482666015625, 0.4881736755371094, 0.48827471923828125]",tokens/s,130.27812821727144,kWh,5.512294617940559e-06,3.019695316624375e-06,9.87660077310139e-06,1.8408590707666325e-05,tokens/kWh,3422315.211438941,,s,1322,10.304203780651093,0.007794405280371478,0.0010281766230401353,0.007729664087295533,0.007859097623825074,0.007991193413734436,0.01569559203147888,"[0.008508416175842285, 0.008594431877136231, 0.008777759552001952, 0.008491999626159668, 0.00799129581451416, 0.008014847755432129, 0.007952383995056152, 0.007979008197784423, 0.007980031967163086, 0.007858176231384278, 0.007865344047546387, 0.00791756820678711, 0.008002559661865234, 0.00802406406402588, 0.008073216438293456, 0.00799948787689209, 0.007865344047546387, 0.008028160095214844, 0.007997439861297608, 0.007987199783325195, 0.008102911949157715, 0.007961599826812745, 0.00806604766845703, 0.008018943786621094, 0.008069120407104492, 0.00787660789489746, 0.00791756820678711, 0.007880703926086426, 0.00790835189819336, 0.007928832054138184, 0.007970816135406494, 0.008086527824401855, 0.008099840164184571, 0.008268799781799317, 0.007961599826812745, 0.007915520191192627, 0.008004608154296875, 0.007958528041839599, 0.00797388792037964, 0.007881760120391846, 0.007931871891021729, 0.007925759792327881, 0.008044608116149902, 0.008069055557250977, 0.00784281587600708, 0.007879680156707763, 0.007976960182189942, 0.00790118408203125, 0.007864319801330566, 0.007878655910491944, 0.007993343830108643, 0.008062975883483887, 0.00801587200164795, 0.007996416091918946, 0.007930880069732665, 0.007913472175598145, 0.007875584125518798, 0.007798783779144287, 0.007759871959686279, 0.007662591934204102, 0.007809023857116699, 0.007976960182189942, 0.01590272045135498, 0.007558144092559814, 0.007910399913787843, 0.008190976142883302, 0.008156160354614257, 0.008259584426879883, 0.008300543785095215, 0.008244223594665527, 0.007879680156707763, 0.007758848190307617, 0.007762944221496582, 0.007763967990875244, 0.0077578239440917966, 0.007787551879882813, 0.007769055843353271, 0.007783423900604248, 0.007803904056549072, 0.007748608112335205, 0.00774348783493042, 0.007717887878417969, 0.007734272003173828, 0.007718912124633789, 0.007740416049957275, 0.00774451208114624, 0.007683072090148926, 0.007753727912902832, 0.0077506561279296875, 0.007745535850524903, 0.007738368034362793, 0.007828479766845703, 0.007776256084442139, 0.00773529577255249, 0.007739391803741455, 0.007742464065551758, 0.007731200218200684, 0.007723008155822754, 0.00773632001876831, 0.007725056171417236, 0.007730175971984863, 0.007706624031066894, 0.007730175971984863, 0.0077578239440917966, 0.007723008155822754, 0.007723008155822754, 0.0077608962059021, 0.007727104187011719, 0.007916543960571289, 0.00779366397857666, 0.007737343788146973, 0.007755775928497314, 0.0077619199752807615, 0.00781824016571045, 0.007758848190307617, 0.007614463806152344, 0.00744652795791626, 0.007478271961212158, 0.007481344223022461, 0.007468031883239746, 0.007501823902130127, 0.007449600219726562, 0.0074700798988342285, 0.007472127914428711, 0.007471168041229248, 0.015690688133239747, 0.007476223945617676, 0.007510015964508057, 0.007475200176239013, 0.007442431926727295, 0.007458816051483155, 0.0075038719177246095, 0.00744755220413208, 0.007483391761779785, 0.007449600219726562, 0.00745472002029419, 0.007482367992401123, 0.007444479942321777, 0.007449632167816162, 0.007460832118988037, 0.007491583824157715, 0.007501823902130127, 0.007459839820861816, 0.007450623989105225, 0.0074629120826721195, 0.007465983867645264, 0.007444479942321777, 0.0074332160949707035, 0.007510015964508057, 0.007449600219726562, 0.007482367992401123, 0.007423999786376953, 0.007452672004699707, 0.007398399829864502, 0.007423999786376953, 0.007443456172943115, 0.007456831932067871, 0.007415743827819824, 0.007476223945617676, 0.007437312126159668, 0.007483391761779785, 0.0074189438819885255, 0.007427008152008057, 0.007461887836456299, 0.00743833589553833, 0.0074291200637817386, 0.00744652795791626, 0.0074106879234313965, 0.0074158720970153805, 0.00740447998046875, 0.007505919933319092, 0.007473152160644531, 0.007453695774078369, 0.007427072048187256, 0.007431168079376221, 0.007455743789672851, 0.00745472002029419, 0.00743225622177124, 0.0074505600929260256, 0.00745472002029419, 0.007449600219726562, 0.007529471874237061, 0.007465983867645264, 0.007425024032592774, 0.007527488231658936, 0.0074577279090881346, 0.007477248191833496, 0.0074711041450500485, 0.01556275177001953, 0.00748748779296875, 0.007435264110565186, 0.007443456172943115, 0.00748748779296875, 0.007444479942321777, 0.007436287879943848, 0.007449600219726562, 0.007491583824157715, 0.007432191848754883, 0.0074403839111328125, 0.00744755220413208, 0.007431168079376221, 0.007464960098266602, 0.007437312126159668, 0.007481344223022461, 0.007469056129455566, 0.0074106879234313965, 0.007450623989105225, 0.007525375843048096, 0.007420928001403809, 0.007444479942321777, 0.007452672004699707, 0.00743833589553833, 0.007431168079376221, 0.007436287879943848, 0.007458816051483155, 0.007412735939025879, 0.007379968166351319, 0.007354368209838867, 0.007333888053894043, 0.007342080116271973, 0.007371776103973389, 0.007355391979217529, 0.007377920150756836, 0.007371776103973389, 0.007353343963623047, 0.007400447845458984, 0.007409664154052734, 0.007434239864349365, 0.007457791805267334, 0.00743936014175415, 0.00744652795791626, 0.007432191848754883, 0.007459839820861816, 0.00739737606048584, 0.007456768035888672, 0.007432191848754883, 0.007456768035888672, 0.007420928001403809, 0.007432191848754883, 0.007419904232025146, 0.007472127914428711, 0.007426047801971435, 0.007445504188537597, 0.00743936014175415, 0.007449600219726562, 0.007418879985809326, 0.00744652795791626, 0.007407616138458252, 0.007426047801971435, 0.007443456172943115, 0.007460864067077637, 0.01560985565185547, 0.007476223945617676, 0.007419904232025146, 0.007423999786376953, 0.007451648235321045, 0.007469056129455566, 0.007458816051483155, 0.007456768035888672, 0.007463935852050781, 0.007425024032592774, 0.007455743789672851, 0.007426047801971435, 0.007436287879943848, 0.00744755220413208, 0.007481344223022461, 0.007449600219726562, 0.007463935852050781, 0.0075008001327514645, 0.007448575973510742, 0.00744758415222168, 0.00744649600982666, 0.007463935852050781, 0.0074403839111328125, 0.007443456172943115, 0.007419904232025146, 0.007425024032592774, 0.00744755220413208, 0.007449600219726562, 0.007456768035888672, 0.007451648235321045, 0.007456768035888672, 0.0074301438331604, 0.007434239864349365, 0.007501823902130127, 0.007458816051483155, 0.007465023994445801, 0.007533504009246826, 0.007479296207427978, 0.007427072048187256, 0.0074711041450500485, 0.007425024032592774, 0.007468031883239746, 0.007444479942321777, 0.007455743789672851, 0.007432191848754883, 0.0074629120826721195, 0.007428095817565918, 0.007431168079376221, 0.007775231838226319, 0.007651328086853027, 0.007633920192718506, 0.007617536067962646, 0.007616511821746826, 0.007590911865234375, 0.007601151943206787, 0.007685120105743408, 0.007747583866119385, 0.007803904056549072, 0.007766016006469726, 0.007753727912902832, 0.0077025599479675294, 0.007787487983703613, 0.007748608112335205, 0.016122880935668944, 0.007781375885009765, 0.007771168231964111, 0.00775267219543457, 0.007747583866119385, 0.0077619199752807615, 0.007795711994171142, 0.007787519931793213, 0.00778547191619873, 0.007723008155822754, 0.007755775928497314, 0.007730175971984863, 0.007714879989624023, 0.00776800012588501, 0.0077608962059021, 0.007733248233795166, 0.007781375885009765, 0.0077608962059021, 0.007739391803741455, 0.007729152202606201, 0.007739391803741455, 0.007781375885009765, 0.007737343788146973, 0.007837696075439453, 0.007806975841522217, 0.00773632001876831, 0.00774348783493042, 0.007716864109039307, 0.007758848190307617, 0.00774348783493042, 0.007776256084442139, 0.007725056171417236, 0.007773183822631836, 0.007763967990875244, 0.007740416049957275, 0.007716864109039307, 0.007721983909606934, 0.007742464065551758, 0.007787519931793213, 0.007755775928497314, 0.007730175971984863, 0.007732223987579345, 0.007762944221496582, 0.007764992237091065, 0.00773529577255249, 0.007731264114379883, 0.007734208106994629, 0.007819263935089112, 0.008202239990234375, 0.007928832054138184, 0.007779327869415284, 0.007803904056549072, 0.007758848190307617, 0.00774348783493042, 0.007759871959686279, 0.007742464065551758, 0.007742464065551758, 0.007753759860992432, 0.007836639881134033, 0.00773529577255249, 0.007723008155822754, 0.007729152202606201, 0.007781375885009765, 0.016199680328369142, 0.007763967990875244, 0.0077578239440917966, 0.007709695816040039, 0.007732223987579345, 0.007733248233795166, 0.0077281279563903805, 0.007723008155822754, 0.00773632001876831, 0.007771135807037354, 0.00779366397857666, 0.007769087791442871, 0.007758848190307617, 0.007738368034362793, 0.007759871959686279, 0.007734272003173828, 0.007738368034362793, 0.0077281279563903805, 0.007790592193603516, 0.007738368034362793, 0.007751679897308349, 0.007738368034362793, 0.007796735763549805, 0.007786496162414551, 0.007770112037658691, 0.007751679897308349, 0.007555071830749512, 0.007490560054779053, 0.007479296207427978, 0.007473152160644531, 0.0074332160949707035, 0.007481344223022461, 0.0074700798988342285, 0.007469056129455566, 0.007477248191833496, 0.007480319976806641, 0.007473152160644531, 0.007476223945617676, 0.0074823999404907224, 0.007457759857177735, 0.007473152160644531, 0.0074670081138610836, 0.007456768035888672, 0.007469056129455566, 0.00745472002029419, 0.007436287879943848, 0.007445504188537597, 0.007453695774078369, 0.0074670081138610836, 0.007497727870941162, 0.007464960098266602, 0.007452672004699707, 0.00743936014175415, 0.00753766393661499, 0.007431168079376221, 0.007497727870941162, 0.007497727870941162, 0.007426047801971435, 0.007455743789672851, 0.007498752117156982, 0.007445504188537597, 0.007449600219726562, 0.007450623989105225, 0.015607808113098144, 0.007452672004699707, 0.007463935852050781, 0.007465983867645264, 0.00744652795791626, 0.007477248191833496, 0.007436287879943848, 0.007468031883239746, 0.007459839820861816, 0.007453695774078369, 0.00744755220413208, 0.0074291200637817386, 0.007426047801971435, 0.007425024032592774, 0.007459839820861816, 0.007465983867645264, 0.007523327827453613, 0.007436287879943848, 0.007417856216430664, 0.007332863807678222, 0.007374911785125732, 0.007379903793334961, 0.0073820161819458, 0.00738918399810791, 0.007435264110565186, 0.0073810238838195805, 0.007386079788208008, 0.0073359360694885255, 0.007392255783081054, 0.0074670081138610836, 0.007459839820861816, 0.007460864067077637, 0.007489535808563232, 0.007436287879943848, 0.0074670081138610836, 0.007419904232025146, 0.007359488010406494, 0.007365632057189942, 0.007321599960327148, 0.007326720237731933, 0.007370751857757568, 0.007377920150756836, 0.0074711041450500485, 0.007374847888946533, 0.007375872135162354, 0.007354368209838867, 0.007400447845458984, 0.007416831970214844, 0.007480319976806641, 0.007457856178283691, 0.007485375881195068, 0.007461887836456299, 0.007448575973510742, 0.007442431926727295, 0.007428095817565918, 0.007469056129455566, 0.007447616100311279, 0.007443391799926758, 0.007435264110565186, 0.007417856216430664, 0.007477248191833496, 0.007345151901245117, 0.0073431038856506346, 0.015643648147583008, 0.007477248191833496, 0.007482367992401123, 0.007453695774078369, 0.0074670081138610836, 0.0074741759300231934, 0.007493631839752197, 0.007479296207427978, 0.007464960098266602, 0.007463935852050781, 0.007481344223022461, 0.007452672004699707, 0.007411712169647216, 0.00745472002029419, 0.007455743789672851, 0.007497727870941162, 0.007490560054779053, 0.007532544136047363, 0.0074741759300231934, 0.007465983867645264, 0.0074670081138610836, 0.007434239864349365, 0.007449600219726562, 0.007457791805267334, 0.007461887836456299, 0.0074291200637817386, 0.007476223945617676, 0.007412735939025879, 0.007457791805267334, 0.007460864067077637, 0.007469056129455566, 0.007480319976806641, 0.007457791805267334, 0.007457791805267334, 0.007428095817565918, 0.007623680114746094, 0.007709695816040039, 0.007723008155822754, 0.007861248016357422, 0.007769087791442871, 0.0077578239440917966, 0.007773183822631836, 0.0077547521591186525, 0.007771135807037354, 0.007770112037658691, 0.007716864109039307, 0.00773529577255249, 0.00773632001876831, 0.007775231838226319, 0.007751679897308349, 0.007750688076019287, 0.007732192039489746, 0.007811071872711181, 0.007732223987579345, 0.007758848190307617, 0.007746560096740723, 0.0077209601402282715, 0.007740416049957275, 0.007811071872711181, 0.007875584125518798, 0.007768064022064209, 0.007725056171417236, 0.007872511863708496, 0.016161792755126952, 0.007508992195129394, 0.007475200176239013, 0.007498752117156982, 0.007436287879943848, 0.0074926080703735356, 0.007461887836456299, 0.007469056129455566, 0.007486464023590088, 0.007472127914428711, 0.007457791805267334, 0.007470111846923828, 0.007460832118988037, 0.0074926080703735356, 0.007473152160644531, 0.007491583824157715, 0.007482367992401123, 0.007449600219726562, 0.007517183780670166, 0.00745472002029419, 0.007416831970214844, 0.007452672004699707, 0.0074711041450500485, 0.007464960098266602, 0.007460864067077637, 0.007431168079376221, 0.007517216205596924, 0.007433184146881103, 0.007455743789672851, 0.007482367992401123, 0.007469056129455566, 0.007449600219726562, 0.007448575973510742, 0.007387135982513428, 0.007453695774078369, 0.007474207878112793, 0.007427040100097656, 0.007457791805267334, 0.00744652795791626, 0.007453728199005127, 0.007421919822692871, 0.007452672004699707, 0.007450623989105225, 0.007456768035888672, 0.007461887836456299, 0.007485439777374267, 0.007464960098266602, 0.007481344223022461, 0.007472127914428711, 0.007513088226318359, 0.007461887836456299, 0.007473152160644531, 0.0074670081138610836, 0.007435264110565186, 0.007391232013702393, 0.007349247932434082, 0.007511040210723877, 0.007370751857757568, 0.007402495861053467, 0.0073820161819458, 0.0074301438331604, 0.007463935852050781, 0.0074403839111328125, 0.015592448234558106, 0.007568384170532226, 0.007469056129455566, 0.0074670081138610836, 0.007455743789672851, 0.007478271961212158, 0.0074741759300231934, 0.0074629120826721195, 0.007459839820861816, 0.007453695774078369, 0.007477248191833496, 0.007478271961212158, 0.007448575973510742, 0.007464960098266602, 0.007472127914428711, 0.007493631839752197, 0.007436287879943848, 0.007450623989105225, 0.007464960098266602, 0.007475200176239013, 0.007444479942321777, 0.007459839820861816, 0.007444479942321777, 0.007418879985809326, 0.0074301438331604, 0.00743833589553833, 0.007480319976806641, 0.007550975799560547, 0.007465983867645264, 0.007417856216430664, 0.00742195177078247, 0.007457791805267334, 0.007445504188537597, 0.007451648235321045, 0.007475200176239013, 0.007486464023590088, 0.0074106879234313965, 0.007425024032592774, 0.007416895866394043, 0.00744543981552124, 0.007641088008880615, 0.007555071830749512, 0.007486464023590088, 0.00758784008026123, 0.008026111602783203, 0.008342528343200683, 0.007891007900238037, 0.007814080238342284, 0.007770112037658691, 0.007767039775848389, 0.007756800174713135, 0.00778547191619873, 0.0077844481468200685, 0.007734272003173828, 0.007767039775848389, 0.007759871959686279, 0.007753727912902832, 0.007836671829223632, 0.0077281279563903805, 0.007794688224792481, 0.007729152202606201, 0.0077199358940124516, 0.007779327869415284, 0.016257024765014647, 0.007768064022064209, 0.007740416049957275, 0.007774208068847656, 0.007770112037658691, 0.007783423900604248, 0.007766016006469726, 0.007773183822631836, 0.0077844481468200685, 0.007715839862823487, 0.007742464065551758, 0.007758848190307617, 0.007749631881713868, 0.00774348783493042, 0.007740416049957275, 0.007753727912902832, 0.007774240016937256, 0.00771888017654419, 0.007773183822631836, 0.007753727912902832, 0.007741439819335938, 0.0077916159629821775, 0.007748608112335205, 0.007779327869415284, 0.007798783779144287, 0.0077619199752807615, 0.0077209601402282715, 0.007767039775848389, 0.007767039775848389, 0.007790592193603516, 0.007730175971984863, 0.007767039775848389, 0.007707647800445557, 0.007769087791442871, 0.00774451208114624, 0.0077608962059021, 0.007870463848114014, 0.007774208068847656, 0.007708672046661377, 0.00774451208114624, 0.007730175971984863, 0.00774451208114624, 0.007752736091613769, 0.007754720211029052, 0.007804927825927735, 0.0077916159629821775, 0.007731200218200684, 0.00773529577255249, 0.00774348783493042, 0.0077547521591186525, 0.007749631881713868, 0.0077578239440917966, 0.007770112037658691, 0.007781375885009765, 0.00775270414352417, 0.007693312168121338, 0.0077209601402282715, 0.007748608112335205, 0.007742464065551758, 0.007756800174713135, 0.007774208068847656, 0.007825407981872558, 0.007748608112335205, 0.016258047103881835, 0.007734272003173828, 0.007729152202606201, 0.008004608154296875, 0.007766047954559326, 0.007758815765380859, 0.00774348783493042, 0.007741439819335938, 0.007798783779144287, 0.007782400131225586, 0.007841792106628418, 0.007746560096740723, 0.007725056171417236, 0.007747583866119385, 0.00775270414352417, 0.007751679897308349, 0.007745535850524903, 0.007738368034362793, 0.0077578239440917966, 0.007727104187011719, 0.007773183822631836, 0.0077547521591186525, 0.007737343788146973, 0.0077619199752807615, 0.0077547521591186525, 0.007748608112335205, 0.0077844481468200685, 0.007723008155822754, 0.00790835189819336, 0.0077578239440917966, 0.007725056171417236, 0.007824384212493896, 0.007727168083190918, 0.00770246410369873, 0.007739391803741455, 0.007766079902648926, 0.00771782398223877, 0.00773632001876831, 0.007724031925201416, 0.007755775928497314, 0.007733248233795166, 0.007893055915832519, 0.007721920013427735, 0.007727104187011719, 0.007717887878417969, 0.007731200218200684, 0.007749631881713868, 0.007769087791442871, 0.007772160053253174, 0.0077619199752807615, 0.00775270414352417, 0.007762944221496582, 0.007753727912902832, 0.007732223987579345, 0.007777279853820801, 0.007770112037658691, 0.007730175971984863, 0.007781375885009765, 0.007710720062255859, 0.007828479766845703, 0.007725056171417236, 0.00785923194885254, 0.007745503902435303, 0.01570406436920166, 0.007460864067077637, 0.007477248191833496, 0.007863296031951903, 0.00785920000076294, 0.007812096118927002, 0.007795711994171142, 0.007817215919494629, 0.007795711994171142, 0.007781375885009765, 0.007751679897308349, 0.007763967990875244, 0.007771135807037354, 0.007713791847229004, 0.007742464065551758, 0.007786496162414551, 0.007777279853820801, 0.007672832012176513, 0.007741439819335938, 0.007729152202606201, 0.007726079940795898, 0.007721983909606934, 0.007758848190307617, 0.007773183822631836, 0.00774348783493042, 0.007808000087738037, 0.007725056171417236, 0.007821311950683594, 0.007725056171417236, 0.007712768077850342, 0.007780352115631104, 0.007746560096740723, 0.007717887878417969, 0.007723008155822754, 0.008517631530761719, 0.008081439971923829, 0.007714784145355225, 0.007745600223541259, 0.007726016044616699, 0.007782400131225586, 0.007710752010345459, 0.007884768009185791, 0.007788576126098633, 0.00775267219543457, 0.007759871959686279, 0.007768064022064209, 0.007732223987579345, 0.007841792106628418, 0.007710720062255859, 0.007764992237091065, 0.007783423900604248, 0.007778304100036621, 0.007753727912902832, 0.007795711994171142, 0.007828479766845703, 0.007805952072143555, 0.00780083179473877, 0.007792640209197998, 0.00780185604095459, 0.007881728172302246, 0.007780352115631104, 0.007748672008514404, 0.007749567985534668, 0.016285696029663087, 0.007734272003173828, 0.007725056171417236, 0.007763967990875244, 0.008056832313537597, 0.007780352115631104, 0.007740416049957275, 0.0077608962059021, 0.00774348783493042, 0.007711775779724121, 0.00775164794921875, 0.007746560096740723, 0.007696383953094482, 0.007753727912902832, 0.007749631881713868, 0.007779327869415284, 0.007779327869415284, 0.007767039775848389, 0.00774451208114624, 0.007766016006469726, 0.007739391803741455, 0.007776256084442139, 0.007712768077850342, 0.007758848190307617, 0.007718912124633789, 0.00774348783493042, 0.007733248233795166, 0.007714816093444824, 0.007713791847229004, 0.007772160053253174, 0.00773632001876831, 0.007773183822631836, 0.007756800174713135, 0.007721983909606934, 0.007753727912902832, 0.007724095821380615, 0.007876543998718262, 0.007746560096740723, 0.007686143875122071, 0.00788479995727539, 0.007709695816040039, 0.007724031925201416, 0.007729152202606201, 0.00774451208114624, 0.007713791847229004, 0.00790118408203125, 0.007751679897308349, 0.007804927825927735, 0.007741439819335938, 0.007759871959686279, 0.007769087791442871, 0.007748608112335205, 0.007786496162414551, 0.007715839862823487, 0.007756832122802734, 0.00776803207397461, 0.007742464065551758, 0.007709695816040039, 0.007729152202606201, 0.007811071872711181, 0.007723008155822754, 0.007689216136932373, 0.007747583866119385, 0.016261119842529297, 0.007762944221496582, 0.007734272003173828, 0.0077578239440917966, 0.0077608962059021, 0.007779327869415284, 0.007746560096740723, 0.0077844481468200685, 0.0077578239440917966, 0.007771135807037354, 0.007763967990875244, 0.008446975708007813, 0.007830527782440186, 0.007772160053253174, 0.00774451208114624, 0.007792640209197998, 0.007777279853820801, 0.007716928005218506, 0.007746496200561523, 0.007732223987579345, 0.007788544178009033, 0.007741439819335938, 0.007874559879302979, 0.007809023857116699, 0.007755775928497314, 0.007788544178009033, 0.00773632001876831, 0.007762944221496582, 0.00793497610092163, 0.008110079765319824, 0.007792640209197998, 0.007771135807037354, 0.007731200218200684, 0.007731200218200684, 0.007730175971984863, 0.007749631881713868, 0.007699456214904785, 0.007751679897308349, 0.007702527999877929, 0.0077506561279296875, 0.007726079940795898, 0.007731200218200684, 0.007755775928497314, 0.007764992237091065, 0.0077547521591186525, 0.007759871959686279, 0.007733248233795166, 0.0077619199752807615, 0.007727104187011719, 0.0077506561279296875, 0.007726079940795898, 0.007766016006469726, 0.007748608112335205, 0.007718944072723389, 0.00770249605178833, 0.007734272003173828, 0.007730175971984863, 0.007742464065551758, 0.007762944221496582, 0.0077506561279296875, 0.00774451208114624, 0.007775231838226319, 0.007748608112335205, 0.016321535110473632, 0.007747583866119385, 0.00775270414352417, 0.007730175971984863, 0.007989247798919678, 0.007869440078735352, 0.007782400131225586, 0.00758784008026123, 0.007621632099151611, 0.007764992237091065, 0.00764518404006958, 0.007696383953094482, 0.00785203218460083, 0.007961599826812745, 0.007786496162414551, 0.007760928153991699, 0.0077803201675415035, 0.007778304100036621, 0.007739391803741455, 0.00773632001876831, 0.00773529577255249, 0.007727104187011719, 0.007702527999877929, 0.007724031925201416, 0.007742464065551758, 0.007675903797149658, 0.007763967990875244, 0.007727104187011719, 0.00774348783493042, 0.007771135807037354, 0.007733248233795166, 0.007741439819335938, 0.0077916159629821775, 0.007797760009765625, 0.007776256084442139, 0.007755775928497314, 0.00773638391494751, 0.007698368072509765, 0.0076902399063110355, 0.00783564805984497, 0.007779327869415284, 0.0076984319686889645, 0.007692287921905518, 0.007730175971984863, 0.0077281279563903805, 0.007745535850524903, 0.0077209601402282715, 0.007774208068847656, 0.007724063873291016, 0.007748576164245605, 0.0077578239440917966, 0.007732223987579345, 0.007887904167175292, 0.00773526382446289, 0.007732223987579345, 0.007702527999877929, 0.007739391803741455, 0.0077281279563903805, 0.007713791847229004, 0.007794688224792481, 0.007707647800445557, 0.007702527999877929, 0.00775270414352417, 0.016236543655395508, 0.007746560096740723, 0.007712768077850342, 0.007741439819335938, 0.007906303882598878, 0.007844863891601562, 0.0077844481468200685, 0.007838719844818116, 0.007863296031951903, 0.007805952072143555, 0.0077547521591186525, 0.007730175971984863, 0.007747583866119385, 0.007692287921905518, 0.007723008155822754, 0.007742464065551758, 0.0077281599044799806, 0.0076830401420593265, 0.007732223987579345, 0.007711743831634522, 0.0076984319686889645, 0.00774451208114624, 0.007742464065551758, 0.007751679897308349, 0.007732223987579345, 0.007708672046661377, 0.007699456214904785, 0.00880742359161377, 0.007705599784851074, 0.007916543960571289, 0.008073216438293456, 0.007772160053253174, 0.0077281279563903805, 0.007732223987579345, 0.007704576015472412, 0.007756800174713135, 0.007705599784851074, 0.007740416049957275, 0.007724031925201416, 0.007725152015686035, 0.0076706881523132325, 0.007463967800140381, 0.0074659519195556644, 0.0074065918922424315, 0.007445504188537597, 0.007480319976806641, 0.007459839820861816, 0.007460864067077637, 0.007483391761779785, 0.007480319976806641, 0.007465983867645264, 0.007479296207427978, 0.0074997758865356446, 0.007459839820861816, 0.007696383953094482, 0.007696383953094482, 0.0077209601402282715, 0.007718912124633789, 0.007734272003173828, 0.007714816093444824, 0.007725120067596436, 0.007790527820587158, 0.007830527782440186, 0.015715328216552735, 0.0074291200637817386, 0.007469056129455566, 0.007446591854095459, 0.0074546561241149905, 0.007479296207427978, 0.008175616264343261, 0.007968768119812012, 0.007771135807037354, 0.0077209601402282715, 0.007739391803741455, 0.007746560096740723, 0.007804927825927735, 0.00774348783493042, 0.007790592193603516, 0.007779327869415284, 0.007768064022064209, 0.007737343788146973, 0.007731200218200684, 0.007742464065551758, 0.007899136066436767, 0.007803904056549072, 0.0077209601402282715, 0.007729152202606201, 0.007768064022064209, 0.007875584125518798, 0.007774208068847656, 0.007701504230499268, 0.007738368034362793, 0.007717887878417969, 0.007858176231384278, 0.007733248233795166, 0.007702591896057129, 0.0077086081504821775, 0.007611455917358398, 0.007697343826293945, 0.007655424118041992, 0.007641088008880615, 0.007839744091033935, 0.007755775928497314, 0.007723008155822754, 0.0077547521591186525, 0.007733248233795166, 0.007679999828338623, 0.007734272003173828, 0.007703584194183349, 0.007840735912322998, 0.007748608112335205, 0.007709695816040039, 0.007747583866119385, 0.007749631881713868, 0.007829504013061523, 0.007705632209777832, 0.00770249605178833, 0.007733248233795166, 0.007531519889831543, 0.007420959949493408, 0.007745503902435303, 0.00783462381362915, 0.007805952072143555, 0.0076871681213378906, 0.007662591934204102, 0.007713791847229004, 0.01619865608215332, 0.00773529577255249, 0.007697408199310303, 0.00774348783493042, 0.007705599784851074, 0.007726079940795898, 0.007779327869415284, 0.007756800174713135, 0.007725056171417236, 0.007711743831634522, 0.007716864109039307, 0.007670783996582031, 0.007718912124633789, 0.00773737621307373, 0.0076973757743835445, 0.007748608112335205, 0.007751679897308349, 0.007701536178588867, 0.007738336086273194, 0.007740416049957275, 0.007787519931793213, 0.007703551769256592, 0.007732223987579345, 0.007745535850524903, 0.007706624031066894, 0.007799808025360107, 0.007709695816040039, 0.007729152202606201, 0.007689216136932373, 0.0076912641525268555, 0.007709695816040039, 0.007713791847229004, 0.007677951812744141, 0.007721983909606934, 0.007733248233795166, 0.007773183822631836, 0.007711743831634522, 0.007751711845397949, 0.008095711708068847, 0.008062975883483887, 0.007741439819335938, 0.00781824016571045, 0.007732223987579345, 0.007692287921905518, 0.007678976058959961, 0.007712768077850342, 0.00781004810333252, 0.007712768077850342, 0.007857151985168457, 0.00820019245147705, 0.007748608112335205, 0.007740416049957275, 0.007766016006469726, 0.007707647800445557, 0.007692287921905518, 0.007710720062255859, 0.007723008155822754, 0.007725056171417236, 0.007773183822631836, 0.007824384212493896, 0.007688191890716553, 0.007516160011291504, 0.007334911823272705, 0.015696895599365233, 0.007783423900604248, 0.007747583866119385, 0.007790592193603516, 0.00787660789489746, 0.007758848190307617, 0.007746560096740723, 0.00799948787689209, 0.00782643222808838, 0.007721983909606934, 0.007871488094329833, 0.007756800174713135, 0.007780352115631104, 0.0077547521591186525, 0.007823359966278077, 0.007765024185180664, 0.0077311677932739254, 0.007716864109039307, 0.007778304100036621, 0.007738368034362793, 0.0077506561279296875, 0.007847936153411865, 0.007766016006469726, 0.007747583866119385, 0.007709695816040039, 0.007783423900604248, 0.007816192150115966, 0.007683072090148926, 0.007713791847229004, 0.0077844481468200685, 0.007692287921905518, 0.007734272003173828, 0.007758848190307617, 0.007688191890716553, 0.00782643222808838, 0.007875584125518798, 0.007775231838226319, 0.007696383953094482, 0.007585792064666748, 0.007632895946502686, 0.007588863849639893, 0.007574528217315674, 0.00758681583404541, 0.007552000045776367, 0.007601151943206787, 0.007579648017883301, 0.0075980801582336424, 0.007836671829223632, 0.0077608962059021, 0.007763967990875244, 0.007860223770141601, 0.007751679897308349, 0.007755775928497314, 0.007772160053253174, 0.007746560096740723, 0.007729152202606201, 0.007723008155822754, 0.0077281279563903805, 0.007744575977325439, 0.007871424198150635, 0.00780185604095459, 0.0077281279563903805, 0.007820288181304931]",tokens/s,128.29715212760152,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,2,2,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/2/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669324e8-0400d2dc76a16f7c6823ee02;042cbc5a-e209-410a-8098-bef74626c6c3) Repository Not Found for url: https://huggingface.co/2/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: 2 is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/xglm-564M,facebook/xglm-564M,cuda,0,42,,,True,,,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",xglm,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-14B,Qwen/Qwen-14B,cuda,0,42,,,True,,,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",qwen,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,huggyllama/llama-7b,huggyllama/llama-7b,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.1,,,,MB,1246.285824,14529.593344,0.0,13883.14624,13705.186304,s,10,0.938450180053711,0.09384501800537108,0.00019577791226699223,0.09376971435546874,0.09399090576171874,0.09418457412719726,0.09433950881958007,"[0.09437824249267578, 0.09368576049804687, 0.09371097564697266, 0.09381791687011719, 0.09378534698486328, 0.09375408172607422, 0.09391398620605469, 0.09371485137939453, 0.09394786834716796, 0.09374114990234375]",tokens/s,2727.9018688594447,kWh,1.1102459435027211e-06,6.082857795076409e-07,5.9449606230467145e-06,7.663492346057078e-06,tokens/kWh,33405135.470868427,MB,1246.285824,14529.593344,0.0,13883.14624,13856.22528,s,10,22.596658203125,2.2596658203124997,0.0014668920688748867,2.25974267578125,2.2615420166015623,2.261759655761719,2.2619337670898436,"[2.2594091796875, 2.258534423828125, 2.25855029296875, 2.260200439453125, 2.258156005859375, 2.260076171875, 2.25726806640625, 2.26099267578125, 2.261977294921875, 2.26149365234375]",tokens/s,27.880228763776866,kWh,2.661313820835339e-05,1.45850039414841e-05,0.000113523023840154,0.0001547211659899915,tokens/kWh,407184.10824331106,,s,629,23.40419585037232,0.037208578458461546,0.011207920818503291,0.03585740661621094,0.035990732574462884,0.03602493438720703,0.13015347290039062,"[0.03576633453369141, 0.03570272064208985, 0.03567923355102539, 0.03569868850708008, 0.03576319885253906, 0.03578777694702148, 0.035759105682373046, 0.035757057189941405, 0.03573964691162109, 0.03572531127929687, 0.03573555374145508, 0.035722240447998044, 0.03572633743286133, 0.035746814727783204, 0.03583590316772461, 0.0357918701171875, 0.035767295837402346, 0.03577142333984375, 0.03578774261474609, 0.03594137573242188, 0.03580928039550781, 0.03577139282226562, 0.03576319885253906, 0.035796993255615236, 0.0358656005859375, 0.035850238800048825, 0.03580313491821289, 0.03581644821166992, 0.03580723190307617, 0.03582156753540039, 0.035829761505126956, 0.035789825439453124, 0.035806209564208984, 0.035860481262207033, 0.03585126495361328, 0.03587583923339844, 0.03589023971557617, 0.03598944091796875, 0.035827713012695314, 0.03584511947631836, 0.03585433578491211, 0.03586969757080078, 0.0358809585571289, 0.035884033203125, 0.03584000015258789, 0.0358922233581543, 0.03591680145263672, 0.03594137573242188, 0.03593830490112305, 0.035888126373291016, 0.03603046417236328, 0.036225025177001956, 0.03597619247436523, 0.03592396926879883, 0.035937278747558594, 0.03602431869506836, 0.035939327239990236, 0.035937278747558594, 0.035939327239990236, 0.035932159423828124, 0.03596799850463867, 0.035972095489501955, 0.13008793640136718, 0.035683326721191407, 0.03566387176513672, 0.035664894104003905, 0.035745792388916016, 0.035748863220214845, 0.035710975646972655, 0.035740673065185545, 0.035768318176269534, 0.03575091171264649, 0.035773441314697264, 0.03581644821166992, 0.03583283233642578, 0.03596492767333984, 0.035958782196044925, 0.03571507263183594, 0.03576115036010742, 0.035740673065185545, 0.03575807952880859, 0.03577139282226562, 0.0357283821105957, 0.03580518341064453, 0.03584000015258789, 0.035773441314697264, 0.03575500869750976, 0.03581542587280274, 0.03581644821166992, 0.03581951904296875, 0.03583488082885742, 0.0358737907409668, 0.035877918243408205, 0.0358502082824707, 0.03582668685913086, 0.03582156753540039, 0.035806209564208984, 0.03584102249145508, 0.03588608169555664, 0.03583488082885742, 0.03588915252685547, 0.03587583923339844, 0.03582156753540039, 0.035866622924804685, 0.03589836883544922, 0.03585433578491211, 0.035862529754638675, 0.03589017486572266, 0.0359444465637207, 0.03592396926879883, 0.0359106559753418, 0.03592396926879883, 0.0358809585571289, 0.03593011093139648, 0.035937278747558594, 0.03591372680664062, 0.03588198471069336, 0.035915775299072264, 0.03594035339355469, 0.0359444465637207, 0.03592601776123047, 0.03593625640869141, 0.0359106559753418, 0.03600281524658203, 0.035985408782958986, 0.13015347290039062, 0.035710975646972655, 0.03567718505859375, 0.03569049453735352, 0.035932159423828124, 0.03577036666870117, 0.035727359771728515, 0.0357283821105957, 0.03576422500610352, 0.035678207397460936, 0.035740673065185545, 0.035724288940429685, 0.03574169540405273, 0.03572326278686523, 0.03567001724243164, 0.035691520690917966, 0.03571712112426758, 0.03574169540405273, 0.035796993255615236, 0.035759105682373046, 0.03581542587280274, 0.035759105682373046, 0.03575296020507813, 0.03583692932128906, 0.03579289627075195, 0.03579084777832031, 0.03580108642578125, 0.035810302734375, 0.03581644821166992, 0.03579904174804688, 0.03584511947631836, 0.035917823791503906, 0.036669441223144535, 0.036071422576904294, 0.03590041732788086, 0.03585945510864258, 0.03585740661621094, 0.035860481262207033, 0.03585433578491211, 0.03581542587280274, 0.03584102249145508, 0.03582259368896484, 0.03581644821166992, 0.035796993255615236, 0.03583488082885742, 0.03584921646118164, 0.03591884613037109, 0.035899391174316404, 0.03589427185058594, 0.03586969757080078, 0.035885055541992186, 0.035929088592529294, 0.03589120101928711, 0.035846145629882815, 0.03587788772583008, 0.035920894622802735, 0.03599871826171875, 0.036001792907714845, 0.036004863739013675, 0.03591680145263672, 0.03592704010009766, 0.03590348815917969, 0.03588915252685547, 0.13025074768066405, 0.03566796875, 0.03567001724243164, 0.03565363311767578, 0.035676158905029294, 0.035737598419189456, 0.035789825439453124, 0.0357283821105957, 0.035730430603027344, 0.035740673065185545, 0.035765281677246095, 0.035783679962158206, 0.035772384643554686, 0.03574476623535156, 0.03571200180053711, 0.03577036666870117, 0.03575500869750976, 0.03575807952880859, 0.03576422500610352, 0.03579391860961914, 0.03581644821166992, 0.0358205451965332, 0.03580316925048828, 0.03578979110717773, 0.03577139282226562, 0.035796993255615236, 0.03582156753540039, 0.03582463836669922, 0.03584307098388672, 0.03584000015258789, 0.03588608169555664, 0.03588710403442383, 0.03585740661621094, 0.03585228729248047, 0.035846145629882815, 0.035896320343017575, 0.03586867141723633, 0.03589120101928711, 0.03587071990966797, 0.035844097137451174, 0.035942401885986325, 0.03601408004760742, 0.03593011093139648, 0.035931137084960936, 0.03589734268188476, 0.035901439666748046, 0.03593523025512695, 0.035952640533447267, 0.035929088592529294, 0.03589427185058594, 0.036087806701660154, 0.036286495208740235, 0.03608572769165039, 0.03600896072387695, 0.0359741439819336, 0.03601510238647461, 0.03602742385864258, 0.03600073623657227, 0.036034561157226565, 0.03601408004760742, 0.0360447998046875, 0.03609395217895508, 0.03596083068847656, 0.13020672607421874, 0.03569868850708008, 0.03577958297729492, 0.03568435287475586, 0.035732479095458985, 0.03572633743286133, 0.03571814346313477, 0.03574784088134766, 0.03570483016967774, 0.03570073699951172, 0.03576115036010742, 0.03582668685913086, 0.03580416107177734, 0.03570278549194336, 0.035743743896484374, 0.03577753448486328, 0.03571507263183594, 0.035794944763183595, 0.03578675079345703, 0.03588915252685547, 0.035796993255615236, 0.035784702301025394, 0.03576319885253906, 0.035765247344970705, 0.03575807952880859, 0.03578572845458984, 0.035817470550537106, 0.03578060913085938, 0.03580108642578125, 0.035810302734375, 0.0358175048828125, 0.0357979850769043, 0.03582668685913086, 0.03579084777832031, 0.03582668685913086, 0.035855358123779296, 0.03586150360107422, 0.035871742248535156, 0.03583180618286133, 0.035937278747558594, 0.035912704467773435, 0.03589427185058594, 0.035915775299072264, 0.03592499160766602, 0.03591884613037109, 0.03590553665161133, 0.03588915252685547, 0.03588608169555664, 0.03590041732788086, 0.03590655899047852, 0.03590553665161133, 0.03592499160766602, 0.03599359893798828, 0.03588198471069336, 0.03594854354858398, 0.03594342422485351, 0.035915775299072264, 0.03591372680664062, 0.035917823791503906, 0.03593318557739258, 0.03594649505615234, 0.0359628791809082, 0.03594342422485351, 0.13020057678222657, 0.035688449859619144, 0.035784702301025394, 0.03594854354858398, 0.0358389778137207, 0.0358389778137207, 0.03575091171264649, 0.03574169540405273, 0.035813377380371096, 0.035707904815673826, 0.035727359771728515, 0.03570483016967774, 0.03576115036010742, 0.03575398254394531, 0.03574272155761719, 0.035727359771728515, 0.03572019195556641, 0.03575603103637695, 0.03575296020507813, 0.03578572845458984, 0.0357724494934082, 0.035773406982421874, 0.03587276840209961, 0.035814399719238284, 0.03580316925048828, 0.03581539154052735, 0.0358656005859375, 0.035862529754638675, 0.03585433578491211, 0.03582156753540039, 0.03606016159057617, 0.03600691223144531, 0.03589734268188476, 0.03590041732788086, 0.03591167831420899, 0.035888126373291016, 0.03592396926879883, 0.035947521209716796, 0.03601203155517578, 0.03584819030761719, 0.035888126373291016, 0.035947521209716796, 0.03592499160766602, 0.035855358123779296, 0.035884033203125, 0.03590553665161133, 0.03591884613037109, 0.03590655899047852, 0.035921985626220704, 0.035943359375, 0.03598438262939453, 0.03597824096679687, 0.035937278747558594, 0.03588710403442383, 0.03595161437988281, 0.03598233413696289, 0.035970046997070314, 0.0359035530090332, 0.03594953536987305, 0.035911647796630856, 0.03595673751831055, 0.0359813117980957, 0.0359659538269043, 0.1303214111328125, 0.03565465545654297, 0.035659774780273434, 0.03564748764038086, 0.035661823272705076, 0.035672065734863284, 0.03572326278686523, 0.03567411041259766, 0.03570483016967774, 0.035675136566162106, 0.03572019195556641, 0.035732479095458985, 0.03579596710205078, 0.03568947219848633, 0.03570073699951172, 0.035765247344970705, 0.03582668685913086, 0.0359024658203125, 0.03593011093139648, 0.035796993255615236, 0.03584511947631836, 0.035748863220214845, 0.03573657608032227, 0.03576422500610352, 0.03569664001464844, 0.035789825439453124, 0.035813377380371096, 0.03577756881713867, 0.03578569412231445, 0.035781631469726564, 0.035844097137451174, 0.03585945510864258, 0.03580518341064453, 0.03577446365356445, 0.03578675079345703, 0.03581542587280274, 0.0358205451965332, 0.03583590316772461, 0.03581951904296875, 0.03579596710205078, 0.03588710403442383, 0.03584000015258789, 0.035833854675292966, 0.03582156753540039, 0.03580313491821289, 0.03583795166015625, 0.035922943115234376, 0.03589836883544922, 0.03601919937133789, 0.036324352264404294, 0.03612057495117187, 0.035912704467773435, 0.03590041732788086, 0.03585228729248047, 0.0358737907409668, 0.03588915252685547, 0.03589734268188476, 0.035972095489501955, 0.03592192077636719, 0.03586355209350586, 0.03591167831420899, 0.035899391174316404, 0.035896320343017575, 0.13015347290039062, 0.03570073699951172, 0.03575500869750976, 0.035707904815673826, 0.03578777694702148, 0.03571712112426758, 0.03575603103637695, 0.03577958297729492, 0.03589734268188476, 0.035814399719238284, 0.035846145629882815, 0.03584819030761719, 0.035781631469726564, 0.035727359771728515, 0.03577958297729492, 0.035775489807128906, 0.03577139282226562, 0.035784702301025394, 0.03577753448486328, 0.03580518341064453, 0.03577753448486328, 0.035879936218261715, 0.035942401885986325, 0.03586764907836914, 0.03588608169555664, 0.035934207916259765, 0.03580108642578125, 0.03582361602783203, 0.035860481262207033, 0.035866622924804685, 0.035945472717285154, 0.0359024658203125, 0.03594342422485351, 0.03587686538696289, 0.03592806243896484, 0.0359024658203125, 0.035896320343017575, 0.035862529754638675, 0.035884033203125, 0.03594956970214844, 0.03592396926879883, 0.03592396926879883, 0.035917823791503906, 0.03590553665161133, 0.03591680145263672, 0.03594137573242188, 0.03592499160766602, 0.03594342422485351, 0.03589324951171875, 0.03595980834960937, 0.036018177032470705, 0.035934207916259765, 0.035942401885986325, 0.03591987228393555, 0.03594342422485351, 0.03597721481323242, 0.035999744415283204, 0.03602329635620117, 0.036050945281982424, 0.036040702819824216, 0.0359628791809082, 0.036013057708740234, 0.03595673751831055, 0.130482177734375, 0.03599155044555664, 0.03595468902587891, 0.03578060913085938, 0.035811328887939455, 0.03574787139892578, 0.03580617523193359, 0.03582566452026367, 0.03579289627075195, 0.03575603103637695, 0.03578879928588867, 0.035813377380371096, 0.035762176513671876, 0.0358287353515625, 0.03574272155761719, 0.035807262420654296, 0.0358502082824707, 0.035880992889404294, 0.035803104400634764, 0.035844097137451174, 0.03581542587280274, 0.035794944763183595, 0.03591680145263672, 0.03589529418945313, 0.03589427185058594, 0.035800064086914066, 0.03591372680664062, 0.035846145629882815, 0.035937278747558594, 0.03590041732788086, 0.03591372680664062, 0.03587481689453125, 0.03584307098388672, 0.03586969757080078, 0.03586764907836914, 0.035961856842041014, 0.03591680145263672, 0.035931137084960936, 0.0358922233581543, 0.035947521209716796, 0.03599052810668945, 0.035985408782958986, 0.03598233413696289, 0.035899391174316404, 0.03587071990966797, 0.03593318557739258, 0.03593523025512695, 0.035920894622802735, 0.035931137084960936, 0.03599564743041992, 0.035983360290527344, 0.03594649505615234, 0.03600896072387695, 0.035952640533447267, 0.035995712280273436, 0.036061119079589844, 0.035997726440429687, 0.03592393493652344, 0.03594342422485351, 0.03594342422485351, 0.03602022552490235, 0.035939327239990236, 0.03595161437988281, 0.1302087707519531, 0.03575296020507813, 0.03567923355102539, 0.03570483016967774, 0.03579391860961914, 0.03578777694702148, 0.03575091171264649, 0.03570278549194336, 0.03577040100097656, 0.03576931381225586, 0.03575500869750976, 0.03574784088134766, 0.03573657608032227, 0.03573350524902344, 0.03577958297729492, 0.035781631469726564, 0.03585638427734375, 0.03585228729248047, 0.03584819030761719, 0.03624448013305664, 0.03592192077636719, 0.0358021125793457, 0.035806209564208984, 0.035850238800048825, 0.0358021125793457, 0.035801120758056644, 0.03581744003295898, 0.03585740661621094, 0.03589529418945313, 0.035783679962158206, 0.035846145629882815, 0.03585228729248047, 0.03590041732788086, 0.03590963363647461, 0.035884033203125, 0.035915775299072264, 0.03589120101928711, 0.035937278747558594, 0.03595775985717774, 0.03597107315063477, 0.03594854354858398, 0.03599462509155273, 0.03586150360107422, 0.035843135833740235, 0.03588294219970703, 0.03596492767333984, 0.03611443328857422, 0.036127742767333985, 0.035958782196044925, 0.03597721481323242, 0.036004863739013675, 0.03599871826171875, 0.035958782196044925, 0.0360263671875, 0.0359659538269043, 0.03599769592285156, 0.03604582214355469, 0.03602431869506836, 0.03596083068847656, 0.03599462509155273, 0.03602534484863281, 0.03600588989257812, 0.035955711364746096]",tokens/s,26.87552283450892,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-180B,tiiuae/falcon-180B,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 403 Client Error: Forbidden for url: https://huggingface.co/tiiuae/falcon-180B/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6693271f-7d8c4c450dcdd82f1bb03763;6865bc64-244f-4c95-b6f1-b4073edcdaf8) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/tiiuae/falcon-180B/resolve/main/config.json. If you are trying to create or update content,make sure you have a token with the `write` role. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1826, in _raise_on_head_call_error raise LocalEntryNotFoundError( huggingface_hub.utils._errors.LocalEntryNotFoundError: An error happened while trying to locate the file on the Hub and we cannot find the requested files in the local cache. Please check your connection and try again or make sure your Internet connection is on. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 445, in cached_file raise EnvironmentError( OSError: We couldn't connect to 'https://huggingface.co' to load this file, couldn't find it in the cached files and it looks like tiiuae/falcon-180B is not the path to a directory containing a file named config.json. Checkout your internet connection or see how to run the library in offline mode at 'https://huggingface.co/docs/transformers/installation#offline-mode'. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/gpt-neo-1.3B,EleutherAI/gpt-neo-1.3B,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1214.91456,3760.717824,0.0,3114.27072,2817.469952,s,12,0.23134873199462888,0.01927906099955241,0.00026746353881322696,0.019206063270568847,0.019287216567993162,0.019680835247039792,0.020060684871673583,"[0.02015564727783203, 0.01920748710632324, 0.0192923526763916, 0.01920921516418457, 0.019201183319091798, 0.019185440063476562, 0.019141151428222657, 0.019204639434814454, 0.019240991592407226, 0.019131391525268555, 0.019214271545410156, 0.019164960861206056]",tokens/s,13278.65501364114,kWh,2.2886718264862004e-07,1.2539962876775114e-07,1.1312189363520563e-06,1.4854857477684275e-06,tokens/kWh,172334201.3779508,MB,1215.24224,3760.717824,0.0,3114.27072,2884.408832,s,12,10.429557800292969,0.8691298166910807,0.010457833467539337,0.8663155822753906,0.876654443359375,0.8881258422851562,0.8985324438476563,"[0.9011340942382813, 0.8660888061523437, 0.8665603637695313, 0.8665423583984375, 0.8603411254882812, 0.8625307006835937, 0.8774827270507812, 0.8691998901367187, 0.8643333129882812, 0.8641051025390625, 0.86664794921875, 0.8645913696289063]",tokens/s,72.48629467097481,kWh,1.0569064838897021e-05,5.7913664041289504e-06,3.189602970764714e-05,4.8256460950673126e-05,tokens/kWh,1305524.6646536605,,s,755,10.626029571533193,0.014074211353024108,0.002366265441901744,0.013703167915344238,0.014147353553771972,0.014442495727539062,0.0333514656829834,"[0.014749695777893066, 0.01497599983215332, 0.0144650239944458, 0.014549023628234863, 0.014733280181884766, 0.014394368171691894, 0.014439423561096192, 0.01427455997467041, 0.014449664115905762, 0.014450688362121582, 0.014438400268554688, 0.014404607772827148, 0.014318592071533203, 0.014330880165100097, 0.014332927703857423, 0.014477312088012695, 0.014540800094604492, 0.014366720199584961, 0.014377984046936035, 0.01439846420288086, 0.014355456352233887, 0.014511103630065919, 0.014454784393310547, 0.014334976196289062, 0.014477312088012695, 0.014304256439208985, 0.014869503974914551, 0.014825471878051758, 0.014395392417907715, 0.014342144012451171, 0.014264320373535156, 0.014314496040344238, 0.014295040130615234, 0.013964287757873535, 0.014592000007629394, 0.014709759712219238, 0.014821375846862793, 0.01461350440979004, 0.014486528396606446, 0.014622719764709472, 0.014363648414611817, 0.014127103805541993, 0.014148608207702636, 0.014051327705383301, 0.014082048416137695, 0.01397862434387207, 0.013642784118652343, 0.013789152145385742, 0.013715456008911133, 0.013648896217346192, 0.014081024169921874, 0.013903871536254882, 0.014090271949768066, 0.014074848175048829, 0.014058496475219727, 0.014044159889221192, 0.014043135643005371, 0.013690879821777344, 0.01387929630279541, 0.013808639526367187, 0.013706303596496583, 0.014089152336120606, 0.0335206413269043, 0.013766655921936035, 0.013659168243408203, 0.013686783790588379, 0.013683679580688477, 0.01366528034210205, 0.013676544189453126, 0.013676544189453126, 0.013717503547668456, 0.013791232109069825, 0.013788160324096679, 0.013659135818481445, 0.013684736251831055, 0.013694975852966309, 0.0136878080368042, 0.013690879821777344, 0.013723648071289063, 0.01363865566253662, 0.013783040046691895, 0.013745152473449707, 0.013737983703613281, 0.013697024345397948, 0.013714431762695312, 0.013633536338806153, 0.013739007949829102, 0.013918208122253419, 0.014413824081420898, 0.014552063941955566, 0.014136320114135742, 0.013760512351989745, 0.013676544189453126, 0.013728768348693847, 0.013740032196044923, 0.013656064033508301, 0.01368064022064209, 0.013684736251831055, 0.013652992248535157, 0.013858880043029785, 0.01379526424407959, 0.013707327842712402, 0.013714367866516114, 0.013659135818481445, 0.013700096130371094, 0.013624320030212403, 0.013650976181030273, 0.013703136444091797, 0.013822976112365723, 0.013744128227233888, 0.013708288192749024, 0.01371446418762207, 0.013699040412902833, 0.013732864379882812, 0.013707263946533203, 0.0136878080368042, 0.013726719856262207, 0.01366528034210205, 0.01367142391204834, 0.013636608123779297, 0.013675519943237305, 0.013666303634643554, 0.013652992248535157, 0.013692928314208984, 0.013647871971130371, 0.03345612716674805, 0.013709312438964843, 0.013775872230529786, 0.013690879821777344, 0.013774911880493164, 0.014145471572875977, 0.013685759544372558, 0.013737983703613281, 0.01369600009918213, 0.013739007949829102, 0.013689855575561523, 0.013719552040100098, 0.013703231811523437, 0.013726655960083007, 0.013789183616638183, 0.013771776199340821, 0.013840383529663085, 0.013757439613342285, 0.013701120376586913, 0.013865983963012696, 0.013702143669128418, 0.013719552040100098, 0.01370419216156006, 0.01376972770690918, 0.013708288192749024, 0.013727744102478028, 0.013745152473449707, 0.013743103981018067, 0.013918208122253419, 0.013974528312683105, 0.013905920028686524, 0.013703167915344238, 0.013731840133666993, 0.013740032196044923, 0.013694975852966309, 0.01366220760345459, 0.013712384223937989, 0.013699071884155273, 0.013748224258422852, 0.013728768348693847, 0.013708288192749024, 0.013955072402954101, 0.0140830717086792, 0.014095359802246094, 0.013832192420959472, 0.01370419216156006, 0.013508607864379883, 0.013549568176269532, 0.013594623565673827, 0.01363049602508545, 0.01372771167755127, 0.013685759544372558, 0.013683712005615235, 0.01367961597442627, 0.013690879821777344, 0.013745152473449707, 0.013546496391296388, 0.01363046360015869, 0.013728768348693847, 0.01364684772491455, 0.013747200012207032, 0.013872127532958984, 0.013808639526367187, 0.0334284782409668, 0.013770751953125, 0.013712384223937989, 0.01366528034210205, 0.013683712005615235, 0.013699071884155273, 0.013598719596862792, 0.013672479629516602, 0.013670368194580078, 0.013652992248535157, 0.01365401554107666, 0.013690912246704101, 0.013695967674255372, 0.014107647895812989, 0.01366528034210205, 0.013645824432373046, 0.013660160064697266, 0.01365401554107666, 0.01367347240447998, 0.01367347240447998, 0.013686847686767579, 0.013681599617004395, 0.013612031936645508, 0.013664256095886231, 0.013616127967834473, 0.013724672317504882, 0.013656064033508301, 0.013698047637939453, 0.013709312438964843, 0.013664256095886231, 0.013620223999023438, 0.013633536338806153, 0.013624320030212403, 0.01366528034210205, 0.013612031936645508, 0.0136878080368042, 0.01367347240447998, 0.013660223960876464, 0.013722559928894044, 0.01368166446685791, 0.013710335731506347, 0.014254079818725587, 0.01470361614227295, 0.014113792419433594, 0.014170111656188965, 0.013760512351989745, 0.013751296043395997, 0.0136878080368042, 0.013628416061401367, 0.013810688018798829, 0.014081024169921874, 0.014034943580627441, 0.01406873607635498, 0.013875200271606445, 0.01369600009918213, 0.01365401554107666, 0.013806591987609864, 0.013683712005615235, 0.0136878080368042, 0.01367347240447998, 0.013592576026916504, 0.013706239700317382, 0.013677568435668945, 0.03340595245361328, 0.013661184310913087, 0.013701120376586913, 0.013631487846374512, 0.0136878080368042, 0.013606911659240722, 0.013637632369995116, 0.013661184310913087, 0.01367142391204834, 0.01367142391204834, 0.013647871971130371, 0.013600768089294434, 0.01355673599243164, 0.01347379207611084, 0.01358233642578125, 0.01370419216156006, 0.013742079734802246, 0.013740032196044923, 0.013698047637939453, 0.013651968002319336, 0.014049280166625976, 0.01355065631866455, 0.013661120414733887, 0.013639679908752441, 0.013692928314208984, 0.013642751693725585, 0.013606911659240722, 0.013629440307617188, 0.013515775680541992, 0.013544447898864746, 0.01345638370513916, 0.013495295524597169, 0.013492223739624023, 0.013442048072814941, 0.01349836826324463, 0.013719552040100098, 0.013659135818481445, 0.013678591728210449, 0.013650943756103515, 0.013698047637939453, 0.013800512313842774, 0.013656000137329101, 0.013660160064697266, 0.01368883228302002, 0.013708288192749024, 0.013617152214050294, 0.01365401554107666, 0.013684736251831055, 0.013676544189453126, 0.013694975852966309, 0.014039039611816406, 0.013917183876037598, 0.013542400360107423, 0.013486080169677735, 0.013524991989135742, 0.013504511833190918, 0.013504511833190918, 0.013707263946533203, 0.013759519577026367, 0.013673439979553222, 0.013633536338806153, 0.013720576286315917, 0.013669376373291015, 0.03328409576416016, 0.013742079734802246, 0.013636608123779297, 0.01365401554107666, 0.013612031936645508, 0.013617152214050294, 0.013622271537780761, 0.013641728401184081, 0.013709312438964843, 0.013705216407775878, 0.013713408470153808, 0.01370419216156006, 0.013658111572265624, 0.013648896217346192, 0.013650943756103515, 0.013798399925231934, 0.013830143928527832, 0.013706239700317382, 0.013674495697021484, 0.013669376373291015, 0.013651968002319336, 0.013650943756103515, 0.013708352088928222, 0.01367852783203125, 0.013643775939941406, 0.013623295783996582, 0.013626399993896485, 0.013650912284851074, 0.013658111572265624, 0.01362230396270752, 0.013684703826904296, 0.014013440132141113, 0.013792256355285644, 0.013689855575561523, 0.013732895851135254, 0.013642720222473145, 0.013642751693725585, 0.01368166446685791, 0.01364684772491455, 0.013664256095886231, 0.01366323184967041, 0.013648896217346192, 0.013670399665832519, 0.013750271797180176, 0.013668352127075196, 0.01368166446685791, 0.013776896476745605, 0.013752320289611816, 0.01363046360015869, 0.01368166446685791, 0.013661184310913087, 0.01360588836669922, 0.013745152473449707, 0.013683712005615235, 0.013625344276428223, 0.013642751693725585, 0.013661184310913087, 0.01366528034210205, 0.01376153564453125, 0.013698047637939453, 0.013649920463562011, 0.01366220760345459, 0.013780991554260253, 0.03333631896972656, 0.013669376373291015, 0.013714431762695312, 0.01365401554107666, 0.014097408294677734, 0.013722623825073242, 0.01368166446685791, 0.013719552040100098, 0.013689855575561523, 0.013758463859558106, 0.013727744102478028, 0.013700096130371094, 0.01367961597442627, 0.013809663772583008, 0.014160896301269531, 0.014683135986328125, 0.014722047805786133, 0.0147640323638916, 0.016072736740112305, 0.014328800201416015, 0.014173184394836426, 0.014113792419433594, 0.013905920028686524, 0.014019583702087402, 0.014217215538024902, 0.014188544273376465, 0.01367142391204834, 0.013690879821777344, 0.01374617576599121, 0.013760512351989745, 0.013669376373291015, 0.013750271797180176, 0.014240768432617188, 0.013876223564147949, 0.013725695610046386, 0.013740032196044923, 0.013723648071289063, 0.013727744102478028, 0.014219264030456542, 0.013730815887451172, 0.013692928314208984, 0.013648896217346192, 0.01377996826171875, 0.013807616233825683, 0.013730815887451172, 0.01377791976928711, 0.01376460838317871, 0.013784095764160157, 0.013779935836791992, 0.01440665626525879, 0.013751296043395997, 0.013748224258422852, 0.013720576286315917, 0.013757439613342285, 0.013796352386474609, 0.01368883228302002, 0.013740032196044923, 0.013669376373291015, 0.013731840133666993, 0.014028800010681153, 0.014090239524841308, 0.014048255920410157, 0.014097408294677734, 0.03338240051269531, 0.013707263946533203, 0.013972479820251465, 0.01415167999267578, 0.013827072143554688, 0.01368166446685791, 0.013698047637939453, 0.013775872230529786, 0.013874176025390626, 0.014022656440734863, 0.01420083236694336, 0.014007295608520508, 0.014128128051757812, 0.013929471969604493, 0.013670399665832519, 0.013521920204162598, 0.013543423652648925, 0.013642751693725585, 0.013641728401184081, 0.013710335731506347, 0.013683712005615235, 0.013677568435668945, 0.013709312438964843, 0.0136878080368042, 0.013805567741394043, 0.014106623649597168, 0.01389568042755127, 0.013809663772583008, 0.01375641632080078, 0.013702143669128418, 0.013699071884155273, 0.013721599578857421, 0.014123007774353028, 0.014116864204406738, 0.01384447956085205, 0.013700096130371094, 0.013693951606750488, 0.01375436782836914, 0.013629440307617188, 0.014020607948303223, 0.013959168434143066, 0.013726719856262207, 0.013694975852966309, 0.01366323184967041, 0.01368883228302002, 0.01366323184967041, 0.014238719940185546, 0.013793279647827148, 0.013757439613342285, 0.013712384223937989, 0.013727744102478028, 0.013650943756103515, 0.013865983963012696, 0.013740032196044923, 0.013732864379882812, 0.013702143669128418, 0.01367961597442627, 0.013657088279724122, 0.013635583877563476, 0.013690879821777344, 0.013837311744689941, 0.013721599578857421, 0.013693951606750488, 0.03337830352783203, 0.013720576286315917, 0.01368166446685791, 0.013658111572265624, 0.013743167877197265, 0.013634495735168457, 0.013697024345397948, 0.01367961597442627, 0.01376255989074707, 0.013670399665832519, 0.013710335731506347, 0.013744128227233888, 0.013669376373291015, 0.01365401554107666, 0.013837311744689941, 0.013932543754577637, 0.013684736251831055, 0.01368883228302002, 0.013750271797180176, 0.013712384223937989, 0.013685759544372558, 0.01363865566253662, 0.013726783752441405, 0.0136693115234375, 0.013649920463562011, 0.013725695610046386, 0.013695008277893066, 0.013632479667663575, 0.013660160064697266, 0.013643775939941406, 0.013670399665832519, 0.01366323184967041, 0.013707263946533203, 0.013705216407775878, 0.013727744102478028, 0.01365401554107666, 0.013706239700317382, 0.013730815887451172, 0.013708288192749024, 0.0136878080368042, 0.013906944274902343, 0.013698047637939453, 0.013700096130371094, 0.013693951606750488, 0.013706239700317382, 0.013700096130371094, 0.013719552040100098, 0.013659135818481445, 0.013694975852966309, 0.01366528034210205, 0.01365503978729248, 0.013674495697021484, 0.013661184310913087, 0.014050304412841797, 0.01374617576599121, 0.01375641632080078, 0.013717503547668456, 0.013738016128540038, 0.013778911590576173, 0.013645855903625489, 0.013735903739929198, 0.013664256095886231, 0.013924351692199707, 0.03334348678588867, 0.013831168174743653, 0.013744128227233888, 0.013660160064697266, 0.013733887672424316, 0.013707263946533203, 0.013643775939941406, 0.013690879821777344, 0.013637632369995116, 0.013721599578857421, 0.013669376373291015, 0.013670399665832519, 0.013648896217346192, 0.013652992248535157, 0.013700096130371094, 0.013663264274597168, 0.013681632041931152, 0.013628416061401367, 0.013652992248535157, 0.01365503978729248, 0.013716480255126954, 0.013822976112365723, 0.013707263946533203, 0.013736960411071777, 0.013656064033508301, 0.01366220760345459, 0.013615103721618652, 0.013661184310913087, 0.01369600009918213, 0.013676544189453126, 0.013649951934814454, 0.013644767761230468, 0.013689855575561523, 0.013652000427246094, 0.013703136444091797, 0.013702143669128418, 0.01363046360015869, 0.013679648399353027, 0.013659104347229004, 0.013622271537780761, 0.013700096130371094, 0.01358131217956543, 0.013882368087768555, 0.013690912246704101, 0.013693984031677247, 0.013658047676086426, 0.013618176460266113, 0.013841407775878906, 0.01364684772491455, 0.013747200012207032, 0.01366323184967041, 0.013660191535949707, 0.013664223670959473, 0.013670399665832519, 0.01367142391204834, 0.013679679870605468, 0.013713343620300293, 0.013647871971130371, 0.013808639526367187, 0.013988863945007325, 0.014156800270080566, 0.014051327705383301, 0.013918208122253419, 0.03378483200073242, 0.013750304222106933, 0.013745120048522949, 0.013670399665832519, 0.013721599578857421, 0.01365401554107666, 0.013743103981018067, 0.013668352127075196, 0.01367347240447998, 0.01364684772491455, 0.013676544189453126, 0.013615103721618652, 0.01476095962524414, 0.01416806411743164, 0.014095359802246094, 0.014045184135437011, 0.013712384223937989, 0.01367347240447998, 0.013834239959716797, 0.013687840461730957, 0.013644767761230468, 0.01367961597442627, 0.013691904067993164, 0.013634559631347656, 0.013689855575561523, 0.013716480255126954, 0.013735936164855958, 0.013625344276428223, 0.013737983703613281, 0.013707263946533203, 0.013727744102478028, 0.013645824432373046, 0.013661184310913087, 0.013726719856262207, 0.013701120376586913, 0.013758463859558106, 0.013808639526367187, 0.013659135818481445, 0.013622271537780761, 0.013657088279724122, 0.013675519943237305, 0.013639679908752441, 0.013660160064697266, 0.013827072143554688, 0.013709343910217286, 0.013709280014038086, 0.01366220760345459, 0.013656064033508301, 0.013691904067993164, 0.013652992248535157, 0.013741056442260742, 0.013643775939941406, 0.013719552040100098, 0.013644800186157227, 0.013651968002319336, 0.013933568000793458, 0.013889535903930664, 0.014051327705383301, 0.013793279647827148, 0.01367347240447998, 0.013793279647827148, 0.013675519943237305, 0.0136561279296875, 0.03336083221435547, 0.013716480255126954, 0.01394380760192871, 0.013728768348693847, 0.013698047637939453, 0.013731840133666993, 0.013748224258422852, 0.014206975936889648, 0.013709312438964843, 0.013652992248535157, 0.013666303634643554, 0.01364896011352539, 0.01367852783203125, 0.014060544013977052, 0.013894656181335448, 0.01406668758392334, 0.013749312400817872, 0.014020544052124023, 0.013848575592041015, 0.013702143669128418, 0.013800448417663574, 0.013694975852966309, 0.013694975852966309, 0.013908991813659668, 0.013658143997192384, 0.013685728073120117, 0.013685759544372558, 0.01364684772491455, 0.013701120376586913, 0.013712384223937989, 0.013736960411071777, 0.0136878080368042, 0.013711359977722168, 0.01366220760345459, 0.013625344276428223, 0.013731840133666993, 0.013642751693725585, 0.013723648071289063, 0.013726719856262207, 0.013741056442260742, 0.013677568435668945, 0.013652031898498534, 0.013719488143920898, 0.013775872230529786, 0.01387929630279541, 0.013709312438964843, 0.013881343841552735, 0.013727744102478028, 0.013524991989135742, 0.01358028793334961, 0.013520895957946777, 0.013627391815185547, 0.013666303634643554, 0.013656064033508301, 0.013658111572265624, 0.01365503978729248, 0.013708288192749024, 0.01368883228302002, 0.01355673599243164, 0.013479935646057128, 0.01348198413848877, 0.013500415802001953, 0.013453311920166015]",tokens/s,71.0519385361604,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,mistralai/Mistral-7B-v0.1,mistralai/Mistral-7B-v0.1,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.1,,,,MB,1289.433088,15987.113984,0.0,15340.66688,15114.865664,s,10,0.8963494186401368,0.08963494186401369,0.002654900259774695,0.08876291275024414,0.09013825759887695,0.09383840827941893,0.09679852882385254,"[0.09753855895996094, 0.08931600189208984, 0.08908214569091796, 0.08915328216552734, 0.08856617736816406, 0.0882488021850586, 0.08849651336669923, 0.08842211151123047, 0.08864672088623046, 0.08887910461425781]",tokens/s,2856.0290738892973,kWh,1.0485136945573514e-06,5.745374322362142e-07,6.507330672920453e-06,8.130381799714018e-06,tokens/kWh,31486836.20355992,MB,1289.433088,15987.113984,0.0,15340.66688,15134.114304,s,10,31.359611328125002,3.1359611328125,0.006538257602936493,3.13266650390625,3.14416083984375,3.144967016601562,3.1456119580078123,"[3.143981689453125, 3.133189697265625, 3.141480224609375, 3.14351806640625, 3.145773193359375, 3.1281259765625, 3.132143310546875, 3.13075732421875, 3.131896484375, 3.128745361328125]",tokens/s,20.089534701438783,kWh,3.688938523933013e-05,2.021716879712013e-05,0.00013762656241187918,0.0001947331164483294,tokens/kWh,323519.70301218104,,s,629,32.10659848785398,0.05104387676924325,0.010630762351609589,0.04962918472290039,0.05035558319091797,0.05054604721069336,0.1386133514404297,"[0.05176627349853516, 0.05060812759399414, 0.05173350524902344, 0.0510300178527832, 0.05065523147583008, 0.05116825485229492, 0.05026201629638672, 0.05043302536010742, 0.05133107376098633, 0.05058560180664062, 0.05035212707519531, 0.05014323043823242, 0.04956979370117188, 0.04954521560668945, 0.04950630569458008, 0.04941619110107422, 0.049560577392578124, 0.049912830352783204, 0.0500398063659668, 0.04959027099609375, 0.04955340957641602, 0.04945612716674805, 0.0495728645324707, 0.04961180877685547, 0.049627105712890626, 0.04956159973144531, 0.04952985763549805, 0.04955340957641602, 0.04956159973144531, 0.04952473449707031, 0.049410049438476565, 0.049593345642089844, 0.049713153839111325, 0.049598464965820314, 0.049565696716308595, 0.04955955123901367, 0.04956265640258789, 0.04957283020019531, 0.049576961517333984, 0.04948787307739258, 0.04947353744506836, 0.04943974304199219, 0.04945100784301758, 0.0495206413269043, 0.049522689819335934, 0.049530879974365234, 0.04960563278198242, 0.049448959350585936, 0.04958720016479492, 0.04955750274658203, 0.049565696716308595, 0.049546241760253906, 0.049514495849609375, 0.04951039886474609, 0.04955955123901367, 0.04949708938598633, 0.049448959350585936, 0.05102796936035156, 0.05060300827026367, 0.050165760040283204, 0.049696769714355465, 0.050162689208984375, 0.13955279541015625, 0.04955337524414063, 0.04958924865722656, 0.04955033493041992, 0.05021388626098633, 0.05041459274291992, 0.05030809783935547, 0.04988927841186523, 0.049460224151611325, 0.04950630569458008, 0.04966912078857422, 0.04952985763549805, 0.049442817687988284, 0.05044224166870117, 0.05031219100952149, 0.049585151672363284, 0.049604606628417966, 0.04972748947143555, 0.0500469741821289, 0.049775615692138675, 0.049584129333496096, 0.049601535797119144, 0.04978176116943359, 0.049568767547607424, 0.04956159973144531, 0.049565696716308595, 0.04966912078857422, 0.049770496368408204, 0.049337345123291014, 0.0499989128112793, 0.04930553436279297, 0.049509376525878904, 0.049588222503662106, 0.049691646575927735, 0.04963942337036133, 0.04956979370117188, 0.049642494201660156, 0.04973567962646484, 0.04995276641845703, 0.050601982116699216, 0.05022617721557617, 0.049620990753173826, 0.04945107269287109, 0.04966598510742187, 0.04956671905517578, 0.04962406539916992, 0.049442817687988284, 0.04979507064819336, 0.04965273666381836, 0.0496363525390625, 0.049665088653564456, 0.049749950408935546, 0.049751041412353515, 0.049588222503662106, 0.049549312591552735, 0.04959231948852539, 0.04971110534667969, 0.04952883148193359, 0.049757183074951174, 0.049860607147216796, 0.04971724700927734, 0.04971212768554688, 0.04980633544921875, 0.13930291748046875, 0.04950732803344727, 0.05038489532470703, 0.05042892837524414, 0.050418689727783204, 0.050427902221679685, 0.05011763381958008, 0.05026816177368164, 0.0503818244934082, 0.05037055969238281, 0.05037055969238281, 0.050536449432373044, 0.050016254425048826, 0.05027635192871094, 0.050121726989746096, 0.04980326461791992, 0.049552383422851565, 0.04955340957641602, 0.049514495849609375, 0.04953702545166016, 0.04959641647338867, 0.04962918472290039, 0.050098175048828124, 0.04995379257202148, 0.04961177444458008, 0.04957593536376953, 0.04957081604003906, 0.049604606628417966, 0.04962713623046875, 0.04953497695922852, 0.049478656768798826, 0.04953497695922852, 0.049568767547607424, 0.04961280059814453, 0.0497786865234375, 0.05025075149536133, 0.049772544860839846, 0.04972851181030274, 0.04970086288452148, 0.049598464965820314, 0.04959743881225586, 0.04958924865722656, 0.049721343994140625, 0.04957081604003906, 0.04943974304199219, 0.04937625503540039, 0.04947455978393555, 0.049552383422851565, 0.04962201690673828, 0.04952883148193359, 0.049576961517333984, 0.049667072296142575, 0.04962815856933594, 0.0496445426940918, 0.049788928985595705, 0.050032638549804685, 0.05043609619140625, 0.04997017669677734, 0.04985958480834961, 0.050411518096923826, 0.05037772750854492, 0.05033574295043945, 0.05064089584350586, 0.14009855651855468, 0.05035724639892578, 0.050334720611572265, 0.05036544036865234, 0.05063372802734375, 0.05133414459228516, 0.05056204986572266, 0.05053235244750977, 0.05031731033325195, 0.050743297576904295, 0.050492416381835936, 0.05053440093994141, 0.050252799987792966, 0.04953804779052735, 0.04966502380371094, 0.049478656768798826, 0.04935987091064453, 0.049667072296142575, 0.04947561645507813, 0.04948579025268555, 0.04950630569458008, 0.04948582458496094, 0.04941516876220703, 0.049667072296142575, 0.0494837760925293, 0.04958720016479492, 0.049704959869384766, 0.04980326461791992, 0.05031321716308594, 0.050321407318115234, 0.04961587142944336, 0.04953804779052735, 0.049560577392578124, 0.04953500747680664, 0.049557472229003904, 0.04969881439208984, 0.04961996841430664, 0.04956979370117188, 0.05037062454223633, 0.0496956787109375, 0.04966195297241211, 0.0498073616027832, 0.04941107177734375, 0.049598464965820314, 0.05002547073364258, 0.04978585433959961, 0.04953702545166016, 0.049865726470947266, 0.049718273162841796, 0.049838081359863284, 0.050405376434326174, 0.05039718246459961, 0.04984524917602539, 0.049503231048583986, 0.0495728645324707, 0.0494837760925293, 0.05007462310791016, 0.050405376434326174, 0.04990771102905273, 0.04962713623046875, 0.04973056030273437, 0.04966195297241211, 0.049620990753173826, 0.14034739685058595, 0.05061228942871094, 0.05054355239868164, 0.04983910369873047, 0.04953190231323242, 0.051095550537109374, 0.05050777435302734, 0.050344959259033206, 0.0504268798828125, 0.050392063140869144, 0.05050470352172851, 0.05045043182373047, 0.05006028747558594, 0.04996505737304688, 0.04969779205322265, 0.050267135620117184, 0.049759231567382815, 0.05034086227416992, 0.04962201690673828, 0.05132287979125977, 0.05077811050415039, 0.05013913726806641, 0.049635326385498044, 0.04962406539916992, 0.049805313110351565, 0.05001728057861328, 0.049670143127441405, 0.04955750274658203, 0.04946739196777344, 0.04950425720214844, 0.04961996841430664, 0.04953395080566406, 0.049547264099121094, 0.04956774520874024, 0.04963840103149414, 0.049955841064453124, 0.049634304046630856, 0.04952678298950195, 0.04958924865722656, 0.049562625885009766, 0.050016254425048826, 0.049585151672363284, 0.04969062423706055, 0.04964556884765625, 0.04955648040771484, 0.04984627151489258, 0.04954111862182617, 0.050195457458496094, 0.04973158264160156, 0.049552417755126955, 0.050172897338867185, 0.04964556884765625, 0.049617919921875, 0.05019647979736328, 0.049944576263427735, 0.05037055969238281, 0.049742847442626956, 0.04965785598754883, 0.04961075210571289, 0.04952371215820312, 0.049530879974365234, 0.04948070526123047, 0.049516544342041016, 0.1384806365966797, 0.04946636962890625, 0.0495022087097168, 0.049582080841064455, 0.049495040893554686, 0.04942745590209961, 0.04957900619506836, 0.04948889541625977, 0.04944179153442383, 0.049519615173339845, 0.05005209732055664, 0.05029171371459961, 0.049696769714355465, 0.049667072296142575, 0.04953804779052735, 0.049821697235107425, 0.05025689697265625, 0.05009920120239258, 0.0495728645324707, 0.049347583770751956, 0.049903617858886716, 0.050032638549804685, 0.04965785598754883, 0.049495040893554686, 0.04966912078857422, 0.04957388687133789, 0.04955855941772461, 0.04934140777587891, 0.04946124649047851, 0.04945612716674805, 0.04948070526123047, 0.04962713623046875, 0.049468448638916016, 0.04953699111938477, 0.049516544342041016, 0.049486846923828126, 0.04951244735717773, 0.04953497695922852, 0.049522689819335934, 0.04966502380371094, 0.04976537704467773, 0.04975513458251953, 0.0496097297668457, 0.04966604614257813, 0.04972851181030274, 0.04957900619506836, 0.04949606323242187, 0.04954111862182617, 0.049495040893554686, 0.04967116928100586, 0.04949606323242187, 0.04952883148193359, 0.04943667221069336, 0.049527809143066405, 0.049549312591552735, 0.04962918472290039, 0.049616897583007816, 0.04955136108398438, 0.049598464965820314, 0.04965683364868164, 0.049914878845214845, 0.050547710418701174, 0.0501923828125, 0.1386649627685547, 0.04951859283447266, 0.04958720016479492, 0.04959641647338867, 0.049683456420898435, 0.04953395080566406, 0.04952678298950195, 0.04991692733764649, 0.04969369506835938, 0.049598464965820314, 0.0496629753112793, 0.05059379196166992, 0.05047808074951172, 0.0496363525390625, 0.04962303924560547, 0.049696769714355465, 0.0500316162109375, 0.0495728645324707, 0.04954316711425781, 0.04957900619506836, 0.049479679107666014, 0.04951859283447266, 0.04937830352783203, 0.04956979370117188, 0.049631233215332034, 0.049719295501708984, 0.05034700775146484, 0.05031935882568359, 0.04985036849975586, 0.0501473274230957, 0.05020364761352539, 0.04962918472290039, 0.049634304046630856, 0.04963328170776367, 0.049527809143066405, 0.049734657287597656, 0.049904640197753904, 0.050233345031738284, 0.049977344512939455, 0.04960665512084961, 0.04992102432250976, 0.04967116928100586, 0.049598464965820314, 0.04958720016479492, 0.04961280059814453, 0.04984729766845703, 0.04958720016479492, 0.04956467056274414, 0.04951244735717773, 0.04949094390869141, 0.0495994873046875, 0.04959743881225586, 0.049516544342041016, 0.049604606628417966, 0.04959539031982422, 0.049547264099121094, 0.04954214477539062, 0.0495810546875, 0.04963020706176758, 0.04962406539916992, 0.04962201690673828, 0.04954316711425781, 0.04952985763549805, 0.1384315185546875, 0.04950729751586914, 0.04957081604003906, 0.049511425018310545, 0.049547264099121094, 0.04958924865722656, 0.049955841064453124, 0.050337791442871094, 0.04959539031982422, 0.04947251129150391, 0.04956671905517578, 0.049658878326416016, 0.05014835357666016, 0.05004185485839844, 0.04967424011230469, 0.049650688171386716, 0.04959641647338867, 0.05004496002197266, 0.049627105712890626, 0.04964966583251953, 0.0497367057800293, 0.04956467056274414, 0.04955852890014648, 0.04960870361328125, 0.049448959350585936, 0.04947455978393555, 0.05081497573852539, 0.05029785537719727, 0.04954118347167969, 0.04969875335693359, 0.049568767547607424, 0.04948992156982422, 0.05003878402709961, 0.04977971267700195, 0.04958617782592773, 0.04957388687133789, 0.04964556884765625, 0.049634304046630856, 0.04957798385620117, 0.04962406539916992, 0.04977766418457031, 0.04961894226074219, 0.0495994873046875, 0.049576961517333984, 0.04956159973144531, 0.04971212768554688, 0.04956774520874024, 0.04968960189819336, 0.049600513458251956, 0.049658878326416016, 0.049527809143066405, 0.04960768127441406, 0.04967222213745117, 0.04973871994018555, 0.0495810546875, 0.04955033493041992, 0.04959027099609375, 0.049616897583007816, 0.049565696716308595, 0.04960259246826172, 0.04988924789428711, 0.04965785598754883, 0.04977664184570312, 0.1387509765625, 0.04978278350830078, 0.05017497634887695, 0.04994047927856445, 0.049650688171386716, 0.049732608795166014, 0.04998758316040039, 0.04980428695678711, 0.04976947021484375, 0.04964863967895508, 0.04976332855224609, 0.04972851181030274, 0.04952883148193359, 0.04971212768554688, 0.04970700836181641, 0.04964044952392578, 0.049704959869384766, 0.04968447875976562, 0.04958003234863281, 0.049535999298095705, 0.049616897583007816, 0.04967935943603516, 0.04959641647338867, 0.04958617782592773, 0.049379390716552736, 0.049415103912353514, 0.0496517105102539, 0.049740798950195314, 0.049788928985595705, 0.04975513458251953, 0.04962815856933594, 0.04978176116943359, 0.04962508773803711, 0.0497367057800293, 0.04992716979980469, 0.04976537704467773, 0.04985760116577148, 0.049624000549316406, 0.049503231048583986, 0.04947558212280274, 0.04990054321289063, 0.049651744842529294, 0.04961180877685547, 0.04998854446411133, 0.04954729461669922, 0.04976022338867187, 0.04952985763549805, 0.04959743881225586, 0.05013401412963867, 0.04964044952392578, 0.0498790397644043, 0.049600513458251956, 0.049788928985595705, 0.04956467056274414, 0.04955750274658203, 0.049718273162841796, 0.049775615692138675, 0.04999270248413086, 0.04965580749511719, 0.049584129333496096, 0.04969472122192383, 0.04979916763305664, 0.049675262451171875, 0.13942681884765626, 0.049716224670410154, 0.04972032165527344, 0.04984012985229492, 0.0497039680480957, 0.04961891174316406, 0.04955852890014648, 0.04950835037231445, 0.04962406539916992, 0.0496517105102539, 0.0496363525390625, 0.049708030700683595, 0.049582080841064455, 0.049721343994140625, 0.04949094390869141, 0.04954214477539062, 0.049514495849609375, 0.04958924865722656, 0.05028252792358399, 0.05035516738891602, 0.04951859283447266, 0.04948582458496094, 0.049492992401123044, 0.04997119903564453, 0.049775615692138675, 0.04958003234863281, 0.04955750274658203, 0.049500160217285157, 0.049498111724853515, 0.04959027099609375, 0.04976128005981445, 0.049588222503662106, 0.04953702545166016, 0.04955340957641602, 0.049452030181884765, 0.04951039886474609, 0.049511425018310545, 0.04958003234863281, 0.04952473449707031, 0.049734657287597656, 0.049635326385498044, 0.049601535797119144, 0.049842174530029294, 0.04967116928100586, 0.04973567962646484, 0.049563648223876954, 0.049653793334960936, 0.04958819198608398, 0.04954009628295898, 0.04952473449707031, 0.04955446243286133, 0.04974998474121094, 0.04962508773803711, 0.04956979370117188, 0.04967731094360352, 0.0496855354309082, 0.05006742477416992, 0.049721343994140625, 0.04965990447998047, 0.049650688171386716, 0.049675262451171875, 0.04963328170776367, 0.049506336212158206]",tokens/s,19.5909884455045,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,google/recurrentgemma-7b,google/recurrentgemma-7b,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/google/recurrentgemma-7b/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66947056-5ef7b8911e1bd81e04545160;5e092540-4d2d-4cf8-9c51-39abf7a25706) Repository Not Found for url: https://huggingface.co/google/recurrentgemma-7b/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: google/recurrentgemma-7b is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,mistralai/Mixtral-8x7B-v0.1,mistralai/Mixtral-8x7B-v0.1,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ self.load_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 250, in load_model_with_no_weights self.pretrained_model = self.automodel_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3710, in from_pretrained model = cls(config, *model_args, **model_kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/mixtral/modeling_mixtral.py"", line 1323, in __init__ self.model = MixtralModel(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/mixtral/modeling_mixtral.py"", line 1087, in __init__ [MixtralDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/mixtral/modeling_mixtral.py"", line 1087, in [MixtralDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/mixtral/modeling_mixtral.py"", line 877, in __init__ self.block_sparse_moe = MixtralSparseMoeBlock(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/mixtral/modeling_mixtral.py"", line 824, in __init__ self.experts = nn.ModuleList([MixtralBlockSparseTop2MLP(config) for _ in range(self.num_experts)]) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/mixtral/modeling_mixtral.py"", line 824, in self.experts = nn.ModuleList([MixtralBlockSparseTop2MLP(config) for _ in range(self.num_experts)]) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/mixtral/modeling_mixtral.py"", line 791, in __init__ self.w2 = nn.Linear(self.ffn_dim, self.hidden_dim, bias=False) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/linear.py"", line 98, in __init__ self.weight = Parameter(torch.empty((out_features, in_features), **factory_kwargs)) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_device.py"", line 78, in __torch_function__ return func(*args, **kwargs) torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 112.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,t,t,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/t/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669473bc-67c1246513079b5054186974;55b27377-a1cd-4583-83fa-050e3ebec93d) Repository Not Found for url: https://huggingface.co/t/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: t is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-72B,Qwen/Qwen-72B,cuda,0,42,,,True,,,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",qwen,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-30b,facebook/opt-30b,cuda,0,42,,,True,,,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",opt,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-MoE-A2.7B,Qwen/Qwen1.5-MoE-A2.7B,cuda,0,42,,,True,,,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",qwen2_moe,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-1.3b,EleutherAI/pythia-1.3b,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1259.839488,3852.992512,0.0,3206.545408,3105.82784,s,10,0.1930521278381348,0.019305212783813475,0.0002542779813739284,0.01924176025390625,0.019335408782958984,0.019699080276489256,0.019990017471313477,"[0.02006275177001953, 0.019244255065917967, 0.019185632705688477, 0.01923993682861328, 0.01925459289550781, 0.01924358367919922, 0.019249919891357423, 0.019162208557128906, 0.019191839218139647, 0.0192174072265625]",tokens/s,13260.667098921807,kWh,2.2924400089528416e-07,1.255498719612745e-07,1.0659343152790816e-06,1.4207281881356402e-06,tokens/kWh,180189287.5342592,MB,1259.839488,3852.992512,0.0,3206.545408,3189.19168,s,10,11.02842041015625,1.102842041015625,0.019420409670015917,1.1083656005859375,1.1194994262695313,1.1242835876464843,1.128110916748047,"[1.1290677490234375, 1.0803919677734375, 1.106476806640625, 1.118436279296875, 1.115659912109375, 1.1168310546875, 1.11025439453125, 1.067032470703125, 1.078049560546875, 1.10622021484375]",tokens/s,57.12513456775939,kWh,1.29441376687585e-05,7.093055422513125e-06,3.657117912772184e-05,5.6608372218993465e-05,tokens/kWh,1112909.5844035237,,s,629,11.182288902282707,0.0177778837874129,0.0023677903324750755,0.017666048049926757,0.01791201286315918,0.018331238555908205,0.03661316253662109,"[0.018852863311767578, 0.018487295150756835, 0.018268159866333008, 0.018164735794067383, 0.01784115219116211, 0.018191360473632814, 0.01842995262145996, 0.01846784019470215, 0.017589248657226563, 0.017549312591552735, 0.01817804718017578, 0.018307071685791015, 0.01883750343322754, 0.018507776260375978, 0.018341888427734376, 0.018276351928710938, 0.018255872726440428, 0.018397184371948243, 0.018229248046875, 0.018317312240600587, 0.018367488861083983, 0.018478080749511717, 0.018318336486816408, 0.01850060844421387, 0.018241535186767577, 0.018151424407958985, 0.017863679885864257, 0.018349056243896485, 0.01829478454589844, 0.018984960556030273, 0.018490367889404297, 0.01776639938354492, 0.017897472381591797, 0.017772544860839845, 0.017763328552246094, 0.017752063751220702, 0.017862655639648437, 0.017739776611328126, 0.018132991790771484, 0.01807360076904297, 0.01779302406311035, 0.017617919921875, 0.017467391967773437, 0.017557504653930665, 0.017487871170043946, 0.017673215866088866, 0.017482751846313475, 0.018217983245849608, 0.018379776000976563, 0.018745344161987306, 0.01706188774108887, 0.016944128036499022, 0.01695542335510254, 0.016973791122436525, 0.01691756820678711, 0.016920511245727538, 0.016898048400878905, 0.016874496459960937, 0.016855039596557618, 0.016759807586669923, 0.016842752456665038, 0.016917503356933594, 0.036670463562011715, 0.01681407928466797, 0.01676089668273926, 0.01675052833557129, 0.01676595115661621, 0.016715776443481444, 0.01677516746520996, 0.01680486488342285, 0.01676905632019043, 0.016758752822875977, 0.016745471954345705, 0.01675775909423828, 0.016688127517700196, 0.016766048431396483, 0.016738208770751953, 0.01676595115661621, 0.01681817626953125, 0.016710655212402344, 0.016698368072509767, 0.016945152282714843, 0.016954368591308593, 0.01685606384277344, 0.016865280151367186, 0.016928768157958983, 0.016911359786987306, 0.01704140853881836, 0.0169881591796875, 0.016911359786987306, 0.016970752716064453, 0.016894975662231446, 0.016891904830932617, 0.01683251190185547, 0.016881664276123046, 0.016900096893310547, 0.016939008712768554, 0.016884735107421875, 0.016945152282714843, 0.016904191970825197, 0.016905216217041014, 0.016911359786987306, 0.01683967971801758, 0.01683967971801758, 0.016959487915039064, 0.01699635124206543, 0.016945152282714843, 0.017855552673339842, 0.01793529510498047, 0.018121728897094725, 0.019534847259521485, 0.018174976348876954, 0.017878015518188475, 0.017827840805053712, 0.017574911117553712, 0.01777663993835449, 0.017648639678955077, 0.01744588851928711, 0.017508352279663086, 0.017743871688842772, 0.017739776611328126, 0.01779199981689453, 0.01784320068359375, 0.01760870361328125, 0.01749504089355469, 0.03660800170898437, 0.016917503356933594, 0.01680588722229004, 0.01679974365234375, 0.01674553680419922, 0.0167259521484375, 0.016698432922363282, 0.01675872039794922, 0.01738956832885742, 0.017556480407714844, 0.017483776092529296, 0.017530879974365234, 0.017614879608154298, 0.01792099189758301, 0.017894399642944335, 0.017804288864135744, 0.017912832260131836, 0.017966079711914062, 0.017904640197753906, 0.017869823455810546, 0.01777561569213867, 0.017711103439331053, 0.01779302406311035, 0.017704959869384765, 0.01781862449645996, 0.01780019187927246, 0.01782067108154297, 0.017912832260131836, 0.0178155517578125, 0.017719295501708983, 0.01780838394165039, 0.017748992919921876, 0.017771520614624024, 0.017682432174682617, 0.01745715141296387, 0.017497087478637697, 0.017803264617919923, 0.017687551498413084, 0.0174704647064209, 0.017497087478637697, 0.01742233657836914, 0.01765888023376465, 0.017712127685546874, 0.017728511810302734, 0.017725439071655275, 0.01739468765258789, 0.017538047790527343, 0.017822719573974608, 0.01743974494934082, 0.017519615173339845, 0.01739571189880371, 0.017533952713012696, 0.01774284744262695, 0.017721343994140625, 0.017812480926513673, 0.01777561569213867, 0.017708032608032227, 0.017458175659179686, 0.01744486427307129, 0.01737932777404785, 0.01744179153442383, 0.017528863906860353, 0.017618911743164063, 0.037647361755371096, 0.017708032608032227, 0.01764249610900879, 0.017723392486572266, 0.01765478324890137, 0.017484800338745117, 0.017862655639648437, 0.017795072555541993, 0.017797119140625, 0.017827840805053712, 0.017979391098022462, 0.01779814338684082, 0.01779199981689453, 0.01777663993835449, 0.017755136489868165, 0.017686527252197267, 0.017796096801757814, 0.017769472122192383, 0.017747007369995117, 0.01774995231628418, 0.017722368240356445, 0.017703935623168944, 0.017748992919921876, 0.017733631134033204, 0.017507360458374022, 0.017522655487060546, 0.017795103073120117, 0.017766368865966796, 0.017722368240356445, 0.017752063751220702, 0.017715200424194336, 0.017731584548950196, 0.017733631134033204, 0.017713151931762695, 0.017748992919921876, 0.017754112243652344, 0.017743871688842772, 0.01765171241760254, 0.017835039138793946, 0.017753055572509766, 0.017719295501708983, 0.017780799865722657, 0.017803199768066408, 0.017796096801757814, 0.017803264617919923, 0.017753087997436523, 0.017773567199707033, 0.017846271514892577, 0.01773465538024902, 0.017744895935058593, 0.017744895935058593, 0.01779302406311035, 0.017810432434082032, 0.017755136489868165, 0.01779302406311035, 0.017524736404418945, 0.017514495849609374, 0.017754112243652344, 0.017769472122192383, 0.01780121612548828, 0.017702911376953127, 0.01778278350830078, 0.017833984375, 0.03754905700683594, 0.0177838077545166, 0.017723392486572266, 0.01768550491333008, 0.01768448066711426, 0.017692703247070313, 0.01792099189758301, 0.017960960388183594, 0.017671167373657228, 0.017859615325927735, 0.017755104064941407, 0.017901567459106444, 0.017895423889160156, 0.017762304306030274, 0.018351104736328124, 0.01788313674926758, 0.017729536056518554, 0.017673215866088866, 0.017712127685546874, 0.017727487564086913, 0.01766092872619629, 0.017682432174682617, 0.017710079193115236, 0.017709056854248048, 0.01781350326538086, 0.01775103950500488, 0.017681407928466796, 0.01776742362976074, 0.017525760650634766, 0.017476608276367187, 0.017549312591552735, 0.017491968154907226, 0.01744588851928711, 0.017564672470092774, 0.017537023544311522, 0.01749504089355469, 0.017494016647338868, 0.017838079452514647, 0.01777459144592285, 0.017760255813598632, 0.01780019187927246, 0.017656831741333007, 0.01779916763305664, 0.017733631134033204, 0.017695743560791014, 0.017674240112304687, 0.017529855728149413, 0.017499135971069335, 0.01740902328491211, 0.017492992401123047, 0.01742131233215332, 0.017546239852905272, 0.017521663665771483, 0.017451007843017577, 0.017605632781982423, 0.01802342414855957, 0.01777663993835449, 0.01780633544921875, 0.017872896194458008, 0.017796096801757814, 0.017735679626464843, 0.017752063751220702, 0.017717248916625978, 0.0375470085144043, 0.01798041534423828, 0.01781452751159668, 0.01777561569213867, 0.017777664184570312, 0.017763328552246094, 0.017730560302734375, 0.017747968673706056, 0.017732608795166017, 0.017708032608032227, 0.017707008361816406, 0.01781760025024414, 0.017702911376953127, 0.01776639938354492, 0.017696767807006835, 0.01769267272949219, 0.017847295761108398, 0.017771520614624024, 0.01760051155090332, 0.017572864532470703, 0.017490943908691405, 0.01780838394165039, 0.01778483200073242, 0.017707008361816406, 0.01746329689025879, 0.01763737678527832, 0.017810432434082032, 0.017733631134033204, 0.017777664184570312, 0.01759129524230957, 0.01744179153442383, 0.017794048309326172, 0.017801279067993163, 0.017785791397094727, 0.017781759262084963, 0.017750015258789064, 0.01772755241394043, 0.01774176025390625, 0.017785856246948242, 0.01778895950317383, 0.017408000946044923, 0.017538015365600585, 0.01767628860473633, 0.017797119140625, 0.017705984115600586, 0.01760665512084961, 0.01741107177734375, 0.017911808013916015, 0.017724416732788087, 0.017768447875976562, 0.01783500862121582, 0.01784217643737793, 0.01784934425354004, 0.017740800857543947, 0.017727487564086913, 0.01748684883117676, 0.017468416213989257, 0.017688575744628905, 0.017757183074951173, 0.017673215866088866, 0.017934335708618163, 0.01785753631591797, 0.017747968673706056, 0.03746406555175781, 0.017761280059814453, 0.01762918472290039, 0.01744076728820801, 0.01757798385620117, 0.017718271255493166, 0.017743871688842772, 0.01766912078857422, 0.017725439071655275, 0.017797119140625, 0.017682432174682617, 0.01775004768371582, 0.01771107292175293, 0.017797119140625, 0.017770496368408203, 0.017787904739379884, 0.01768550491333008, 0.017687551498413084, 0.017704959869384765, 0.01779814338684082, 0.017702911376953127, 0.01779302406311035, 0.017755136489868165, 0.017667072296142578, 0.01785241508483887, 0.017755136489868165, 0.017722368240356445, 0.017752063751220702, 0.017694719314575197, 0.017682432174682617, 0.017657855987548828, 0.017819648742675782, 0.0175861759185791, 0.017703935623168944, 0.017757183074951173, 0.01765171241760254, 0.017705984115600586, 0.017765375137329103, 0.01765990447998047, 0.01757900810241699, 0.018147327423095702, 0.02009600067138672, 0.01846886444091797, 0.018339839935302735, 0.017802240371704102, 0.01784115219116211, 0.01780121612548828, 0.017709056854248048, 0.017761280059814453, 0.0177807674407959, 0.01766294479370117, 0.017320959091186524, 0.016931840896606445, 0.01679769515991211, 0.016695295333862305, 0.0168089599609375, 0.01666662406921387, 0.016684032440185546, 0.016743423461914063, 0.016713727951049806, 0.01681100845336914, 0.016916479110717773, 0.016869375228881836, 0.036615169525146485, 0.016953344345092772, 0.016892927169799805, 0.016926719665527345, 0.016941055297851563, 0.016905216217041014, 0.016889856338500975, 0.016903200149536134, 0.016926687240600587, 0.016855039596557618, 0.016750591278076172, 0.016792575836181642, 0.016792575836181642, 0.01681715202331543, 0.016776191711425782, 0.016773120880126953, 0.01679052734375, 0.01677004814147949, 0.016762943267822267, 0.01679865646362305, 0.01759231948852539, 0.017087488174438475, 0.016904191970825197, 0.016900096893310547, 0.01684787178039551, 0.016928768157958983, 0.01703424072265625, 0.01700044822692871, 0.01696463966369629, 0.016958431243896486, 0.016948223114013672, 0.016913408279418944, 0.016921600341796874, 0.016924671173095703, 0.017105920791625977, 0.01699839973449707, 0.016933887481689454, 0.01695232009887695, 0.017077247619628907, 0.017062911987304686, 0.01700454330444336, 0.016949247360229493, 0.016945152282714843, 0.016932863235473633, 0.016895999908447267, 0.01685606384277344, 0.01699942398071289, 0.017191999435424803, 0.016944063186645507, 0.016890880584716796, 0.016948223114013672, 0.016889856338500975, 0.01700556755065918, 0.016946176528930663, 0.016942079544067384, 0.016912384033203123, 0.016894975662231446, 0.016902143478393555, 0.016888832092285155, 0.016898048400878905, 0.016878591537475587, 0.016914432525634765, 0.016924671173095703, 0.036580352783203124, 0.016909311294555664, 0.016905216217041014, 0.016869375228881836, 0.016904256820678712, 0.016869312286376954, 0.01676595115661621, 0.01683251190185547, 0.016913408279418944, 0.016890880584716796, 0.01684480094909668, 0.016977920532226562, 0.016920576095581053, 0.0168724479675293, 0.016862207412719727, 0.016934911727905275, 0.016939008712768554, 0.016947200775146484, 0.017140735626220704, 0.016925695419311524, 0.016917503356933594, 0.016869375228881836, 0.016957439422607423, 0.016869375228881836, 0.016745471954345705, 0.016750591278076172, 0.01684480094909668, 0.016875520706176757, 0.016962591171264647, 0.017020896911621095, 0.0169564151763916, 0.016917503356933594, 0.016889856338500975, 0.016925695419311524, 0.016928768157958983, 0.01686835289001465, 0.016888832092285155, 0.016927743911743166, 0.016861183166503906, 0.016863231658935548, 0.016946176528930663, 0.016723968505859374, 0.01701375961303711, 0.016940031051635742, 0.016894975662231446, 0.016903167724609376, 0.017778688430786133, 0.017624063491821287, 0.01766912078857422, 0.017708032608032227, 0.01763737678527832, 0.017683456420898438, 0.0176680965423584, 0.017656831741333007, 0.01767628860473633, 0.017722368240356445, 0.017747968673706056, 0.017617919921875, 0.017492992401123047, 0.017539072036743163, 0.017688575744628905, 0.017716224670410157, 0.017570816040039062, 0.0366561279296875, 0.01699942398071289, 0.016900096893310547, 0.017897472381591797, 0.01784524726867676, 0.01762611198425293, 0.017694719314575197, 0.01736396789550781, 0.01777663993835449, 0.017726463317871095, 0.017678335189819337, 0.01775103950500488, 0.01763020706176758, 0.01762303924560547, 0.01762099266052246, 0.01760256004333496, 0.01768448066711426, 0.017715200424194336, 0.01763430404663086, 0.017666048049926757, 0.01767628860473633, 0.01765068817138672, 0.01785958480834961, 0.017679359436035155, 0.01764352035522461, 0.017698816299438477, 0.0178657283782959, 0.01766912078857422, 0.01740083122253418, 0.017321983337402345, 0.017284095764160155, 0.017510400772094727, 0.017604608535766602, 0.01763839912414551, 0.01764556884765625, 0.017546239852905272, 0.017330175399780275, 0.017311744689941407, 0.01734758377075195, 0.017346559524536134, 0.017286144256591796, 0.017346559524536134, 0.017350656509399414, 0.017344512939453126, 0.017362943649291994, 0.018486272811889647, 0.018918399810791017, 0.017944576263427735, 0.017695743560791014, 0.01777561569213867, 0.017500160217285156, 0.017252351760864256, 0.017312768936157227, 0.017314815521240236, 0.01764659118652344, 0.017624063491821287, 0.01764352035522461, 0.01770086479187012, 0.017698816299438477, 0.01765376091003418, 0.017006591796875, 0.016745471954345705, 0.016749568939208984]",tokens/s,56.24966458088899,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-1.4b,EleutherAI/pythia-1.4b,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1257.570304,3852.992512,0.0,3206.545408,3105.82784,s,10,0.1930878410339356,0.019308784103393553,0.0002982267366102465,0.019219376564025878,0.019381788444519044,0.01978655824661255,0.020110374088287353,"[0.020191328048706055, 0.019291839599609374, 0.019276735305786132, 0.01921321678161621, 0.01916364860534668, 0.01917958450317383, 0.019126815795898436, 0.019245439529418946, 0.01922553634643555, 0.019173696517944337]",tokens/s,13258.214428686244,kWh,2.2930128148513863e-07,1.255967355000605e-07,1.0614075028720838e-06,1.416305519857283e-06,tokens/kWh,180751960.93692863,MB,1257.570304,3852.992512,0.0,3206.545408,3189.19168,s,10,10.936780883789062,1.0936780883789061,0.01630772174915674,1.0941107177734375,1.1115546997070311,1.1175878967285156,1.122414454345703,"[1.12362109375, 1.1102139892578125, 1.09401953125, 1.0732847900390625, 1.08555517578125, 1.0743533935546874, 1.074810791015625, 1.1100771484375, 1.094201904296875, 1.09664306640625]",tokens/s,57.603787320436446,kWh,1.3152869895716662e-05,7.204484050677799e-06,3.6702956041928975e-05,5.706030998832343e-05,tokens/kWh,1104094.9481853857,,s,629,11.091198944091792,0.017633066683770744,0.0023944620599391686,0.017116159439086915,0.017849549102783205,0.01795563507080078,0.03670208465576172,"[0.018229215621948243, 0.018549760818481444, 0.017408000946044923, 0.017102848052978514, 0.017147903442382813, 0.017110015869140623, 0.017257471084594727, 0.018291711807250977, 0.017951744079589844, 0.017933311462402343, 0.017864704132080078, 0.01785651206970215, 0.0177838077545166, 0.017855487823486327, 0.017953792572021485, 0.017913856506347657, 0.017951744079589844, 0.01780633544921875, 0.017902591705322265, 0.01785139274597168, 0.01780121612548828, 0.017910783767700195, 0.017886207580566405, 0.01785856056213379, 0.01803980827331543, 0.017917951583862304, 0.017895423889160156, 0.017977344512939454, 0.01785036849975586, 0.017992704391479493, 0.018061311721801757, 0.018448383331298827, 0.01819340705871582, 0.018375680923461913, 0.018152448654174806, 0.017811456680297853, 0.01784934425354004, 0.01785753631591797, 0.017934335708618163, 0.017708032608032227, 0.01780121612548828, 0.017750015258789064, 0.017701887130737306, 0.017633279800415038, 0.0176363525390625, 0.017761280059814453, 0.017696767807006835, 0.017795072555541993, 0.017719327926635744, 0.017763296127319337, 0.017741823196411134, 0.017702911376953127, 0.017727487564086913, 0.01759539222717285, 0.017497087478637697, 0.017490943908691405, 0.017516544342041016, 0.017794048309326172, 0.017703935623168944, 0.017696767807006835, 0.017746944427490235, 0.01806438446044922, 0.038354942321777344, 0.017970176696777345, 0.01780735969543457, 0.017812480926513673, 0.017769472122192383, 0.017735679626464843, 0.017709056854248048, 0.01767628860473633, 0.017687551498413084, 0.017733631134033204, 0.017695743560791014, 0.01775923156738281, 0.017435647964477538, 0.01744691276550293, 0.017467391967773437, 0.017694719314575197, 0.017770496368408203, 0.01766912078857422, 0.017729536056518554, 0.017715200424194336, 0.017704959869384765, 0.01780633544921875, 0.017680383682250975, 0.01770086479187012, 0.01742131233215332, 0.017605632781982423, 0.017557504653930665, 0.017727487564086913, 0.017717248916625978, 0.017753087997436523, 0.017693695068359376, 0.01768448066711426, 0.017746944427490235, 0.017753087997436523, 0.017786880493164063, 0.01760972785949707, 0.01762303924560547, 0.01769267272949219, 0.017777664184570312, 0.017779712677001954, 0.017726463317871095, 0.017743871688842772, 0.01746329689025879, 0.017124351501464845, 0.017110015869140623, 0.017118207931518553, 0.01713363265991211, 0.01714067268371582, 0.01701478385925293, 0.0170700798034668, 0.017059839248657227, 0.017005632400512696, 0.01714681625366211, 0.017054719924926756, 0.01703628730773926, 0.01705881690979004, 0.01726470375061035, 0.017803199768066408, 0.018060287475585936, 0.018311168670654295, 0.018736127853393555, 0.017971200942993162, 0.01798041534423828, 0.03761056137084961, 0.01788307189941406, 0.017714176177978515, 0.017531904220581054, 0.017548288345336914, 0.0176312313079834, 0.017663999557495116, 0.01767628860473633, 0.0175994873046875, 0.01717452812194824, 0.017043455123901367, 0.017135616302490234, 0.017088512420654296, 0.017139711380004884, 0.017073152542114257, 0.017084415435791016, 0.01705369567871094, 0.017082368850708008, 0.01705881690979004, 0.017102848052978514, 0.017081344604492187, 0.017098751068115235, 0.017138687133789063, 0.01691961669921875, 0.01686419105529785, 0.016903167724609376, 0.01702707290649414, 0.017031167984008787, 0.017102848052978514, 0.017107967376708985, 0.017155071258544922, 0.017126399993896483, 0.017105920791625977, 0.017156095504760743, 0.017155071258544922, 0.017076223373413087, 0.017083391189575196, 0.017118207931518553, 0.017176576614379883, 0.017192960739135742, 0.017133600234985353, 0.017084384918212892, 0.01710086441040039, 0.01713657569885254, 0.017050655364990234, 0.01709667205810547, 0.017385471343994142, 0.017781759262084963, 0.017729536056518554, 0.017755136489868165, 0.017765375137329103, 0.017789951324462892, 0.017768447875976562, 0.01796403121948242, 0.017794048309326172, 0.017901567459106444, 0.017799200057983397, 0.017763296127319337, 0.01777561569213867, 0.017900543212890627, 0.01767628860473633, 0.017939456939697264, 0.01772960090637207, 0.03676153564453125, 0.01713667106628418, 0.017050592422485352, 0.017043455123901367, 0.01706188774108887, 0.017006591796875, 0.01704243278503418, 0.017047552108764647, 0.017046592712402345, 0.017060800552368163, 0.017036352157592773, 0.01705465507507324, 0.01706188774108887, 0.016924671173095703, 0.01686425590515137, 0.01685196876525879, 0.017171455383300782, 0.01706188774108887, 0.01702809524536133, 0.017079296112060546, 0.01700966453552246, 0.01701375961303711, 0.01701273536682129, 0.01703628730773926, 0.01701171112060547, 0.0170700798034668, 0.017150976181030272, 0.016921600341796874, 0.016855039596557618, 0.016867328643798828, 0.01681407928466797, 0.016857088088989256, 0.017037376403808594, 0.017106880187988283, 0.0170700798034668, 0.017093631744384767, 0.017100799560546876, 0.0170250244140625, 0.01702911949157715, 0.017063936233520507, 0.016974847793579103, 0.01700556755065918, 0.01700966453552246, 0.01702400016784668, 0.017074176788330078, 0.017004575729370117, 0.017019872665405274, 0.017072128295898437, 0.01701785659790039, 0.017126399993896483, 0.01701580810546875, 0.01701375961303711, 0.017077247619628907, 0.01706598472595215, 0.01703731155395508, 0.017039360046386717, 0.017068031311035157, 0.01700556755065918, 0.017133567810058595, 0.017047552108764647, 0.017039360046386717, 0.017073152542114257, 0.01705881690979004, 0.036703231811523435, 0.017076223373413087, 0.016969728469848632, 0.017027103424072265, 0.0170229434967041, 0.01701785659790039, 0.016950271606445313, 0.01700147247314453, 0.01704140853881836, 0.01705881690979004, 0.01702195167541504, 0.017073152542114257, 0.017082368850708008, 0.017115135192871094, 0.017113088607788086, 0.017035263061523438, 0.017077247619628907, 0.017112064361572265, 0.01705779266357422, 0.017094655990600584, 0.017035263061523438, 0.01701478385925293, 0.01707423973083496, 0.017034175872802735, 0.017117183685302736, 0.017089536666870117, 0.01701171112060547, 0.017090560913085938, 0.01699839973449707, 0.017085439682006837, 0.017087488174438475, 0.017122304916381836, 0.01704960060119629, 0.017059839248657227, 0.017161216735839844, 0.016950271606445313, 0.016908287048339844, 0.016870399475097657, 0.016880640029907225, 0.016849920272827147, 0.016917503356933594, 0.016842752456665038, 0.01682329559326172, 0.016858112335205077, 0.017116159439086915, 0.01768550491333008, 0.017736703872680663, 0.017755136489868165, 0.017727487564086913, 0.01782374382019043, 0.017672191619873046, 0.017710079193115236, 0.017710079193115236, 0.017671199798583986, 0.017700895309448243, 0.01774995231628418, 0.017737728118896484, 0.017778688430786133, 0.017732608795166017, 0.017682432174682617, 0.017697792053222656, 0.01775923156738281, 0.01770086479187012, 0.03675651168823242, 0.01709667205810547, 0.01705881690979004, 0.017040384292602538, 0.01700556755065918, 0.01703731155395508, 0.017185792922973633, 0.0170199031829834, 0.01704652786254883, 0.01700966453552246, 0.01706188774108887, 0.01704550361633301, 0.0169932804107666, 0.01703321647644043, 0.017060863494873048, 0.01698918342590332, 0.017071104049682616, 0.01701683235168457, 0.017031167984008787, 0.017125375747680666, 0.017085439682006837, 0.017022975921630858, 0.016879615783691407, 0.01684480094909668, 0.016871423721313478, 0.017044479370117188, 0.017147903442382813, 0.01702195167541504, 0.017073152542114257, 0.017062911987304686, 0.017063936233520507, 0.017052671432495118, 0.01703628730773926, 0.01702911949157715, 0.017097728729248047, 0.017052671432495118, 0.017052671432495118, 0.01702707290649414, 0.017076223373413087, 0.017068031311035157, 0.017007680892944337, 0.017015743255615234, 0.01702195167541504, 0.017054719924926756, 0.016990207672119142, 0.017092607498168946, 0.017091583251953125, 0.017130495071411133, 0.01703219223022461, 0.01704652786254883, 0.017101823806762697, 0.0169932804107666, 0.017097728729248047, 0.017079296112060546, 0.017085439682006837, 0.017064960479736328, 0.01704550361633301, 0.01700454330444336, 0.017067007064819336, 0.01703731155395508, 0.017083423614501953, 0.017076192855834962, 0.017073152542114257, 0.03666534423828125, 0.017054719924926756, 0.01701580810546875, 0.01699737548828125, 0.01701785659790039, 0.01698406410217285, 0.01701478385925293, 0.01696051216125488, 0.01700249671936035, 0.017031167984008787, 0.01699839973449707, 0.017047552108764647, 0.01698099136352539, 0.0170383358001709, 0.016995328903198242, 0.01702604866027832, 0.017060863494873048, 0.016998464584350587, 0.017014720916748046, 0.01704243278503418, 0.016991231918334963, 0.01700044822692871, 0.017119232177734374, 0.01696051216125488, 0.016877567291259766, 0.01683558464050293, 0.01683456039428711, 0.01683558464050293, 0.017042463302612304, 0.01697587203979492, 0.017061920166015626, 0.017036224365234377, 0.0169932804107666, 0.017068031311035157, 0.017142784118652343, 0.01706188774108887, 0.01700761604309082, 0.017069055557250978, 0.01700351905822754, 0.017074176788330078, 0.017135616302490234, 0.017006591796875, 0.01704140853881836, 0.017095680236816405, 0.01700454330444336, 0.01703014373779297, 0.017105920791625977, 0.0170250244140625, 0.01702911949157715, 0.017112096786499022, 0.017001440048217773, 0.017257471084594727, 0.01775103950500488, 0.017670143127441407, 0.017475584030151366, 0.017163263320922852, 0.016849920272827147, 0.016915456771850586, 0.017067007064819336, 0.017088512420654296, 0.01702400016784668, 0.017076223373413087, 0.017040384292602538, 0.03828736114501953, 0.01795686340332031, 0.017936384201049805, 0.017796096801757814, 0.017743871688842772, 0.017718271255493166, 0.01774284744262695, 0.017711103439331053, 0.017745920181274414, 0.017721343994140625, 0.017699840545654297, 0.01779199981689453, 0.01704960060119629, 0.01702911949157715, 0.01700556755065918, 0.01701273536682129, 0.017117183685302736, 0.017110015869140623, 0.017073152542114257, 0.017067007064819336, 0.017063936233520507, 0.017076223373413087, 0.01705062484741211, 0.017048576354980468, 0.01701683235168457, 0.017090560913085938, 0.017129472732543945, 0.017894399642944335, 0.01790771293640137, 0.017769472122192383, 0.017679359436035155, 0.017559551239013673, 0.017468416213989257, 0.017479679107666016, 0.017914880752563478, 0.017819648742675782, 0.01778892707824707, 0.017777664184570312, 0.017718271255493166, 0.01761587142944336, 0.017488895416259767, 0.01784217643737793, 0.017911808013916015, 0.017744895935058593, 0.017735679626464843, 0.017711103439331053, 0.017797119140625, 0.01780019187927246, 0.017739776611328126, 0.01771321678161621, 0.01774176025390625, 0.017657855987548828, 0.017821695327758787, 0.01844534492492676, 0.017991647720336915, 0.0177838077545166, 0.017703935623168944, 0.017689599990844726, 0.017736703872680663, 0.017702911376953127, 0.017743871688842772, 0.017736703872680663, 0.017960992813110352, 0.037421024322509766, 0.01785036849975586, 0.01764761543273926, 0.017670143127441407, 0.017686527252197267, 0.017675264358520508, 0.017625087738037108, 0.017694719314575197, 0.01764352035522461, 0.017839103698730468, 0.0176629753112793, 0.01775103950500488, 0.017680383682250975, 0.017765375137329103, 0.017729536056518554, 0.01787494468688965, 0.01780940818786621, 0.017709056854248048, 0.017698816299438477, 0.017839103698730468, 0.017739776611328126, 0.017663999557495116, 0.017716224670410157, 0.017731584548950196, 0.01769267272949219, 0.01781760025024414, 0.017522687911987304, 0.017731584548950196, 0.01720627212524414, 0.016941055297851563, 0.017051647186279297, 0.017103872299194335, 0.01700249671936035, 0.017087488174438475, 0.017083423614501953, 0.017078239440917967, 0.017039360046386717, 0.017112064361572265, 0.01702809524536133, 0.0170199031829834, 0.01707827186584473, 0.017031167984008787, 0.017074176788330078, 0.017120256423950195, 0.01701888084411621, 0.01705887985229492, 0.0170761604309082, 0.01701785659790039, 0.01703321647644043, 0.017093631744384767, 0.01700966453552246, 0.017006591796875, 0.01705881690979004, 0.01716531181335449, 0.017303552627563477, 0.017102848052978514, 0.01704652786254883, 0.017072160720825194, 0.01709974479675293, 0.017110015869140623, 0.017093631744384767, 0.017156095504760743, 0.01700966453552246, 0.03669913482666016, 0.017085439682006837, 0.01701375961303711, 0.01701785659790039, 0.01701683235168457, 0.01721446418762207, 0.017079296112060546, 0.017060863494873048, 0.01707526397705078, 0.01710483169555664, 0.01703219223022461, 0.017082368850708008, 0.01696767997741699, 0.01706598472595215, 0.01701068878173828, 0.01701481628417969, 0.01707209587097168, 0.01702195167541504, 0.01698406410217285, 0.016966655731201173, 0.016978944778442383, 0.01705062484741211, 0.017083391189575196, 0.0169932804107666, 0.017012767791748047, 0.01703727912902832, 0.017089536666870117, 0.017022975921630858, 0.017048576354980468, 0.017031167984008787, 0.017154048919677735, 0.01779916763305664, 0.01785651206970215, 0.017703935623168944, 0.017731584548950196, 0.01775103950500488, 0.017712127685546874, 0.017710079193115236, 0.017723392486572266, 0.01783500862121582, 0.017762304306030274, 0.017665023803710937, 0.01767628860473633, 0.017745920181274414, 0.017752063751220702, 0.017717248916625978, 0.017730560302734375, 0.017872896194458008, 0.017682432174682617, 0.017737728118896484, 0.017763328552246094, 0.017656831741333007, 0.01782476806640625, 0.0178155517578125, 0.01776639938354492, 0.017732608795166017, 0.017755136489868165, 0.017681407928466796, 0.017770496368408203, 0.017688575744628905, 0.017688575744628905, 0.017702911376953127, 0.017737728118896484]",tokens/s,56.7116326350871,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,x,x,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/x/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66932481-227ec7106cfeaef2056f0988;bf792241-f91e-4f5f-9aeb-c29b0d86c3bf) Repository Not Found for url: https://huggingface.co/x/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: x is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,01-ai/Yi-34B,01-ai/Yi-34B,cuda,0,42,,,True,,,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",llama,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/xglm-4.5B,facebook/xglm-4.5B,cuda,0,42,,,True,,,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",xglm,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/xglm-7.5B,facebook/xglm-7.5B,cuda,0,42,,,True,,,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",xglm,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-14B,Qwen/Qwen1.5-14B,cuda,0,42,,,True,,,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",qwen2,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,meta-llama/Llama-2-70b-hf,meta-llama/Llama-2-70b-hf,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ self.load_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 250, in load_model_with_no_weights self.pretrained_model = self.automodel_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3710, in from_pretrained model = cls(config, *model_args, **model_kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 1101, in __init__ self.model = LlamaModel(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 884, in __init__ [LlamaDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 884, in [LlamaDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 679, in __init__ self.mlp = LlamaMLP(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 192, in __init__ self.gate_proj = nn.Linear(self.hidden_size, self.intermediate_size, bias=config.mlp_bias) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/linear.py"", line 98, in __init__ self.weight = Parameter(torch.empty((out_features, in_features), **factory_kwargs)) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_device.py"", line 78, in __torch_function__ return func(*args, **kwargs) torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 448.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Deci/DeciCoder-1b,Deci/DeciCoder-1b,cuda,0,42,,,True,,,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",llama,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,meta-llama/Meta-Llama-3-70B,meta-llama/Meta-Llama-3-70B,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ self.load_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 250, in load_model_with_no_weights self.pretrained_model = self.automodel_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3710, in from_pretrained model = cls(config, *model_args, **model_kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 1101, in __init__ self.model = LlamaModel(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 884, in __init__ [LlamaDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 884, in [LlamaDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 679, in __init__ self.mlp = LlamaMLP(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 193, in __init__ self.up_proj = nn.Linear(self.hidden_size, self.intermediate_size, bias=config.mlp_bias) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/linear.py"", line 98, in __init__ self.weight = Parameter(torch.empty((out_features, in_features), **factory_kwargs)) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_device.py"", line 78, in __torch_function__ return func(*args, **kwargs) torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 448.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,r,r,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/r/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669473fa-683404ae216356980b7abc0c;adfc29e0-9ecf-4f29-be56-b9bea6981d19) Repository Not Found for url: https://huggingface.co/r/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: r is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,togethercomputer/RedPajama-INCITE-Base-7B-v0.1,togethercomputer/RedPajama-INCITE-Base-7B-v0.1,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1269.313536,15192.293376,0.0,14545.846272,14287.956992,s,10,0.7803749465942383,0.07803749465942382,0.0003365333754513757,0.07791838455200195,0.07815171661376953,0.07859150619506836,0.07894333786010743,"[0.0790312957763672, 0.07800313568115234, 0.07786608123779297, 0.07805398559570312, 0.07795494079589843, 0.07787420654296875, 0.07792044830322266, 0.07785648345947266, 0.07789804840087891, 0.07791632080078124]",tokens/s,3280.47435552937,kWh,9.232611757599646e-07,5.059041854963197e-07,4.9811237765623895e-06,6.410289137818674e-06,tokens/kWh,39935796.10780443,MB,1269.313536,15192.293376,0.0,14545.846272,14491.309056,s,10,21.413564453124998,2.1413564453125,0.0010744466200075123,2.14093408203125,2.142409912109375,2.1432635498046873,2.1439464599609375,"[2.1441171875, 2.141706298828125, 2.140923095703125, 2.14222021484375, 2.14094384765625, 2.140333984375, 2.14092431640625, 2.1413837890625, 2.140415771484375, 2.140595947265625]",tokens/s,29.420603999819406,kWh,2.521940104145971e-05,1.382070252872495e-05,0.00011872725296783733,0.00015776735653802202,tokens/kWh,399322.1499202655,,s,629,22.08051300048829,0.03510415421381285,0.009304286167305212,0.0339865608215332,0.03410124969482422,0.034136472320556645,0.11219831176757812,"[0.03388108825683594, 0.03379097747802735, 0.033850399017333985, 0.0338267822265625, 0.03391078567504883, 0.03383305740356445, 0.03385641479492187, 0.033841152191162106, 0.03392102432250976, 0.03395686340332031, 0.03393638229370117, 0.03385036849975586, 0.0339128303527832, 0.033982463836669925, 0.034146305084228515, 0.03455900955200195, 0.03415548706054688, 0.03393638229370117, 0.03395072174072265, 0.03390873718261719, 0.033896446228027344, 0.033931262969970705, 0.03421388626098633, 0.03390566253662109, 0.03395379257202148, 0.0340049934387207, 0.03401523208618164, 0.033980415344238284, 0.033972225189208984, 0.03395072174072265, 0.03406028747558594, 0.03395993423461914, 0.03399679946899414, 0.033942527770996093, 0.03398963165283203, 0.034044929504394535, 0.034385921478271485, 0.034685951232910156, 0.03420569610595703, 0.034028545379638675, 0.03403571319580078, 0.03401113510131836, 0.0340316162109375, 0.03397324752807617, 0.034012161254882815, 0.03402649688720703, 0.03410124969482422, 0.0340049934387207, 0.034062335968017575, 0.03406950378417969, 0.034070526123046875, 0.034086910247802735, 0.034083839416503905, 0.034067455291748046, 0.03410636901855469, 0.03407462310791016, 0.03403366470336914, 0.034081790924072264, 0.03406028747558594, 0.03407360076904297, 0.03407564926147461, 0.034138111114501955, 0.1125580825805664, 0.03385036849975586, 0.03400703811645508, 0.03381248092651367, 0.03390054321289063, 0.03387596893310547, 0.034121726989746096, 0.03399168014526367, 0.03396505737304688, 0.03389952087402344, 0.03397119903564453, 0.03382374572753906, 0.033914878845214845, 0.033857536315917966, 0.0339046401977539, 0.03384012985229492, 0.03394559860229492, 0.033844223022460936, 0.0339343376159668, 0.03390156936645508, 0.0339046401977539, 0.033896446228027344, 0.03393740844726562, 0.03390054321289063, 0.03391897583007813, 0.03396915054321289, 0.03396710586547851, 0.033941505432128906, 0.03395276641845703, 0.03393948745727539, 0.03396918487548828, 0.03393939208984375, 0.033972225189208984, 0.033976318359375, 0.03400396728515625, 0.03402649688720703, 0.033977344512939454, 0.03398963165283203, 0.03400908660888672, 0.03402137756347656, 0.03401932907104492, 0.03401728057861328, 0.03405209732055664, 0.034028545379638675, 0.03401113510131836, 0.034051071166992186, 0.03406950378417969, 0.034016254425048825, 0.034124801635742184, 0.03406028747558594, 0.034105342864990236, 0.03407564926147461, 0.034065406799316404, 0.03406335830688476, 0.03408281707763672, 0.034043968200683596, 0.03411859130859375, 0.03411251068115234, 0.03411251068115234, 0.0340684814453125, 0.03411763381958008, 0.03408793640136719, 0.034103294372558594, 0.11217100524902343, 0.033811454772949216, 0.03379404830932617, 0.033805313110351565, 0.033838081359863284, 0.03381760025024414, 0.033842174530029294, 0.03390054321289063, 0.0338524169921875, 0.03393843078613281, 0.03382070541381836, 0.03390153503417969, 0.033873920440673826, 0.033955841064453124, 0.03386470413208008, 0.033982463836669925, 0.033883167266845704, 0.033962974548339846, 0.03385548782348633, 0.033926143646240234, 0.03392921447753906, 0.03397119903564453, 0.033893375396728515, 0.033964031219482424, 0.033893375396728515, 0.033964031219482424, 0.03394559860229492, 0.03397529602050781, 0.0338974723815918, 0.0339681282043457, 0.03396710586547851, 0.03403984069824219, 0.03409507369995117, 0.03400908660888672, 0.03394559860229492, 0.034049022674560545, 0.03397119903564453, 0.03400601577758789, 0.03400089645385742, 0.03402239990234375, 0.03399679946899414, 0.03408793640136719, 0.034028545379638675, 0.03405619049072266, 0.03399270248413086, 0.0340766716003418, 0.03399884796142578, 0.034083839416503905, 0.0340582389831543, 0.034046112060546876, 0.03402428817749024, 0.0340582389831543, 0.03404390335083008, 0.034078720092773435, 0.03403776168823242, 0.034103294372558594, 0.0341401596069336, 0.03412377548217774, 0.034078784942626957, 0.03414931106567383, 0.03406041717529297, 0.034095008850097655, 0.03407049560546875, 0.1126328353881836, 0.03379097747802735, 0.03413094329833984, 0.03426611328125, 0.03392409515380859, 0.033811454772949216, 0.0339128303527832, 0.033860607147216795, 0.03386368179321289, 0.033854465484619144, 0.03385958480834961, 0.03386470413208008, 0.03387187194824219, 0.033883201599121095, 0.03390047836303711, 0.033876991271972655, 0.03392102432250976, 0.03385548782348633, 0.03387289428710937, 0.033888256072998044, 0.03390566253662109, 0.03390054321289063, 0.03390054321289063, 0.03389132690429687, 0.03393024063110352, 0.033925121307373046, 0.033977344512939454, 0.033928192138671875, 0.03397324752807617, 0.033941505432128906, 0.03400396728515625, 0.033963008880615236, 0.03400806427001953, 0.03398144149780274, 0.03401420974731445, 0.03399168014526367, 0.03402444839477539, 0.03398758316040039, 0.03400908660888672, 0.0339865608215332, 0.034028545379638675, 0.03397324752807617, 0.0340316162109375, 0.034010112762451174, 0.03403571319580078, 0.03402547073364258, 0.03408588790893555, 0.034034687042236327, 0.03403878402709961, 0.03418316650390625, 0.03463782501220703, 0.034219009399414066, 0.03407257461547852, 0.03406335830688476, 0.03404803085327148, 0.034077663421630856, 0.03407462310791016, 0.03407462310791016, 0.03409920120239258, 0.034078720092773435, 0.034103294372558594, 0.034108417510986325, 0.034118656158447266, 0.11236863708496093, 0.03385343933105469, 0.0337786865234375, 0.033849342346191406, 0.03384012985229492, 0.03395993423461914, 0.033821697235107424, 0.03391692733764649, 0.03384012985229492, 0.03386470413208008, 0.03383910369873047, 0.03392204666137695, 0.033858558654785154, 0.03392409515380859, 0.03391692733764649, 0.0339046401977539, 0.0338155517578125, 0.03390771102905273, 0.03386265563964844, 0.03393228912353516, 0.033957889556884766, 0.0339128303527832, 0.033919998168945316, 0.03400089645385742, 0.033909759521484374, 0.033939456939697264, 0.03395891189575195, 0.033957950592041014, 0.03394451141357422, 0.03398553466796875, 0.03399782562255859, 0.03397017669677734, 0.0339865608215332, 0.03398860931396484, 0.033990657806396485, 0.03398963165283203, 0.03400089645385742, 0.033999870300292966, 0.034006046295166015, 0.03399164962768555, 0.03401318359375, 0.03400294494628906, 0.034034687042236327, 0.034028545379638675, 0.03403673553466797, 0.03403571319580078, 0.03401631927490235, 0.034071487426757814, 0.034049022674560545, 0.034070526123046875, 0.034034687042236327, 0.03406643295288086, 0.0340766716003418, 0.03413401412963867, 0.03402547073364258, 0.03407257461547852, 0.0341104621887207, 0.034236415863037106, 0.03409100723266602, 0.034108417510986325, 0.03407974243164062, 0.03410230255126953, 0.034064350128173826, 0.11215257263183594, 0.033833984375, 0.03380223846435547, 0.03381760025024414, 0.033801246643066406, 0.03384521484375, 0.03384524917602539, 0.03382374572753906, 0.03386265563964844, 0.03382476806640625, 0.03388415908813477, 0.03385139083862305, 0.03387494277954101, 0.03386777496337891, 0.03392716979980469, 0.033854465484619144, 0.033931262969970705, 0.0338524169921875, 0.033887233734130856, 0.03386880111694336, 0.0339343376159668, 0.03388927841186523, 0.03393024063110352, 0.03398553466796875, 0.03391692733764649, 0.03395072174072265, 0.033964031219482424, 0.033966079711914066, 0.033941505432128906, 0.033947647094726564, 0.033944576263427735, 0.033993728637695314, 0.03415654373168946, 0.03420159912109375, 0.03400089645385742, 0.033976383209228515, 0.03401004791259766, 0.03399679946899414, 0.03398144149780274, 0.033966079711914066, 0.03399884796142578, 0.034035743713378905, 0.03400086212158203, 0.03402649688720703, 0.03400294494628906, 0.03402137756347656, 0.034044929504394535, 0.034000926971435544, 0.03408585739135742, 0.034050048828125, 0.034051071166992186, 0.034037792205810546, 0.03407155227661133, 0.03403567886352539, 0.0340469741821289, 0.03401932907104492, 0.034100223541259765, 0.034049022674560545, 0.034083839416503905, 0.03407974243164062, 0.03408998489379883, 0.034097152709960936, 0.03416371154785156, 0.11231027221679687, 0.03382681655883789, 0.033832000732421874, 0.03384313583374023, 0.03385145568847656, 0.033833919525146486, 0.033841152191162106, 0.033879039764404296, 0.03387801742553711, 0.03386777496337891, 0.033870849609375, 0.0338974723815918, 0.033909759521484374, 0.03391897583007813, 0.033898494720458985, 0.03390259170532227, 0.03384729766845703, 0.033947647094726564, 0.03386470413208008, 0.03393024063110352, 0.033893375396728515, 0.033966079711914066, 0.033919998168945316, 0.03397119903564453, 0.033887233734130856, 0.0339681282043457, 0.033939456939697264, 0.03395276641845703, 0.03392716979980469, 0.0339865608215332, 0.033949695587158206, 0.033995777130126956, 0.03393952178955078, 0.034002880096435546, 0.033972225189208984, 0.03404390335083008, 0.0339681282043457, 0.034028545379638675, 0.03399168014526367, 0.03402348709106445, 0.03398649597167969, 0.03407462310791016, 0.034070526123046875, 0.03406028747558594, 0.03402751922607422, 0.03405414581298828, 0.034028545379638675, 0.03407360076904297, 0.03403571319580078, 0.034113536834716796, 0.03405414581298828, 0.03410124969482422, 0.03405311965942383, 0.03410124969482422, 0.03405619049072266, 0.034108417510986325, 0.03407462310791016, 0.03409612655639648, 0.03404185485839844, 0.03410943984985351, 0.03406131362915039, 0.034105342864990236, 0.03409612655639648, 0.11220893096923829, 0.03380118560791016, 0.03380121612548828, 0.03381452941894531, 0.03379097747802735, 0.033838081359863284, 0.033838081359863284, 0.033911808013916016, 0.033914878845214845, 0.03386982345581055, 0.03384320068359375, 0.033854465484619144, 0.033838081359863284, 0.03392409515380859, 0.03385958480834961, 0.03385651016235352, 0.03393228912353516, 0.03395481491088867, 0.033890304565429685, 0.03387494277954101, 0.0339159049987793, 0.033964031219482424, 0.033906688690185545, 0.03392921447753906, 0.03392716979980469, 0.03392003250122071, 0.03399983978271484, 0.03392409515380859, 0.033966079711914066, 0.03395379257202148, 0.03395276641845703, 0.03392921447753906, 0.03399168014526367, 0.03396505737304688, 0.033993728637695314, 0.03425689697265625, 0.034716670989990234, 0.03441561508178711, 0.03403878402709961, 0.03402751922607422, 0.03401318359375, 0.03402342224121094, 0.03399270248413086, 0.03400806427001953, 0.03400908660888672, 0.03401113510131836, 0.03400908660888672, 0.03404185485839844, 0.03400908660888672, 0.03404185485839844, 0.03401113510131836, 0.03404800033569336, 0.03404083251953125, 0.03402444839477539, 0.03405516815185547, 0.03405209732055664, 0.03405414581298828, 0.03405619049072266, 0.03406643295288086, 0.034111488342285154, 0.03412688064575195, 0.03407561492919922, 0.03406959915161133, 0.11224976348876953, 0.03382681655883789, 0.033788928985595705, 0.03384627151489258, 0.03383193588256836, 0.03385036849975586, 0.03379203033447266, 0.033864673614501954, 0.03383295822143555, 0.03382271957397461, 0.0338524169921875, 0.03385343933105469, 0.033876991271972655, 0.033893375396728515, 0.033906688690185545, 0.03386880111694336, 0.033966079711914066, 0.034146305084228515, 0.03413094329833984, 0.03394047927856445, 0.03390873718261719, 0.03388313674926758, 0.033928192138671875, 0.033909793853759765, 0.03394761657714844, 0.03392921447753906, 0.03396915054321289, 0.03392716979980469, 0.03391692733764649, 0.03392409515380859, 0.03393638229370117, 0.033949695587158206, 0.03395686340332031, 0.03394867324829102, 0.03395897674560547, 0.03401516723632812, 0.03398860931396484, 0.03396915054321289, 0.03399168014526367, 0.03400396728515625, 0.03401113510131836, 0.034017345428466794, 0.033984447479248045, 0.03400396728515625, 0.0339865608215332, 0.03402035140991211, 0.0340582389831543, 0.034062335968017575, 0.03405721664428711, 0.03407360076904297, 0.03404083251953125, 0.03404294586181641, 0.03405612945556641, 0.03413094329833984, 0.03408998489379883, 0.03403366470336914, 0.03404390335083008, 0.034081790924072264, 0.034088958740234376, 0.034103294372558594, 0.034083839416503905, 0.03408486557006836, 0.034100223541259765, 0.11226112365722657, 0.03381043243408203, 0.03382988739013672, 0.033807361602783206, 0.03385139083862305, 0.03378688049316406, 0.03384729766845703, 0.03380326461791992, 0.033834049224853516, 0.03382675170898437, 0.03390566253662109, 0.033923072814941405, 0.03394559860229492, 0.03385343933105469, 0.034025665283203124, 0.0339403190612793, 0.03388412857055664, 0.03391801452636719, 0.03388204956054688, 0.03388620758056641, 0.0339128303527832, 0.03394559860229492, 0.033914878845214845, 0.03392102432250976, 0.03391897583007813, 0.03395072174072265, 0.03398963165283203, 0.03399168014526367, 0.033964031219482424, 0.033977344512939454, 0.03395686340332031, 0.03400294494628906, 0.03395276641845703, 0.0339865608215332, 0.03398553466796875, 0.033977344512939454, 0.03399782562255859, 0.03401420974731445, 0.03399168014526367, 0.033977344512939454, 0.033993728637695314, 0.03401318359375, 0.03400908660888672, 0.034012161254882815, 0.0339947509765625, 0.034016254425048825, 0.03405516815185547, 0.034081790924072264, 0.03405209732055664, 0.03404288101196289, 0.03402239990234375, 0.034051071166992186, 0.03404800033569336, 0.03408793640136719, 0.034044929504394535, 0.0342108154296875, 0.03411251068115234, 0.034098175048828124, 0.034100223541259765, 0.034098175048828124, 0.03406950378417969, 0.034080768585205076, 0.03410943984985351]",tokens/s,28.486656989631104,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-72B,Qwen/Qwen1.5-72B,cuda,0,42,,,True,,,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",qwen2,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,internlm/internlm-20b,internlm/internlm-20b,cuda,0,42,,,True,,,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",internlm,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-110B,Qwen/Qwen1.5-110B,cuda,0,42,,,True,,,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",qwen2,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Salesforce/codegen-16B-nl,Salesforce/codegen-16B-nl,cuda,0,42,,,True,,,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",codegen,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,m,m,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/m/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669472f8-26458458117a660f6c201551;55c0804a-6026-44d1-86aa-b7cc0d729371) Repository Not Found for url: https://huggingface.co/m/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: m is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,openai-community/gpt2-large,openai-community/gpt2-large,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1207.242752,2663.907328,0.0,2017.460224,1692.280832,s,10,0.19551939201354981,0.019551939201354982,0.0001408010870508943,0.01950425624847412,0.019655907821655272,0.01979617748260498,0.01990839321136475,"[0.01993644714355469, 0.019506912231445312, 0.019434688568115234, 0.019480192184448242, 0.019532320022583007, 0.019482336044311523, 0.019429727554321288, 0.019590431213378907, 0.01962473678588867, 0.01950160026550293]",tokens/s,13093.330403884376,kWh,2.3315087003245994e-07,1.2775568627509025e-07,8.694002133806692e-07,1.2303067696882196e-06,tokens/kWh,208078185.3007886,MB,1207.570432,2663.907328,0.0,2017.460224,1740.079616,s,10,12.674806030273437,1.2674806030273438,0.004450098605968816,1.267710693359375,1.2720069702148438,1.2741922058105468,1.2759403942871095,"[1.27637744140625, 1.2654049072265625, 1.266474365234375, 1.26938330078125, 1.268022216796875, 1.2593397216796876, 1.262382080078125, 1.26850146484375, 1.267399169921875, 1.2715213623046875]",tokens/s,49.704902662436155,kWh,1.474103270263074e-05,8.077823779609422e-06,3.138148669841923e-05,5.420034318065939e-05,tokens/kWh,1162354.2638837136,,s,629,12.829667316436774,0.020396927371123637,0.0023886441308962796,0.020084735870361328,0.02030510673522949,0.020455014419555665,0.040146983337402346,"[0.020393983840942383, 0.020355072021484375, 0.020923391342163086, 0.021122047424316406, 0.02011238479614258, 0.02020966339111328, 0.020320255279541014, 0.020076543807983398, 0.02016160011291504, 0.019937215805053712, 0.019975168228149414, 0.020174848556518556, 0.02065510368347168, 0.02089574432373047, 0.02012876892089844, 0.020192256927490236, 0.0204902400970459, 0.020776960372924806, 0.020920320510864256, 0.020962303161621093, 0.020864000320434572, 0.02084556770324707, 0.0208353271484375, 0.02148249626159668, 0.020967424392700194, 0.02123263931274414, 0.02061311912536621, 0.020355072021484375, 0.02027212715148926, 0.02010419273376465, 0.02010419273376465, 0.020106239318847655, 0.02012774467468262, 0.020091903686523437, 0.020002815246582033, 0.020075519561767577, 0.02002227210998535, 0.020143104553222657, 0.0200263671875, 0.020068351745605468, 0.020011007308959963, 0.02021683120727539, 0.020182016372680665, 0.02007347106933594, 0.02001817512512207, 0.019999744415283204, 0.02011238479614258, 0.02011238479614258, 0.01925939178466797, 0.019268608093261717, 0.019260416030883788, 0.020050943374633787, 0.020153343200683595, 0.01982361602783203, 0.019703807830810546, 0.019733503341674806, 0.019766271591186522, 0.02012057685852051, 0.019975168228149414, 0.020113407135009767, 0.02003455924987793, 0.02021171188354492, 0.040156158447265625, 0.020145151138305666, 0.019971071243286134, 0.019992576599121094, 0.02008678436279297, 0.02004377555847168, 0.020047903060913086, 0.019992544174194337, 0.020011007308959963, 0.02000486373901367, 0.020107263565063475, 0.019989503860473632, 0.019721216201782226, 0.019736576080322265, 0.02003558349609375, 0.020337696075439452, 0.02027516746520996, 0.019992576599121094, 0.02012876892089844, 0.02011142349243164, 0.02012870407104492, 0.020050943374633787, 0.020084735870361328, 0.020098047256469728, 0.02003660774230957, 0.02001408004760742, 0.020171775817871093, 0.02007859230041504, 0.020025344848632814, 0.020126720428466797, 0.020116479873657226, 0.02003046417236328, 0.020039680480957032, 0.020142080307006836, 0.020002815246582033, 0.02008166313171387, 0.02002124786376953, 0.020076543807983398, 0.02005299186706543, 0.020105215072631837, 0.020007936477661133, 0.02002124786376953, 0.020000768661499024, 0.020207616806030275, 0.020083711624145507, 0.02011955261230469, 0.02007142448425293, 0.020048896789550782, 0.020200447082519533, 0.020101119995117187, 0.02005401611328125, 0.020152320861816408, 0.02003558349609375, 0.019951616287231445, 0.020075519561767577, 0.019931135177612306, 0.02004275131225586, 0.020053024291992187, 0.02018611145019531, 0.020665311813354494, 0.02025267219543457, 0.02005606460571289, 0.020273151397705077, 0.04022784042358398, 0.02009600067138672, 0.019969024658203126, 0.020125696182250977, 0.02000486373901367, 0.019994623184204103, 0.020121599197387697, 0.019992576599121094, 0.019958784103393554, 0.02003455924987793, 0.02004582405090332, 0.02004377555847168, 0.020025344848632814, 0.0200581111907959, 0.020060159683227538, 0.020074495315551756, 0.020185087203979494, 0.02009702491760254, 0.01998028755187988, 0.02005401611328125, 0.02002227210998535, 0.020753408432006838, 0.020385791778564453, 0.020281343460083007, 0.02008166313171387, 0.020100095748901366, 0.020065343856811524, 0.020086719512939454, 0.019982400894165038, 0.019982271194458008, 0.01996086311340332, 0.020017120361328126, 0.020017152786254884, 0.019986431121826173, 0.020238336563110353, 0.02012876892089844, 0.02007040023803711, 0.02008064079284668, 0.02029363250732422, 0.020305952072143554, 0.020038623809814454, 0.02009600067138672, 0.019993600845336915, 0.020011007308959963, 0.020093952178955078, 0.020017152786254884, 0.020117536544799804, 0.020212703704833985, 0.019991552352905274, 0.02023628807067871, 0.020115455627441405, 0.02011238479614258, 0.01998028755187988, 0.020141056060791016, 0.020050943374633787, 0.02007040023803711, 0.02009600067138672, 0.020106239318847655, 0.020125696182250977, 0.02011238479614258, 0.02001919937133789, 0.02003455924987793, 0.020153343200683595, 0.040202239990234374, 0.020002815246582033, 0.020148223876953125, 0.019974143981933593, 0.019977279663085937, 0.019976192474365235, 0.019997695922851562, 0.019969984054565428, 0.020006912231445313, 0.019976192474365235, 0.02039091110229492, 0.021129215240478515, 0.020379648208618165, 0.02009702491760254, 0.020148223876953125, 0.020436992645263673, 0.02011955261230469, 0.02004684829711914, 0.020164608001708984, 0.02043903923034668, 0.020131839752197265, 0.020093952178955078, 0.020002815246582033, 0.020157472610473633, 0.02004374313354492, 0.020182016372680665, 0.020059135437011717, 0.020208639144897463, 0.02009600067138672, 0.02007040023803711, 0.020032512664794923, 0.020099071502685546, 0.02000588798522949, 0.020130815505981444, 0.020091903686523437, 0.020349952697753908, 0.02027724838256836, 0.020106239318847655, 0.02007961654663086, 0.02008166313171387, 0.020100095748901366, 0.020191232681274415, 0.020092927932739257, 0.020236352920532226, 0.020328384399414062, 0.020140031814575195, 0.020076543807983398, 0.02026905632019043, 0.02024448013305664, 0.020065280914306642, 0.020074495315551756, 0.02006220817565918, 0.02011136054992676, 0.020142080307006836, 0.020148223876953125, 0.02008883285522461, 0.020139007568359374, 0.02002841567993164, 0.020084735870361328, 0.02003865623474121, 0.020304895401000975, 0.020032512664794923, 0.020048959732055664, 0.04022675323486328, 0.02045132827758789, 0.020122623443603514, 0.02011955261230469, 0.020141056060791016, 0.020282367706298828, 0.020355072021484375, 0.020136959075927736, 0.020351999282836913, 0.020109312057495117, 0.02007142448425293, 0.02005606460571289, 0.020126720428466797, 0.020076543807983398, 0.020032512664794923, 0.02008678436279297, 0.02010419273376465, 0.020122623443603514, 0.0200447998046875, 0.020191232681274415, 0.02033459281921387, 0.020131839752197265, 0.019994623184204103, 0.019971071243286134, 0.019999744415283204, 0.020116479873657226, 0.020092927932739257, 0.02004991912841797, 0.02008780860900879, 0.020218944549560545, 0.020102079391479493, 0.02001817512512207, 0.020035648345947267, 0.020107200622558594, 0.02007040023803711, 0.020122623443603514, 0.02006937599182129, 0.020182016372680665, 0.020048896789550782, 0.020025344848632814, 0.020084735870361328, 0.020197376251220703, 0.01997209548950195, 0.020202560424804686, 0.020340671539306642, 0.020196352005004883, 0.020214784622192384, 0.020151296615600587, 0.019940351486206053, 0.02002022361755371, 0.020388959884643554, 0.020189088821411134, 0.02002841567993164, 0.019949567794799804, 0.01998028755187988, 0.020082719802856447, 0.02014409637451172, 0.020162559509277343, 0.020147199630737304, 0.019990528106689453, 0.020142112731933594, 0.019996639251708984, 0.020074495315551756, 0.039378944396972655, 0.01926348876953125, 0.019346431732177736, 0.019355648040771483, 0.01925119972229004, 0.019312639236450196, 0.01923686408996582, 0.01985945510864258, 0.02006220817565918, 0.020032512664794923, 0.020106239318847655, 0.020073503494262696, 0.02009187126159668, 0.020303871154785155, 0.02019327926635742, 0.020110336303710938, 0.020092927932739257, 0.019993600845336915, 0.02002943992614746, 0.02005299186706543, 0.02009600067138672, 0.020001792907714845, 0.02005401611328125, 0.019989503860473632, 0.019998720169067383, 0.020067327499389647, 0.020289535522460937, 0.020015104293823242, 0.020060159683227538, 0.02003455924987793, 0.020034624099731446, 0.020005823135375977, 0.02005504035949707, 0.020047903060913086, 0.02004374313354492, 0.019997695922851562, 0.020256767272949217, 0.020185087203979494, 0.020222976684570314, 0.020076543807983398, 0.020031488418579102, 0.020093952178955078, 0.020129791259765627, 0.019982336044311523, 0.02006118392944336, 0.019962879180908204, 0.02000383949279785, 0.020110336303710938, 0.020144128799438478, 0.020051967620849608, 0.019952640533447266, 0.019900415420532228, 0.02003865623474121, 0.019939327239990236, 0.019965951919555663, 0.019949567794799804, 0.02001919937133789, 0.020171775817871093, 0.02022809600830078, 0.02021171188354492, 0.02012057685852051, 0.020025344848632814, 0.02007347106933594, 0.040123390197753905, 0.0200263671875, 0.020023296356201172, 0.02017078399658203, 0.020047840118408204, 0.02006630325317383, 0.019958784103393554, 0.020011007308959963, 0.020097087860107422, 0.019639232635498046, 0.019715072631835938, 0.01969254493713379, 0.019559423446655275, 0.01942630386352539, 0.019340288162231444, 0.020154367446899413, 0.020296703338623046, 0.02021171188354492, 0.020064319610595703, 0.020114368438720703, 0.02003865623474121, 0.020023296356201172, 0.020246528625488282, 0.02030080032348633, 0.02025369644165039, 0.02026188850402832, 0.019975168228149414, 0.020124671936035156, 0.020024320602416993, 0.020060159683227538, 0.020040704727172853, 0.019963903427124022, 0.019994623184204103, 0.019965951919555663, 0.02003763198852539, 0.019957759857177734, 0.020060224533081053, 0.020115392684936523, 0.02022809600830078, 0.02006937599182129, 0.02001919937133789, 0.02003660774230957, 0.020015104293823242, 0.02020147132873535, 0.02020249557495117, 0.02006220817565918, 0.02007347106933594, 0.019894271850585937, 0.02004991912841797, 0.020033536911010744, 0.02000486373901367, 0.019924991607666014, 0.02002124786376953, 0.02023423957824707, 0.019992576599121094, 0.020161535263061522, 0.02001408004760742, 0.019975168228149414, 0.020075519561767577, 0.020084735870361328, 0.020114431381225584, 0.02008576011657715, 0.02006118392944336, 0.040180736541748044, 0.020031488418579102, 0.02008576011657715, 0.02004377555847168, 0.019945472717285157, 0.019954687118530275, 0.02002739143371582, 0.02001408004760742, 0.02003763198852539, 0.02006220817565918, 0.020102144241333008, 0.019951616287231445, 0.020031488418579102, 0.020143104553222657, 0.020396032333374024, 0.02024345588684082, 0.02000588798522949, 0.020117504119873047, 0.020230144500732423, 0.019950592041015625, 0.020008960723876954, 0.020065280914306642, 0.020060159683227538, 0.020165632247924805, 0.020102144241333008, 0.020105215072631837, 0.02003558349609375, 0.02008883285522461, 0.02016972732543945, 0.019997695922851562, 0.02028441619873047, 0.020076543807983398, 0.020133888244628906, 0.02019327926635742, 0.020153343200683595, 0.020298751831054687, 0.020110336303710938, 0.02010316848754883, 0.02005299186706543, 0.02003660774230957, 0.02021887969970703, 0.020124671936035156, 0.020063232421875, 0.02003558349609375, 0.020711423873901368, 0.0204083194732666, 0.020197376251220703, 0.020249599456787108, 0.020051967620849608, 0.020093984603881836, 0.020074464797973632, 0.020126720428466797, 0.020338687896728515, 0.02029363250732422, 0.020333568572998048, 0.020184064865112306, 0.020108287811279296, 0.020053024291992187, 0.020273120880126953, 0.020185087203979494, 0.020040704727172853, 0.020065280914306642, 0.020126720428466797, 0.040476673126220705, 0.020242431640625, 0.0198973445892334, 0.019926015853881835, 0.020009983062744142, 0.019998720169067383, 0.02007347106933594, 0.019985408782958985, 0.02003763198852539, 0.02014419174194336, 0.02019731140136719, 0.021082111358642578, 0.0200949764251709, 0.020171775817871093, 0.020351999282836913, 0.020172800064086914, 0.02023423957824707, 0.0202106876373291, 0.020184064865112306, 0.020057088851928712, 0.020237312316894532, 0.020165632247924805, 0.020287488937377928, 0.02007347106933594, 0.02007040023803711, 0.020172800064086914, 0.0200949764251709, 0.020001792907714845, 0.02007859230041504, 0.020032512664794923, 0.019971071243286134, 0.019968000411987305, 0.020115455627441405, 0.02002022361755371, 0.020171775817871093, 0.02001408004760742, 0.020139007568359374, 0.019960832595825196, 0.019953664779663087, 0.020190208435058594, 0.020133888244628906, 0.02002124786376953, 0.019969024658203126, 0.020020288467407228, 0.020175807952880858, 0.02004684829711914, 0.020029472351074218, 0.019990495681762695, 0.020108287811279296, 0.020160512924194338, 0.02005504035949707, 0.02045747184753418, 0.020131839752197265, 0.02009702491760254, 0.02004991912841797, 0.020008960723876954, 0.02007961654663086, 0.020060159683227538, 0.02006220817565918, 0.020183040618896485, 0.019998720169067383, 0.019924991607666014, 0.02003046417236328, 0.04023603057861328, 0.02008166313171387, 0.020350976943969725, 0.020196352005004883, 0.019957759857177734, 0.020129791259765627, 0.020102144241333008, 0.02007142448425293, 0.02002841567993164, 0.020175872802734376, 0.020140031814575195, 0.020168703079223634, 0.0200581111907959, 0.020001792907714845, 0.020118528366088868, 0.020137983322143553, 0.020376575469970702, 0.020065280914306642, 0.020009983062744142, 0.02007040023803711, 0.02012057685852051, 0.02003046417236328, 0.02003558349609375, 0.020337663650512695, 0.02022502326965332, 0.02029363250732422, 0.020059135437011717, 0.02012774467468262, 0.02011136054992676, 0.020130815505981444, 0.020341760635375978, 0.020148223876953125, 0.02022809600830078, 0.020065280914306642, 0.02008576011657715, 0.020107328414916994, 0.0201778564453125, 0.020136959075927736, 0.020068351745605468, 0.020109312057495117, 0.02004377555847168, 0.019990528106689453, 0.020231168746948244, 0.02005299186706543, 0.021133312225341795, 0.021317632675170898, 0.02044825553894043, 0.020297727584838866, 0.020121599197387697, 0.020345855712890625, 0.02021171188354492, 0.020091903686523437, 0.020142080307006836, 0.02004991912841797, 0.020139007568359374, 0.020115455627441405, 0.02009702491760254, 0.02004684829711914, 0.02011238479614258, 0.02005299186706543, 0.020072511672973633, 0.0201276798248291, 0.020389888763427736]",tokens/s,49.02699224274933,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,huggyllama/llama-65b,huggyllama/llama-65b,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ self.load_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 250, in load_model_with_no_weights self.pretrained_model = self.automodel_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3710, in from_pretrained model = cls(config, *model_args, **model_kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 1101, in __init__ self.model = LlamaModel(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 884, in __init__ [LlamaDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 884, in [LlamaDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 677, in __init__ self.self_attn = LLAMA_ATTENTION_CLASSES[config._attn_implementation](config=config, layer_idx=layer_idx) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 387, in __init__ super().__init__(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 264, in __init__ self.v_proj = nn.Linear(self.hidden_size, self.num_key_value_heads * self.head_dim, bias=config.attention_bias) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/linear.py"", line 98, in __init__ self.weight = Parameter(torch.empty((out_features, in_features), **factory_kwargs)) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_device.py"", line 78, in __torch_function__ return func(*args, **kwargs) torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 128.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,l,l,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/l/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6694746a-0c240cfa4fe5ad0030b27285;5d7d01f6-486a-4d78-9f4c-46cb35a86805) Repository Not Found for url: https://huggingface.co/l/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: l is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,B,B,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/B/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6693251b-363ad7ce3d434be813171af5;97d07606-3078-4424-aff7-a8539c4f7bea) Repository Not Found for url: https://huggingface.co/B/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: B is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,huggyllama/llama-30b,huggyllama/llama-30b,cuda,0,42,,,True,,,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",llama,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,meta-llama/Llama-2-13b-hf,meta-llama/Llama-2-13b-hf,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ self.load_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 250, in load_model_with_no_weights self.pretrained_model = self.automodel_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3710, in from_pretrained model = cls(config, *model_args, **model_kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 1101, in __init__ self.model = LlamaModel(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 884, in __init__ [LlamaDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 884, in [LlamaDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 679, in __init__ self.mlp = LlamaMLP(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 192, in __init__ self.gate_proj = nn.Linear(self.hidden_size, self.intermediate_size, bias=config.mlp_bias) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/linear.py"", line 98, in __init__ self.weight = Parameter(torch.empty((out_features, in_features), **factory_kwargs)) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_device.py"", line 78, in __torch_function__ return func(*args, **kwargs) torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 136.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-7B,,cuda,0,42,,,,,True,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,217063f5c507ed7cc255df7e1f64c4333a0b4dfe,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,,MB,1292.10368,17102.798848,0.0,16456.351744,16359.853056,s,10,1.0287446746826174,0.10287446746826172,0.001574352666792247,0.1023089599609375,0.1031449493408203,0.10536452407836913,0.1071401838684082,"[0.10758409881591798, 0.10240496063232422, 0.10230368041992187, 0.1022434539794922, 0.10224969482421875, 0.10231423950195312, 0.10226022338867187, 0.10228489685058594, 0.10244771575927734, 0.1026517105102539]",tokens/s,2488.4697466743123,kWh,1.2122054962344181e-06,6.642319068268247e-07,5.7926718676940615e-06,7.669109270755305e-06,tokens/kWh,33380669.248801485,MB,1292.10368,17102.798848,0.0,16456.351744,16416.242176,s,10,29.410879882812498,2.94108798828125,0.004424772369057995,2.9409384765624997,2.9470216064453125,2.9476450805664065,2.9481438598632814,"[2.9482685546875, 2.946883056640625, 2.935814697265625, 2.939655517578125, 2.93367822265625, 2.9402763671875, 2.943142822265625, 2.937574951171875, 2.9416005859375, 2.943985107421875]",tokens/s,21.42064441833198,kWh,3.477936609903701e-05,1.9060741952693285e-05,0.00012457632131630225,0.00017841642936803254,tokens/kWh,353106.4948623387,,s,629,30.28513584136962,0.0481480697001107,0.012214817356114677,0.04664934539794922,0.046851686096191406,0.04733276138305664,0.14937149353027343,"[0.04658995056152344, 0.04657766342163086, 0.046698558807373045, 0.046656448364257815, 0.046625823974609376, 0.04664214324951172, 0.04668928146362305, 0.04677427291870117, 0.04667903900146484, 0.04730060958862305, 0.047685630798339845, 0.047546367645263675, 0.047473663330078124, 0.046508033752441405, 0.046502910614013675, 0.04651520156860352, 0.04653363037109375, 0.04659814453125, 0.0465203857421875, 0.046744510650634764, 0.04672819137573242, 0.04664012908935547, 0.04670361709594727, 0.0478474235534668, 0.04763033676147461, 0.04754227066040039, 0.047580158233642575, 0.04751462554931641, 0.04651520156860352, 0.04645580673217774, 0.046604286193847655, 0.04659302520751953, 0.04655718231201172, 0.046581760406494144, 0.046568447113037106, 0.04662681579589844, 0.046615550994873044, 0.046622718811035156, 0.04671897506713867, 0.046635009765625, 0.04675993728637695, 0.046663681030273435, 0.046671871185302735, 0.046639102935791016, 0.04694015884399414, 0.04679270553588867, 0.04659609603881836, 0.04648755264282227, 0.04719619369506836, 0.047478816986083985, 0.04671583938598633, 0.046545921325683595, 0.046712833404541014, 0.046562305450439455, 0.046691326141357424, 0.046581760406494144, 0.046690303802490236, 0.04663603210449219, 0.04666572952270508, 0.04665651321411133, 0.04668009567260742, 0.04654691314697266, 0.14947225952148438, 0.046685184478759766, 0.04649881744384766, 0.04677119827270508, 0.04667801666259765, 0.04733030319213867, 0.046312446594238284, 0.04638412857055664, 0.04655615997314453, 0.046706687927246096, 0.04676095962524414, 0.04660326385498047, 0.04668108749389648, 0.046676990509033206, 0.046637054443359374, 0.046688255310058595, 0.046682113647460936, 0.04665958404541016, 0.04665446472167969, 0.04666470336914062, 0.04671692657470703, 0.04721667098999023, 0.0465305290222168, 0.04648857498168945, 0.046688255310058595, 0.047391742706298826, 0.04675686264038086, 0.04659711837768555, 0.046637054443359374, 0.046690303802490236, 0.04670873641967774, 0.046870529174804686, 0.046693374633789066, 0.04668415832519531, 0.04679679870605469, 0.046548992156982424, 0.04660326385498047, 0.04666163253784179, 0.04666777420043945, 0.04665958404541016, 0.0465530891418457, 0.04662169647216797, 0.04656947326660156, 0.04639129638671875, 0.046317569732666014, 0.046427135467529294, 0.04659199905395508, 0.04667084884643555, 0.04727500915527344, 0.04753919982910156, 0.04715315246582031, 0.04726476669311523, 0.04731903839111328, 0.047230976104736325, 0.04743884658813476, 0.047612926483154294, 0.04740300750732422, 0.04728422546386719, 0.04637388610839844, 0.04665446472167969, 0.04674764633178711, 0.04660326385498047, 0.0465530891418457, 0.14944256591796876, 0.04665651321411133, 0.04665139389038086, 0.04667289733886719, 0.046685184478759766, 0.046666751861572264, 0.046604286193847655, 0.04664831924438476, 0.046650367736816405, 0.046611457824707034, 0.04656742477416992, 0.0464721908569336, 0.04624486541748047, 0.04662793731689453, 0.046518241882324216, 0.04654892730712891, 0.046483455657958986, 0.04661043167114258, 0.04659814453125, 0.04659814453125, 0.046227455139160156, 0.0463012466430664, 0.046314430236816403, 0.04667903900146484, 0.04665446472167969, 0.04671180725097656, 0.04667596817016602, 0.04950223922729492, 0.04683055877685547, 0.046429183959960936, 0.04655104064941406, 0.04668108749389648, 0.046615615844726566, 0.04664825439453125, 0.04652851104736328, 0.04653772735595703, 0.0461998405456543, 0.04648137664794922, 0.0464640007019043, 0.04653055953979492, 0.04647116851806641, 0.04653772735595703, 0.04628377532958984, 0.04633804702758789, 0.046306304931640625, 0.04667084884643555, 0.04674662399291992, 0.04675481414794922, 0.04651827239990235, 0.04665651321411133, 0.046676990509033206, 0.046714881896972656, 0.047148033142089846, 0.046496768951416016, 0.04640256118774414, 0.046532608032226565, 0.04649574279785156, 0.046255104064941405, 0.046281726837158206, 0.04637388610839844, 0.046496768951416016, 0.04657356643676758, 0.046311424255371096, 0.14914764404296876, 0.04664012908935547, 0.046483455657958986, 0.04638412857055664, 0.04694940948486328, 0.0466728630065918, 0.04648448181152344, 0.046637054443359374, 0.04677529525756836, 0.04654489517211914, 0.04660019302368164, 0.046690303802490236, 0.04721664047241211, 0.04734668731689453, 0.046809120178222655, 0.04668412780761719, 0.04654489517211914, 0.04669440078735351, 0.04667494583129883, 0.04680089569091797, 0.046740478515625, 0.0468592643737793, 0.0464721908569336, 0.04657664108276367, 0.046663681030273435, 0.04656947326660156, 0.04665657424926758, 0.046613441467285154, 0.046676990509033206, 0.046615585327148434, 0.04660323333740234, 0.04671692657470703, 0.04668115234375, 0.04666259384155273, 0.046532608032226565, 0.04660326385498047, 0.046332992553710935, 0.04632160186767578, 0.04636876678466797, 0.046581760406494144, 0.046584831237792966, 0.046669822692871094, 0.04667903900146484, 0.046733375549316405, 0.046623680114746095, 0.04668108749389648, 0.0466063346862793, 0.04657356643676758, 0.04671590423583984, 0.046720001220703126, 0.046721023559570314, 0.046696449279785154, 0.046581760406494144, 0.04684492874145508, 0.04660531234741211, 0.04657254409790039, 0.04667494583129883, 0.04660224151611328, 0.04659814453125, 0.046690303802490236, 0.046721023559570314, 0.04664422225952149, 0.04674969482421875, 0.14936781311035155, 0.046698558807373045, 0.0464967041015625, 0.04665958404541016, 0.046565376281738284, 0.04652851104736328, 0.04612607955932617, 0.04642303848266602, 0.04643840026855469, 0.04656332778930664, 0.046639102935791016, 0.04674764633178711, 0.046644287109375, 0.04670969772338867, 0.04668928146362305, 0.04668108749389648, 0.046647296905517575, 0.04663296127319336, 0.04651827239990235, 0.0466319351196289, 0.04624486541748047, 0.046437374114990236, 0.046513214111328124, 0.046523326873779296, 0.046496768951416016, 0.04671897506713867, 0.0466431999206543, 0.046622718811035156, 0.04674460983276367, 0.04663292694091797, 0.04659609603881836, 0.046650367736816405, 0.04655923080444336, 0.0465715217590332, 0.046192638397216795, 0.04637900924682617, 0.04635443115234375, 0.04660940933227539, 0.04667193603515625, 0.046634944915771484, 0.04691251373291016, 0.04655411148071289, 0.04658182525634766, 0.046705600738525394, 0.04657766342163086, 0.046532608032226565, 0.046318622589111326, 0.04634006500244141, 0.04625305557250976, 0.046529537200927736, 0.046615550994873044, 0.04675174331665039, 0.046496768951416016, 0.04628070449829102, 0.04632166290283203, 0.04637696075439453, 0.046491649627685545, 0.04671897506713867, 0.04667494583129883, 0.046992385864257816, 0.0464640007019043, 0.046620670318603515, 0.046650367736816405, 0.1494958038330078, 0.04671590423583984, 0.046483455657958986, 0.04697804641723633, 0.046701568603515625, 0.046617599487304685, 0.046876670837402344, 0.04656025695800781, 0.04643635177612305, 0.0465530891418457, 0.04657459259033203, 0.04664115142822266, 0.046683135986328124, 0.04667801666259765, 0.04664524841308594, 0.046706687927246096, 0.04666777420043945, 0.046635009765625, 0.04670159912109375, 0.04681622314453125, 0.0465797119140625, 0.04664524841308594, 0.046655487060546875, 0.04663603210449219, 0.04665446472167969, 0.046668800354003906, 0.04650905609130859, 0.04662783813476563, 0.046840831756591796, 0.04665139389038086, 0.04664422225952149, 0.04681625747680664, 0.04722687911987305, 0.04669440078735351, 0.046342144012451174, 0.04629913711547851, 0.04659097671508789, 0.04660531234741211, 0.04656742477416992, 0.04653772735595703, 0.046461952209472655, 0.04648448181152344, 0.04646092987060547, 0.04656639862060547, 0.0465049934387207, 0.04683055877685547, 0.047608833312988284, 0.04683366394042969, 0.046835712432861325, 0.04674355316162109, 0.046410751342773435, 0.04667596817016602, 0.04659404754638672, 0.04681932830810547, 0.04665961456298828, 0.04665135955810547, 0.04656335830688477, 0.04672201538085938, 0.04672716903686523, 0.04662886428833008, 0.04666163253784179, 0.046676990509033206, 0.04654796981811524, 0.1495665283203125, 0.04660831832885742, 0.04655001449584961, 0.04666470336914062, 0.04655104064941406, 0.04733440017700195, 0.04753100967407226, 0.04713369750976563, 0.046584831237792966, 0.04646912002563477, 0.04659609603881836, 0.04654182434082031, 0.04656639862060547, 0.046635009765625, 0.04662886428833008, 0.04675174331665039, 0.04667494583129883, 0.04662988662719727, 0.046668800354003906, 0.046704639434814454, 0.046622718811035156, 0.04669440078735351, 0.04665139389038086, 0.046653438568115234, 0.046663681030273435, 0.046653438568115234, 0.04661248016357422, 0.046744640350341794, 0.04667180633544922, 0.046660606384277346, 0.04667398452758789, 0.04658476638793945, 0.04784537506103516, 0.04724428939819336, 0.04643942260742188, 0.04662169647216797, 0.0466319351196289, 0.04667084884643555, 0.04673535919189453, 0.04664524841308594, 0.046599166870117184, 0.04664934539794922, 0.04667596817016602, 0.046881790161132815, 0.04650700759887695, 0.04687257766723633, 0.04653363037109375, 0.046806015014648435, 0.046450687408447267, 0.046581760406494144, 0.04668415832519531, 0.04667903900146484, 0.046620670318603515, 0.04671184158325195, 0.04664009475708008, 0.04669747161865234, 0.04664217758178711, 0.04663296127319336, 0.046676990509033206, 0.04667903900146484, 0.046630912780761716, 0.046734336853027345, 0.0467322883605957, 0.1493729248046875, 0.04665958404541016, 0.04658073425292969, 0.04659609603881836, 0.046723072052001956, 0.04662988662719727, 0.04681727981567383, 0.046669822692871094, 0.04658380889892578, 0.04666572952270508, 0.04672204971313477, 0.04682547378540039, 0.046524417877197265, 0.046562305450439455, 0.04651007843017578, 0.04664934539794922, 0.04657254409790039, 0.04631961441040039, 0.04642508697509766, 0.046489601135253904, 0.04651724624633789, 0.04665958404541016, 0.04655615997314453, 0.046683135986328124, 0.04667494583129883, 0.04654182434082031, 0.04669440078735351, 0.046685184478759766, 0.046752769470214846, 0.04653363037109375, 0.046653438568115234, 0.0466165771484375, 0.04659711837768555, 0.04657049560546875, 0.046317569732666014, 0.04642508697509766, 0.04642816162109375, 0.046440448760986325, 0.04673535919189453, 0.046595073699951174, 0.04671590423583984, 0.046650367736816405, 0.04665753555297852, 0.046655487060546875, 0.04668620681762695, 0.04669235229492188, 0.04664115142822266, 0.04665862274169922, 0.046642112731933597, 0.046680065155029295, 0.04661862564086914, 0.0466431999206543, 0.04664115142822266, 0.046611457824707034, 0.04676403045654297, 0.04673126220703125, 0.0466165771484375, 0.04662988662719727, 0.04675174331665039, 0.04685823822021484, 0.04659711837768555, 0.04656435012817383, 0.046615550994873044, 0.149465087890625, 0.046635009765625, 0.0465715217590332, 0.046706687927246096, 0.04679065704345703, 0.046611457824707034, 0.046601215362548826, 0.04664422225952149, 0.0466431999206543, 0.04664012908935547, 0.04672512054443359, 0.04656332778930664, 0.04672716903686523, 0.046723072052001956, 0.04667391967773438, 0.04664934539794922, 0.04663603210449219, 0.04660019302368164, 0.04676095962524414, 0.04679372787475586, 0.04657561492919922, 0.04669440078735351, 0.046620670318603515, 0.04666265487670898, 0.0466431999206543, 0.04668620681762695, 0.04655110549926758, 0.046789569854736326, 0.04666777420043945, 0.046691326141357424, 0.04660019302368164, 0.04668620681762695, 0.046614528656005856, 0.04667289733886719, 0.04666163253784179, 0.04668620681762695, 0.046699520111083984, 0.04664934539794922, 0.04661862564086914, 0.04671692657470703, 0.046717952728271485, 0.04668928146362305, 0.04659711837768555, 0.04671590423583984, 0.046614528656005856, 0.046611457824707034, 0.04657254409790039, 0.046720001220703126, 0.046682113647460936, 0.046720001220703126, 0.0468326416015625, 0.04770611190795898, 0.04695449447631836, 0.04664627075195313, 0.04650188827514649, 0.047132671356201174, 0.046467071533203126, 0.046578689575195314, 0.04655001449584961, 0.046739456176757815, 0.04655411148071289, 0.046827518463134765, 0.046588958740234374, 0.1502340850830078, 0.046944255828857424, 0.04677734375, 0.046630912780761716, 0.046676990509033206, 0.04668012619018555, 0.0466677131652832, 0.046707710266113284, 0.04665753555297852, 0.04670361709594727, 0.04655718231201172, 0.046690303802490236, 0.04661964797973633, 0.04659814453125, 0.04674457550048828, 0.04671897506713867, 0.046699520111083984, 0.04664115142822266, 0.046653438568115234, 0.04670259094238281, 0.04666572952270508, 0.046699520111083984, 0.046631999969482425, 0.04676396942138672, 0.04763238525390625, 0.04698726272583008, 0.046691326141357424, 0.04662681579589844, 0.04671078491210937, 0.04660326385498047, 0.04668316650390625, 0.04659811019897461, 0.046653438568115234, 0.0466954231262207, 0.046723072052001956, 0.04668723297119141, 0.046653438568115234, 0.04678144073486328, 0.04657664108276367, 0.046843902587890625, 0.04685004806518555, 0.04661043167114258, 0.04671180725097656, 0.04659609603881836, 0.04652646255493164, 0.04692582321166992, 0.04676505661010742, 0.046535678863525394, 0.046565376281738284, 0.04657766342163086, 0.04649369430541992, 0.04792934417724609, 0.04684809494018555, 0.04660931015014649, 0.046568447113037106, 0.04664422225952149, 0.04667903900146484, 0.04660838317871094, 0.04666777420043945, 0.04661862564086914, 0.046700607299804686, 0.04675683212280273, 0.0466728630065918]",tokens/s,20.76926460870561,,,main,False,False,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,M,M,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/M/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669322af-196bd1317a96c28d6c51dc7c;6bf02f62-4268-41d2-ab31-a61c89855a19) Repository Not Found for url: https://huggingface.co/M/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: M is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,stabilityai/stablelm-base-alpha-3b,stabilityai/stablelm-base-alpha-3b,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1269.198848,8550.612992,0.0,7904.165888,7764.225024,s,10,0.38919372940063474,0.03891937294006348,0.00026700765164320884,0.03882700729370117,0.03899450340270996,0.03935338039398193,0.039640481986999515,"[0.039712257385253906, 0.038828094482421874, 0.03882592010498047, 0.03882271957397461, 0.038823009490966794, 0.03876937484741211, 0.03886867141723633, 0.03878736114501953, 0.038841567993164065, 0.03891475296020508]",tokens/s,6577.701043494316,kWh,4.612804399750937e-07,2.5276008109358367e-07,2.4215875403813143e-06,3.1356280614499915e-06,tokens/kWh,81642336.07528672,MB,1269.198848,8550.612992,0.0,7904.165888,7866.521088,s,10,11.456949218750001,1.145694921875,0.006738720415438151,1.1433384399414064,1.1468847900390626,1.1563556640625,1.16393236328125,"[1.1658265380859374, 1.143456298828125, 1.1447801513671876, 1.1439632568359375, 1.1432205810546876, 1.14321533203125, 1.1430030517578125, 1.14250244140625, 1.144014892578125, 1.1429666748046876]",tokens/s,54.988460537903606,kWh,1.3450753445742298e-05,7.3706515967933215e-06,6.063529623841905e-05,8.145670128095467e-05,tokens/kWh,773417.0302662377,,s,629,11.787404273986809,0.018739911405384446,0.004657064328016564,0.018147327423095702,0.01825605125427246,0.018713971710205078,0.05732671569824219,"[0.01882828712463379, 0.018766847610473633, 0.01880678367614746, 0.0188538875579834, 0.018861183166503905, 0.018824064254760742, 0.018768896102905275, 0.01885081672668457, 0.018788415908813475, 0.01879750442504883, 0.01884979248046875, 0.018896896362304686, 0.01882931137084961, 0.018790399551391602, 0.01883340835571289, 0.01887129592895508, 0.018800703048706055, 0.018734016418457032, 0.018874368667602538, 0.018889728546142577, 0.018765823364257812, 0.018766847610473633, 0.018683904647827147, 0.01860505676269531, 0.0186562557220459, 0.018677759170532226, 0.018647039413452148, 0.018744319915771485, 0.0184453125, 0.018487295150756835, 0.01842995262145996, 0.01840230369567871, 0.01836953544616699, 0.018307071685791015, 0.01839206314086914, 0.018295808792114256, 0.01820364761352539, 0.018256895065307616, 0.018316287994384766, 0.01827020835876465, 0.018199552536010744, 0.018166784286499024, 0.01822208023071289, 0.01816783905029297, 0.018224096298217772, 0.018176000595092775, 0.01819340705871582, 0.018166784286499024, 0.018197504043579102, 0.018165760040283203, 0.01814630317687988, 0.018174976348876954, 0.018159616470336915, 0.01819340705871582, 0.01818009567260742, 0.018197504043579102, 0.01820569610595703, 0.018199552536010744, 0.0182108154296875, 0.01830297660827637, 0.01824358367919922, 0.018174976348876954, 0.0573757438659668, 0.018140159606933593, 0.018117631912231445, 0.018044992446899413, 0.018091968536376953, 0.018070528030395508, 0.01809715270996094, 0.018092031478881835, 0.018091007232666014, 0.018060287475585936, 0.018077695846557617, 0.018074623107910158, 0.018100223541259765, 0.018081792831420897, 0.018128896713256838, 0.018076671600341796, 0.018086912155151368, 0.018096128463745118, 0.018109439849853515, 0.01808076858520508, 0.018148351669311523, 0.018117631912231445, 0.018111488342285157, 0.018092031478881835, 0.01808896064758301, 0.01813811111450195, 0.018147327423095702, 0.018152448654174806, 0.018134016036987305, 0.0182108154296875, 0.018145280838012694, 0.018159616470336915, 0.018125823974609375, 0.018139135360717772, 0.018128896713256838, 0.018147327423095702, 0.01818726348876953, 0.018293760299682618, 0.018119680404663087, 0.018137088775634767, 0.018153472900390624, 0.01813199996948242, 0.018185184478759765, 0.018159616470336915, 0.0181790714263916, 0.01817190361022949, 0.018211872100830077, 0.01825584030151367, 0.018158592224121094, 0.018189311981201172, 0.018176000595092775, 0.018186368942260743, 0.01815951919555664, 0.01826505661010742, 0.018173952102661133, 0.01824870491027832, 0.018145280838012694, 0.018173952102661133, 0.018309152603149415, 0.018159584045410158, 0.018174976348876954, 0.01816268730163574, 0.018190336227416993, 0.0573573112487793, 0.018115583419799804, 0.01807360076904297, 0.01805619239807129, 0.01808076858520508, 0.018101247787475586, 0.01810534477233887, 0.018102272033691406, 0.018102272033691406, 0.01816681671142578, 0.018122751235961913, 0.01812681579589844, 0.018121728897094725, 0.01819545555114746, 0.018164735794067383, 0.018082815170288084, 0.018102272033691406, 0.018112512588500978, 0.018148351669311523, 0.018078720092773438, 0.018099199295043944, 0.018113536834716795, 0.018127872467041017, 0.018124799728393554, 0.018147327423095702, 0.018117631912231445, 0.0181790714263916, 0.018106367111206053, 0.018135040283203126, 0.018116607666015624, 0.018149375915527344, 0.018119680404663087, 0.018132991790771484, 0.018120704650878908, 0.018131967544555663, 0.018113536834716795, 0.01814630317687988, 0.018143232345581056, 0.018144256591796876, 0.01817190361022949, 0.018151424407958985, 0.01820057678222656, 0.018516992568969725, 0.018562047958374024, 0.01821286392211914, 0.01819443130493164, 0.018239488601684572, 0.018198528289794923, 0.01824051284790039, 0.018183168411254884, 0.018215936660766603, 0.01820876884460449, 0.01820979118347168, 0.018211904525756835, 0.018230207443237306, 0.01821183967590332, 0.01822412872314453, 0.018242559432983398, 0.018197504043579102, 0.018197504043579102, 0.01820467185974121, 0.0182108154296875, 0.018207744598388673, 0.05735833740234375, 0.01815449523925781, 0.018061311721801757, 0.01808076858520508, 0.018110464096069336, 0.018114559173583983, 0.018084863662719726, 0.018147327423095702, 0.01808076858520508, 0.018147327423095702, 0.018052095413208007, 0.018125823974609375, 0.01824665641784668, 0.018356224060058594, 0.018131967544555663, 0.018117631912231445, 0.018160640716552736, 0.018082815170288084, 0.018109439849853515, 0.018116607666015624, 0.018128896713256838, 0.018115583419799804, 0.018152448654174806, 0.018096128463745118, 0.018156543731689453, 0.018108415603637695, 0.018144287109375, 0.018145248413085936, 0.018153472900390624, 0.018153472900390624, 0.01816166305541992, 0.018145280838012694, 0.018139135360717772, 0.018132991790771484, 0.018138143539428712, 0.018134048461914062, 0.01812985610961914, 0.01816476821899414, 0.018181087493896485, 0.018185216903686522, 0.01816985511779785, 0.018149375915527344, 0.018145280838012694, 0.01817087936401367, 0.01818009567260742, 0.018158687591552734, 0.01816156768798828, 0.01821388816833496, 0.01817190361022949, 0.018149471282958983, 0.01820047950744629, 0.018184192657470705, 0.01819340705871582, 0.018174976348876954, 0.018192384719848635, 0.018181119918823242, 0.01816268730163574, 0.018166784286499024, 0.01819647979736328, 0.01817804718017578, 0.01819647979736328, 0.018231296539306642, 0.01818009567260742, 0.05733990478515625, 0.018051136016845704, 0.018078655242919923, 0.01802956771850586, 0.01809715270996094, 0.018062335968017578, 0.018101247787475586, 0.018055200576782227, 0.01808176040649414, 0.018067455291748045, 0.018083839416503905, 0.018068479537963866, 0.018069503784179687, 0.018062335968017578, 0.018107391357421874, 0.01840332794189453, 0.018251775741577148, 0.01825382423400879, 0.018104320526123048, 0.01810534477233887, 0.018101247787475586, 0.018134016036987305, 0.018111488342285157, 0.018102272033691406, 0.018106367111206053, 0.018126848220825196, 0.018128896713256838, 0.018115583419799804, 0.018137088775634767, 0.018099199295043944, 0.01815449523925781, 0.01811667251586914, 0.01812575912475586, 0.018120704650878908, 0.018152448654174806, 0.018124799728393554, 0.018155519485473632, 0.018151424407958985, 0.01816985511779785, 0.018157567977905274, 0.018142208099365235, 0.018117631912231445, 0.018163711547851562, 0.018155519485473632, 0.018136064529418947, 0.01816985511779785, 0.018167808532714845, 0.018157567977905274, 0.018166784286499024, 0.018199552536010744, 0.018172927856445312, 0.01820364761352539, 0.018182144165039063, 0.018190336227416993, 0.018190336227416993, 0.018166784286499024, 0.01817087936401367, 0.01820159912109375, 0.0181790714263916, 0.018165760040283203, 0.01818828773498535, 0.018198528289794923, 0.01822003173828125, 0.05729280090332031, 0.018076671600341796, 0.01803468894958496, 0.018093055725097656, 0.01804595184326172, 0.018086912155151368, 0.018085887908935547, 0.018131967544555663, 0.018132991790771484, 0.018111488342285157, 0.018096128463745118, 0.018127872467041017, 0.01808793640136719, 0.018124799728393554, 0.018132991790771484, 0.018102272033691406, 0.018096128463745118, 0.018093055725097656, 0.018098175048828127, 0.018099199295043944, 0.018118656158447266, 0.01810534477233887, 0.018147327423095702, 0.018086048126220704, 0.01809596824645996, 0.018096128463745118, 0.01811574363708496, 0.01809699249267578, 0.018102272033691406, 0.018125823974609375, 0.018185312271118165, 0.018139039993286133, 0.018113536834716795, 0.018092031478881835, 0.018125823974609375, 0.018150463104248046, 0.0181278076171875, 0.018155519485473632, 0.01819443130493164, 0.018339935302734374, 0.01841244888305664, 0.01820979118347168, 0.018150400161743165, 0.018140159606933593, 0.018129919052124025, 0.018141183853149414, 0.01813811111450195, 0.01816166305541992, 0.01820364761352539, 0.018182144165039063, 0.01821696090698242, 0.01814031982421875, 0.018174816131591796, 0.018156543731689453, 0.01815449523925781, 0.01820467185974121, 0.018177024841308592, 0.01815353584289551, 0.018132928848266602, 0.018177024841308592, 0.018157567977905274, 0.0182108154296875, 0.01819443130493164, 0.057485313415527345, 0.018053119659423827, 0.018099199295043944, 0.01802239990234375, 0.01810534477233887, 0.018100223541259765, 0.018102272033691406, 0.018082815170288084, 0.018129919052124025, 0.0180633602142334, 0.018113536834716795, 0.018093055725097656, 0.01808076858520508, 0.018115583419799804, 0.018096128463745118, 0.01808896064758301, 0.018082815170288084, 0.018120704650878908, 0.018155519485473632, 0.018140159606933593, 0.018153472900390624, 0.018172927856445312, 0.018107456207275392, 0.01813907241821289, 0.018096128463745118, 0.018141183853149414, 0.018126848220825196, 0.018124799728393554, 0.018137088775634767, 0.018116607666015624, 0.01813811111450195, 0.01810438346862793, 0.0181278076171875, 0.018101247787475586, 0.018137088775634767, 0.018121728897094725, 0.01816268730163574, 0.018136064529418947, 0.01816985511779785, 0.018117631912231445, 0.01815449523925781, 0.018144256591796876, 0.018181119918823242, 0.018147327423095702, 0.018130943298339842, 0.018172927856445312, 0.018157567977905274, 0.018166784286499024, 0.01815449523925781, 0.0181790714263916, 0.01814630317687988, 0.018172927856445312, 0.018168832778930662, 0.01820159912109375, 0.018197504043579102, 0.01820057678222656, 0.018158592224121094, 0.01819545555114746, 0.01819443130493164, 0.018163711547851562, 0.018197504043579102, 0.018173952102661133, 0.018217983245849608, 0.05728255844116211, 0.018058240890502928, 0.018035711288452147, 0.018051071166992186, 0.01804595184326172, 0.01808492851257324, 0.018046911239624024, 0.018094079971313477, 0.018050048828125, 0.018083839416503905, 0.01808793640136719, 0.018103296279907227, 0.01807974433898926, 0.018069503784179687, 0.01807257652282715, 0.01808793640136719, 0.018095104217529297, 0.018125823974609375, 0.018153472900390624, 0.01808896064758301, 0.018120704650878908, 0.018112512588500978, 0.018114559173583983, 0.018092031478881835, 0.018110464096069336, 0.018114559173583983, 0.018153472900390624, 0.018117631912231445, 0.018139135360717772, 0.018114559173583983, 0.01813811111450195, 0.018127872467041017, 0.018149375915527344, 0.018149375915527344, 0.018136064529418947, 0.01813811111450195, 0.018137088775634767, 0.018157567977905274, 0.018126848220825196, 0.018156543731689453, 0.018128896713256838, 0.01817804718017578, 0.018176000595092775, 0.018182144165039063, 0.018156543731689453, 0.018197504043579102, 0.018158592224121094, 0.018158592224121094, 0.018165760040283203, 0.018149375915527344, 0.01817804718017578, 0.018173952102661133, 0.01819545555114746, 0.01817087936401367, 0.0181790714263916, 0.01818009567260742, 0.01816985511779785, 0.01813811111450195, 0.018174976348876954, 0.018163711547851562, 0.01817804718017578, 0.01818828773498535, 0.018174976348876954, 0.05737984085083008, 0.018052095413208007, 0.01807155227661133, 0.01805516815185547, 0.01807974433898926, 0.018069503784179687, 0.018078720092773438, 0.018052095413208007, 0.01808076858520508, 0.018053119659423827, 0.01808896064758301, 0.01806438446044922, 0.018125823974609375, 0.01808793640136719, 0.018106367111206053, 0.018052095413208007, 0.018108415603637695, 0.018116607666015624, 0.018110464096069336, 0.01802649688720703, 0.018082815170288084, 0.018074623107910158, 0.018116607666015624, 0.01810534477233887, 0.018104320526123048, 0.018119680404663087, 0.018124799728393554, 0.018192384719848635, 0.01860403251647949, 0.018531328201293946, 0.018182144165039063, 0.018129919052124025, 0.018155519485473632, 0.018127872467041017, 0.018192384719848635, 0.018125823974609375, 0.018129919052124025, 0.018099199295043944, 0.018130943298339842, 0.018135040283203126, 0.018150400161743165, 0.018159616470336915, 0.01816166305541992, 0.018123775482177733, 0.01816268730163574, 0.01818726348876953, 0.018164735794067383, 0.018147327423095702, 0.018148351669311523, 0.01817190361022949, 0.018184192657470705, 0.01818009567260742, 0.018182144165039063, 0.018250751495361327, 0.018235424041748046, 0.018298847198486328, 0.018316287994384766, 0.018198528289794923, 0.01823539161682129, 0.01817087936401367, 0.01820057678222656, 0.018168832778930662, 0.018309120178222657, 0.057490432739257816, 0.018091007232666014, 0.01807360076904297, 0.018098175048828127, 0.018069568634033202, 0.018108352661132813, 0.018121728897094725, 0.018094079971313477, 0.018059263229370116, 0.018078720092773438, 0.018075647354125975, 0.018085887908935547, 0.018111520767211915, 0.018097120285034178, 0.01809715270996094, 0.018092031478881835, 0.018081792831420897, 0.018101247787475586, 0.01810534477233887, 0.018086912155151368, 0.018153472900390624, 0.018084863662719726, 0.018134016036987305, 0.018113536834716795, 0.018168832778930662, 0.01820979118347168, 0.018177024841308592, 0.018140159606933593, 0.018121728897094725, 0.018137088775634767, 0.018110464096069336, 0.018128896713256838, 0.018124799728393554, 0.018123775482177733, 0.018132991790771484, 0.018141183853149414, 0.018145280838012694, 0.018155519485473632, 0.018136064529418947, 0.018182144165039063, 0.018141183853149414, 0.018165760040283203, 0.018167808532714845, 0.018137088775634767, 0.01816783905029297, 0.01814735984802246, 0.018144191741943358, 0.018134016036987305, 0.01820159912109375, 0.018166784286499024, 0.01819443130493164, 0.01817087936401367, 0.018199552536010744, 0.018181119918823242, 0.018142208099365235, 0.018173952102661133, 0.018168895721435548, 0.018181055068969727, 0.018176000595092775, 0.018185216903686522, 0.018176000595092775, 0.01822208023071289, 0.01816166305541992]",tokens/s,53.362045228915804,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,google/recurrentgemma-9b,google/recurrentgemma-9b,cuda,0,42,,,True,,,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",recurrent_gemma,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,mistralai/Mixtral-8x22B-v0.1,mistralai/Mixtral-8x22B-v0.1,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/base.py"", line 43, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 54, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1021, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 403 Client Error: Forbidden for url: https://huggingface.co/mistralai/Mixtral-8x22B-v0.1/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 398, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 321, in hf_raise_for_status raise GatedRepoError(message, response) from e huggingface_hub.utils._errors.GatedRepoError: 403 Client Error. (Request ID: Root=1-66475096-7f0cc4644edeb78b6cb2a47e;8da5b03a-d279-4200-9536-517a57406562) Cannot access gated repo for url https://huggingface.co/mistralai/Mixtral-8x22B-v0.1/resolve/main/config.json. Access to model mistralai/Mixtral-8x22B-v0.1 is restricted and you are not in the authorized list. Visit https://huggingface.co/mistralai/Mixtral-8x22B-v0.1 to ask for access. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 87, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/base.py"", line 56, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 37, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 62, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 928, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 631, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 686, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 416, in cached_file raise EnvironmentError( OSError: You are trying to access a gated repo. Make sure to have access to it at https://huggingface.co/mistralai/Mixtral-8x22B-v0.1. 403 Client Error. (Request ID: Root=1-66475096-7f0cc4644edeb78b6cb2a47e;8da5b03a-d279-4200-9536-517a57406562) Cannot access gated repo for url https://huggingface.co/mistralai/Mixtral-8x22B-v0.1/resolve/main/config.json. Access to model mistralai/Mixtral-8x22B-v0.1 is restricted and you are not in the authorized list. Visit https://huggingface.co/mistralai/Mixtral-8x22B-v0.1 to ask for access. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,01-ai/Yi-6B,01-ai/Yi-6B,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1263.824896,13032.226816,0.0,12385.779712,12315.692032,s,10,0.8444192581176757,0.08444192581176757,0.0002549405837826158,0.0843782081604004,0.08449670562744141,0.08484374313354492,0.08512137313842774,"[0.08519078063964844, 0.08431529235839844, 0.08436259460449219, 0.08435084533691406, 0.08441446685791015, 0.08441782379150391, 0.08439382171630859, 0.08441958618164062, 0.08427855682373046, 0.0842754898071289]",tokens/s,3031.669369676131,kWh,9.99760012778025e-07,5.478005807682556e-07,5.162805250689099e-06,6.710365844235379e-06,tokens/kWh,38149931.90273223,MB,1263.824896,13032.226816,0.0,12385.779712,12315.694592,s,10,20.553834350585937,2.0553834350585936,0.014585121476491863,2.0566546630859373,2.0749329589843746,2.0752970458984374,2.0755883154296875,"[2.031691162109375, 2.0353634033203125, 2.05194091796875, 2.07485205078125, 2.042063720703125, 2.05729638671875, 2.056012939453125, 2.061626953125, 2.0756611328125, 2.06732568359375]",tokens/s,30.65121520656998,kWh,2.415962840356524e-05,1.3240143448422583e-05,0.00010038725141151057,0.0001377870232634984,tokens/kWh,457227.38257812074,,s,629,21.279567874908462,0.03383079153403568,0.010042286309707655,0.032718879699707035,0.033363916015625,0.033898496246337895,0.11673862396240235,"[0.033157119750976564, 0.033181697845458984, 0.032911361694335936, 0.032666622161865236, 0.03325439834594727, 0.0331069450378418, 0.03298406219482422, 0.03319807815551758, 0.033160190582275394, 0.03338547134399414, 0.03304959869384766, 0.03265331268310547, 0.03326566314697266, 0.03453440093994141, 0.03389235305786133, 0.03242496109008789, 0.03221401596069336, 0.03210342407226562, 0.03183103942871094, 0.031937536239624024, 0.03167231941223145, 0.03204710388183594, 0.03203276824951172, 0.03179520034790039, 0.03179110336303711, 0.03176755142211914, 0.032094207763671875, 0.03178598403930664, 0.03176345634460449, 0.03182592010498047, 0.03180953598022461, 0.03176755142211914, 0.03185868835449219, 0.03183001518249512, 0.03176345634460449, 0.0316866569519043, 0.03179929542541504, 0.032026622772216795, 0.03211468887329102, 0.03208294296264649, 0.03205017471313477, 0.032118785858154295, 0.03196928024291992, 0.032024574279785153, 0.031974399566650394, 0.03203071975708008, 0.03199488067626953, 0.03216588973999023, 0.031855615615844726, 0.03193139266967773, 0.031730688095092774, 0.03200102233886719, 0.032000064849853516, 0.03174188804626465, 0.03176755142211914, 0.0318023681640625, 0.03177676773071289, 0.03172147178649903, 0.03168972778320313, 0.0317573127746582, 0.031834112167358396, 0.031715328216552735, 0.11657727813720703, 0.031698944091796875, 0.03160576057434082, 0.03168358421325684, 0.0317379207611084, 0.03178079986572266, 0.03171327972412109, 0.031676416397094724, 0.031735807418823245, 0.03199795150756836, 0.031699968338012696, 0.0320706558227539, 0.0332677116394043, 0.033192958831787106, 0.033508350372314456, 0.031937536239624024, 0.03173990440368652, 0.031753215789794925, 0.03169075202941894, 0.03197235107421875, 0.03160985565185547, 0.03172659111022949, 0.03163750457763672, 0.031719423294067385, 0.03169484710693359, 0.03168569564819336, 0.031652799606323244, 0.031627264022827145, 0.031682559967041016, 0.03167747116088867, 0.03174908828735352, 0.031699968338012696, 0.03172761535644531, 0.03332505416870117, 0.033331199645996096, 0.03328204727172852, 0.03313663864135742, 0.033470462799072266, 0.03321241760253906, 0.031704063415527346, 0.03168563270568848, 0.031591424942016604, 0.031693887710571286, 0.031582143783569334, 0.031615999221801756, 0.0328540153503418, 0.03329228973388672, 0.03171737670898438, 0.03193343925476074, 0.033454078674316406, 0.03178803253173828, 0.03321855926513672, 0.03335168075561523, 0.0332421760559082, 0.03321132659912109, 0.03328409576416016, 0.03313049697875976, 0.03275161743164062, 0.03320729446411133, 0.033099777221679685, 0.033186817169189455, 0.03318476867675781, 0.033230846405029296, 0.11675039672851563, 0.031877056121826175, 0.031678464889526366, 0.03162627220153809, 0.03166307258605957, 0.03167027282714844, 0.03160985565185547, 0.031677440643310545, 0.03313971328735352, 0.03311001586914063, 0.03283660888671875, 0.033977344512939454, 0.03405516815185547, 0.03446579360961914, 0.034149375915527344, 0.031947839736938474, 0.0317644157409668, 0.031765504837036135, 0.03166924858093262, 0.031747072219848634, 0.03179110336303711, 0.031749120712280275, 0.03162521553039551, 0.03166720008850098, 0.03194777679443359, 0.03203583908081055, 0.0333496322631836, 0.03390259170532227, 0.03343564987182617, 0.03376639938354492, 0.033964031219482424, 0.03343155288696289, 0.033323009490966796, 0.033329151153564454, 0.033372161865234375, 0.03329536056518555, 0.03323904037475586, 0.03322367858886719, 0.033309696197509765, 0.03324415969848633, 0.03364761734008789, 0.03391385650634766, 0.0333199348449707, 0.031916032791137694, 0.03197235107421875, 0.03200204849243164, 0.03203276824951172, 0.03181158447265625, 0.031722528457641605, 0.034181087493896485, 0.03315507125854492, 0.03307622528076172, 0.03163750457763672, 0.03159244728088379, 0.03176140785217285, 0.03192422485351563, 0.031731712341308595, 0.03158016014099121, 0.031644672393798826, 0.03174092864990234, 0.03178291130065918, 0.03281510543823242, 0.03318783950805664, 0.11670835113525391, 0.03165798377990723, 0.0317890567779541, 0.03196416091918945, 0.03308236694335937, 0.03162623977661133, 0.0331960334777832, 0.0333199348449707, 0.03334143829345703, 0.03187507247924805, 0.03192422485351563, 0.031650848388671875, 0.03162928009033203, 0.031682559967041016, 0.03160371208190918, 0.03163545608520508, 0.03165184020996094, 0.03165695953369141, 0.031664127349853514, 0.03162931251525879, 0.03298406219482422, 0.033963008880615236, 0.033926143646240234, 0.033827838897705076, 0.03390259170532227, 0.03197747230529785, 0.03177267265319824, 0.031666208267211914, 0.031919071197509766, 0.03328409576416016, 0.0333199348449707, 0.033127422332763674, 0.03336806488037109, 0.03381350326538086, 0.0340582389831543, 0.03338444900512695, 0.03323904037475586, 0.033718334197998044, 0.0333639030456543, 0.03326566314697266, 0.03326873779296875, 0.033165313720703124, 0.03315302276611328, 0.03160780715942383, 0.031838207244873046, 0.033255424499511715, 0.03315097427368164, 0.03320115280151367, 0.03308544158935547, 0.03183718490600586, 0.03703500747680664, 0.03437875366210937, 0.0340684814453125, 0.03399782562255859, 0.033911808013916016, 0.033882110595703126, 0.03391078567504883, 0.033925121307373046, 0.03385961532592773, 0.03314896011352539, 0.03321440124511719, 0.0332042236328125, 0.033963008880615236, 0.11689984130859375, 0.03200307083129883, 0.03191910362243652, 0.031927295684814457, 0.03211673736572265, 0.031940607070922854, 0.0319180793762207, 0.03196723175048828, 0.03200307083129883, 0.031748096466064454, 0.03163750457763672, 0.0316180477142334, 0.03172147178649903, 0.03167436790466309, 0.031628288269042966, 0.03177676773071289, 0.03199180793762207, 0.03331071853637695, 0.033363967895507815, 0.03278335952758789, 0.03239424133300781, 0.032589824676513675, 0.032927745819091796, 0.03295641708374023, 0.032517120361328124, 0.032524288177490236, 0.03253452682495117, 0.03261337661743164, 0.03251814270019531, 0.03254886245727539, 0.03219046401977539, 0.03240652847290039, 0.032732192993164065, 0.03283248138427734, 0.032707584381103515, 0.0325591049194336, 0.03215359878540039, 0.03275980758666992, 0.03268096160888672, 0.03263488006591797, 0.03259904098510742, 0.032519168853759765, 0.03267379379272461, 0.03263897705078125, 0.032233470916748046, 0.03232460784912109, 0.03289907073974609, 0.03268915176391601, 0.0324956169128418, 0.03264409637451172, 0.03274137496948242, 0.032484352111816404, 0.032543743133544925, 0.032487422943115234, 0.032570369720458986, 0.032508926391601564, 0.03257551956176758, 0.03252016067504883, 0.032942081451416014, 0.03260723114013672, 0.03256320190429687, 0.03243315124511719, 0.03211468887329102, 0.11717222595214843, 0.032064510345458985, 0.03212595367431641, 0.03266457748413086, 0.03296255874633789, 0.03267583847045898, 0.03261644744873047, 0.03273830413818359, 0.03273011016845703, 0.03249049758911133, 0.03278950500488281, 0.03305267333984375, 0.03290633773803711, 0.033069984436035156, 0.03290726470947265, 0.03286531066894531, 0.03286832046508789, 0.03278643035888672, 0.03266355133056641, 0.03285708618164063, 0.03219660949707031, 0.031922176361083986, 0.0317388801574707, 0.03162112045288086, 0.03178803253173828, 0.0317255687713623, 0.03182284736633301, 0.03180031967163086, 0.03258265686035156, 0.0321607666015625, 0.032140289306640625, 0.032126976013183595, 0.03213516616821289, 0.031908863067626955, 0.032438270568847655, 0.032824321746826174, 0.03326259231567383, 0.033192958831787106, 0.03247212982177734, 0.03293587112426758, 0.03311718368530273, 0.03308544158935547, 0.03308544158935547, 0.03291545486450195, 0.033309696197509765, 0.03321855926513672, 0.0326379508972168, 0.033068031311035154, 0.03287449645996094, 0.03292979049682617, 0.0329881591796875, 0.032492542266845705, 0.032952320098876955, 0.032925697326660154, 0.03300864028930664, 0.03290828704833984, 0.03304345703125, 0.03304243087768555, 0.03266252899169922, 0.03316428756713867, 0.0330967025756836, 0.032791553497314455, 0.03261542510986328, 0.1174466552734375, 0.03282534408569336, 0.03230310440063477, 0.03265024185180664, 0.03296768188476563, 0.0325294075012207, 0.03245772933959961, 0.03293695831298828, 0.03261849594116211, 0.03257753753662109, 0.03259084701538086, 0.03260927963256836, 0.032478206634521486, 0.03252326583862305, 0.0324505615234375, 0.032570369720458986, 0.0325928955078125, 0.03260416030883789, 0.032586753845214846, 0.03242700958251953, 0.03182899284362793, 0.03262156677246094, 0.032589824676513675, 0.03259494400024414, 0.03196211242675781, 0.03184435272216797, 0.032643070220947264, 0.03284275054931641, 0.03266355133056641, 0.03298611068725586, 0.03287859344482422, 0.032996353149414064, 0.033185791015625, 0.0330250244140625, 0.032194561004638675, 0.03297587203979492, 0.032438304901123045, 0.03302601623535156, 0.03316121673583984, 0.03305779266357422, 0.032726016998291016, 0.03254988861083984, 0.033023998260498046, 0.032699390411376955, 0.03223756790161133, 0.032361473083496094, 0.03228876876831055, 0.03202150344848633, 0.031730688095092774, 0.03171123123168945, 0.03162931251525879, 0.032331775665283204, 0.03228160095214844, 0.032570369720458986, 0.032595966339111326, 0.03284073638916016, 0.03343254470825195, 0.03374387359619141, 0.03342335891723633, 0.03318476867675781, 0.03310182571411133, 0.03302297592163086, 0.032661502838134765, 0.1168752670288086, 0.03203071975708008, 0.03198566436767578, 0.03241471862792969, 0.032054271697998044, 0.032026622772216795, 0.031953920364379884, 0.03195084762573242, 0.031736831665039066, 0.03171327972412109, 0.031735807418823245, 0.03175526428222656, 0.03170099258422852, 0.032382976531982424, 0.03298918533325195, 0.03249663925170899, 0.03283148956298828, 0.03306598281860352, 0.03301273727416992, 0.032920574188232424, 0.03282636642456055, 0.033050624847412106, 0.03378073501586914, 0.033716224670410154, 0.03337830352783203, 0.03306598281860352, 0.03311820983886719, 0.03324006271362305, 0.033056766510009765, 0.0328611831665039, 0.032709632873535156, 0.03297382354736328, 0.03302195358276367, 0.03316121673583984, 0.0330332145690918, 0.03299123382568359, 0.03306291198730469, 0.03287551879882813, 0.033445888519287106, 0.03342540740966797, 0.03301068878173828, 0.03295948791503906, 0.03300352096557617, 0.03294003295898437, 0.03292364883422851, 0.03303936004638672, 0.033001472473144534, 0.03265024185180664, 0.032950271606445314, 0.03215871810913086, 0.03313356781005859, 0.033185791015625, 0.033104896545410156, 0.03301171112060547, 0.03304243087768555, 0.03225395202636719, 0.032718879699707035, 0.0325959358215332, 0.032574462890625, 0.032454654693603514, 0.03205734252929687, 0.032589824676513675, 0.032290817260742184, 0.11744358062744141, 0.03276697540283203, 0.03361075210571289, 0.03300966262817383, 0.03330047988891602, 0.03308031845092774, 0.03261849594116211, 0.03315097427368164, 0.03309260940551758, 0.03298918533325195, 0.03285094451904297, 0.032865280151367186, 0.03209318542480469, 0.03180748748779297, 0.032072704315185545, 0.032024574279785153, 0.033277950286865234, 0.0330885124206543, 0.03291340637207031, 0.03322367858886719, 0.03300864028930664, 0.032299007415771484, 0.03378483200073242, 0.03298303985595703, 0.03202560043334961, 0.03185766410827637, 0.03199283218383789, 0.032059391021728514, 0.03587481689453125, 0.03463577651977539, 0.03335987091064453, 0.032998401641845705, 0.0331038703918457, 0.033175552368164066, 0.03281203079223633, 0.03294412612915039, 0.032476158142089845, 0.03288371276855469, 0.03279667282104492, 0.03319091033935547, 0.03303424072265625, 0.03276595306396484, 0.03242393493652344, 0.0328540153503418, 0.03304857635498047, 0.03281308746337891, 0.03308848190307617, 0.033148929595947264, 0.03322880172729492, 0.03302707290649414, 0.032975936889648436, 0.03328915023803711, 0.033142784118652346, 0.0327720947265625, 0.03211468887329102, 0.032927745819091796, 0.03281203079223633, 0.032761856079101564, 0.033018878936767575, 0.03339571380615235, 0.03352883148193359, 0.03331584167480469, 0.03295641708374023, 0.11700531005859376, 0.03209625625610352, 0.03211775970458984, 0.03201945495605469, 0.032043006896972655, 0.03282534408569336, 0.032942081451416014, 0.03284172821044922, 0.033165313720703124, 0.03299225616455078, 0.03286937713623047, 0.03216588973999023, 0.03213721466064453, 0.03219046401977539, 0.0321003532409668, 0.03212595367431641, 0.03208806228637695, 0.032091136932373046, 0.03336191940307617, 0.03315097427368164, 0.0330332145690918, 0.03301375961303711, 0.033083393096923826, 0.03317148971557617, 0.03288777542114258, 0.033097728729248044, 0.03291648101806641, 0.03298713684082031, 0.033363967895507815, 0.032844863891601565, 0.03304851150512695, 0.03313151931762695, 0.033086463928222655, 0.03229388809204101, 0.03313151931762695, 0.03313663864135742, 0.03313971328735352, 0.03318476867675781, 0.03311001586914063, 0.0331141128540039, 0.033135616302490234, 0.03314585494995117, 0.032753662109375, 0.033107967376708985, 0.03303628921508789, 0.033037311553955076, 0.03290521621704102, 0.03301171112060547, 0.03306393432617188, 0.03301171112060547, 0.03280384063720703, 0.03295948791503906, 0.03257958221435547, 0.0328950080871582, 0.03327385711669922, 0.03312329483032227, 0.03271475219726563, 0.03205120086669922, 0.03226729583740234, 0.032914398193359375, 0.0330250244140625, 0.03271680068969727, 0.03299430465698242]",tokens/s,29.558870917753836,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,/,/,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 373, in cached_file raise EnvironmentError( OSError: / does not appear to have a file named config.json. Checkout 'https://huggingface.co///tree/None' for available files. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,TencentARC/Mistral_Pro_8B_v0.1,,cuda,0,42,,,,,True,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,d35829e539df8480b726c647eeabf91e41eae047,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,,MB,1273.831424,19665.518592,0.0,19019.071488,18747.268096,s,10,1.1004881896972656,0.11004881896972656,0.0023048996004499615,0.10927312088012694,0.11063434829711914,0.11377157325744627,0.116281353225708,"[0.11690879821777343, 0.10935049438476563, 0.10919574737548828, 0.10904892730712891, 0.10894719696044922, 0.10919388580322266, 0.10957100677490235, 0.10940447998046875, 0.10893046569824219, 0.10993718719482422]",tokens/s,2326.2403213107023,kWh,1.2923166515746556e-06,7.081279742657858e-07,7.338004058803867e-06,9.338448684644309e-06,tokens/kWh,27413546.79401451,MB,1273.831424,19665.518592,0.0,19019.071488,18774.938112,s,10,38.08989453125,3.808989453125,0.019078616113550908,3.8025064697265627,3.8398253906249997,3.840421728515625,3.8408987988281247,"[3.83969287109375, 3.817088623046875, 3.84101806640625, 3.788548583984375, 3.787248779296875, 3.803858154296875, 3.792437255859375, 3.796133544921875, 3.80115478515625, 3.8227138671875]",tokens/s,16.539820016648527,kWh,4.4770425936018206e-05,2.4536668305228256e-05,0.0001569185995685957,0.0002262256938098422,tokens/kWh,278482.9562859279,,s,629,39.01295923614504,0.06202378256938794,0.013030398047888114,0.06020505523681641,0.061349887847900395,0.06157086715698242,0.1696836053466797,"[0.06014566421508789, 0.060393470764160156, 0.06012211227416992, 0.06210662460327149, 0.061674495697021485, 0.06131814575195312, 0.06132633590698242, 0.06183731079101563, 0.061502464294433595, 0.061158401489257816, 0.0613939208984375, 0.061315071105957034, 0.06130995178222656, 0.06123212814331055, 0.061502464294433595, 0.06128742218017578, 0.061246463775634766, 0.06126489639282227, 0.061238273620605466, 0.061445121765136716, 0.061418495178222655, 0.06131097412109375, 0.06129151916503906, 0.061282302856445314, 0.061238273620605466, 0.06134783935546875, 0.06123519897460938, 0.06125260925292969, 0.06193561553955078, 0.061625343322753906, 0.061363201141357425, 0.061295616149902345, 0.06102220916748047, 0.06139084625244141, 0.061284351348876956, 0.06127718353271484, 0.0612567024230957, 0.06181785583496094, 0.06134067153930664, 0.06156288146972656, 0.061243392944335937, 0.06116454315185547, 0.06133964920043945, 0.06128947067260742, 0.060061695098876954, 0.060083198547363284, 0.060060672760009766, 0.05997158432006836, 0.06000435256958008, 0.059845630645751956, 0.05991526412963867, 0.05991731262207031, 0.06007807922363281, 0.06015180969238281, 0.06000435256958008, 0.06011699295043945, 0.06089932632446289, 0.06028799819946289, 0.05994393539428711, 0.060091392517089844, 0.060091392517089844, 0.059921409606933596, 0.16969625854492187, 0.060061695098876954, 0.06082252883911133, 0.061591552734375, 0.06101196670532227, 0.06013030242919922, 0.06010572814941406, 0.0602501106262207, 0.06028799819946289, 0.05997158432006836, 0.06132735824584961, 0.06047436904907227, 0.0601077766418457, 0.05990092849731445, 0.0601077766418457, 0.05982617568969727, 0.060012542724609375, 0.05986918258666992, 0.059991039276123044, 0.06011801528930664, 0.060034046173095705, 0.06138163375854492, 0.06113075256347656, 0.0602081298828125, 0.06157619094848633, 0.061553665161132816, 0.06057164764404297, 0.06133964920043945, 0.060126209259033205, 0.060037120819091794, 0.06013542556762695, 0.06003507232666016, 0.060055553436279295, 0.0602347526550293, 0.06018048095703125, 0.060273662567138675, 0.06016921615600586, 0.06019071960449219, 0.0602081298828125, 0.060012542724609375, 0.05991424179077148, 0.06002483367919922, 0.06015385437011719, 0.06015488052368164, 0.06020403289794922, 0.06049894332885742, 0.06033715057373047, 0.060829696655273435, 0.06012723159790039, 0.06236671829223633, 0.06161612701416016, 0.061259777069091796, 0.0615464973449707, 0.06137036895751953, 0.06137036895751953, 0.06138163375854492, 0.06123929595947265, 0.061475841522216794, 0.061618175506591794, 0.061358081817626954, 0.061193214416503904, 0.06094643020629883, 0.06036275100708008, 0.1696942138671875, 0.059963390350341796, 0.060080127716064455, 0.06005759811401367, 0.05998387145996094, 0.06013132858276367, 0.059937793731689455, 0.06092595291137695, 0.060170238494873046, 0.06026342391967773, 0.06095052719116211, 0.061537281036376956, 0.0614389762878418, 0.061521919250488284, 0.06150143814086914, 0.061532161712646485, 0.06149631881713867, 0.061521919250488284, 0.061491199493408207, 0.061297664642333986, 0.06147481536865235, 0.061431808471679686, 0.06186086273193359, 0.061652992248535154, 0.061462528228759764, 0.06092390441894531, 0.060642303466796874, 0.0603955192565918, 0.060916736602783204, 0.06016614532470703, 0.061292545318603515, 0.06056447982788086, 0.06122393417358398, 0.060611583709716796, 0.06010060882568359, 0.06057779312133789, 0.06003302383422852, 0.06037606430053711, 0.060478462219238284, 0.06204723358154297, 0.06168166351318359, 0.06137036895751953, 0.06162432098388672, 0.06139084625244141, 0.061276161193847656, 0.0609699821472168, 0.06122086334228516, 0.061429759979248044, 0.06141952133178711, 0.060104705810546874, 0.06034636688232422, 0.060295169830322265, 0.060012542724609375, 0.06083174514770508, 0.060141632080078125, 0.06118598556518555, 0.06067507171630859, 0.06174822235107422, 0.06235136032104492, 0.0645560302734375, 0.06109593582153321, 0.06043648147583008, 0.06032179260253906, 0.1696563262939453, 0.06014771270751953, 0.06008524703979492, 0.0599736328125, 0.05992038345336914, 0.06009142303466797, 0.0601712646484375, 0.060379104614257814, 0.06002790451049805, 0.05975040054321289, 0.06011391830444336, 0.060055553436279295, 0.060083198547363284, 0.06000230407714844, 0.059878398895263675, 0.06011391830444336, 0.05972582244873047, 0.05986713409423828, 0.059979774475097655, 0.05998188781738281, 0.060198848724365234, 0.060044288635253906, 0.060235774993896485, 0.06028799819946289, 0.06016204833984375, 0.060219390869140625, 0.06016204833984375, 0.060167263031005856, 0.06033295822143555, 0.059843582153320314, 0.05988044738769531, 0.060273662567138675, 0.06010265731811523, 0.06018355178833008, 0.0602347526550293, 0.060148735046386716, 0.05998796844482422, 0.06057369613647461, 0.06069760131835938, 0.06060441589355469, 0.06010265731811523, 0.06028595352172852, 0.060276737213134764, 0.060007423400878904, 0.059924480438232425, 0.05986918258666992, 0.05994291305541992, 0.059979774475097655, 0.06023372650146484, 0.06008422470092773, 0.059840511322021485, 0.06009548950195313, 0.06000435256958008, 0.060025856018066405, 0.05995622253417969, 0.06044672012329102, 0.06012211227416992, 0.060260353088378904, 0.05996134567260742, 0.06030847930908203, 0.06129151916503906, 0.06032998275756836, 0.05996646499633789, 0.16963388061523438, 0.05992643356323242, 0.05969305419921875, 0.06001766586303711, 0.060075008392333984, 0.06026342391967773, 0.06012211227416992, 0.060006401062011716, 0.06002175903320312, 0.060114944458007816, 0.06008729553222656, 0.05980876922607422, 0.05993267059326172, 0.06014879989624024, 0.06008211135864258, 0.060050430297851565, 0.06004940795898438, 0.06002995300292969, 0.06032588958740234, 0.060096511840820314, 0.060633087158203126, 0.060158977508544924, 0.05991116714477539, 0.059924480438232425, 0.060093441009521485, 0.06010163116455078, 0.06045491027832031, 0.06027468872070312, 0.06011391830444336, 0.06007295989990234, 0.06010265731811523, 0.05996236801147461, 0.05997568130493164, 0.05994905471801758, 0.060020736694335934, 0.05989376068115235, 0.06005350494384765, 0.060093441009521485, 0.06013030242919922, 0.06020505523681641, 0.06084505462646484, 0.06030847930908203, 0.059930622100830076, 0.0600002555847168, 0.06010060882568359, 0.06002175903320312, 0.06013337707519531, 0.060058624267578124, 0.059937793731689455, 0.060128257751464846, 0.060175361633300783, 0.05999411010742187, 0.06016716766357422, 0.060082176208496096, 0.0601712646484375, 0.06014054489135742, 0.060316673278808595, 0.06019276809692383, 0.06014976119995117, 0.06023168182373047, 0.06025932693481445, 0.06012723159790039, 0.060148735046386716, 0.1697617950439453, 0.06006272125244141, 0.06000537490844727, 0.05998591995239258, 0.06001971054077149, 0.05995929718017578, 0.060022785186767576, 0.06013132858276367, 0.060252159118652344, 0.06019583892822265, 0.06055833435058594, 0.06031568145751953, 0.06125052642822266, 0.060652542114257815, 0.06033715057373047, 0.06034534454345703, 0.060827648162841794, 0.060903423309326174, 0.060485633850097656, 0.06008627319335937, 0.06022041702270508, 0.06029926300048828, 0.06068838500976562, 0.06021836853027344, 0.060548095703125, 0.06034739303588867, 0.06013542556762695, 0.06061670303344727, 0.060484638214111326, 0.06049174499511719, 0.06041702270507812, 0.06026444625854492, 0.06048767852783203, 0.060142593383789064, 0.06002175903320312, 0.06013849639892578, 0.05992755126953125, 0.06064332962036133, 0.06011084747314453, 0.06056447982788086, 0.06036684799194336, 0.0600709114074707, 0.06015488052368164, 0.0601343994140625, 0.059947006225585936, 0.06070272064208984, 0.06040371322631836, 0.06091475296020508, 0.06036678314208985, 0.06094028854370117, 0.06049075317382813, 0.06029209518432617, 0.060216320037841796, 0.060012542724609375, 0.06012108612060547, 0.06060441589355469, 0.06030438232421875, 0.06061568069458008, 0.06065971374511719, 0.0607303695678711, 0.061126655578613284, 0.06054195022583008, 0.060440574645996094, 0.17018675231933594, 0.05994803237915039, 0.05994291305541992, 0.060028926849365234, 0.06000128173828125, 0.05982822418212891, 0.06005452728271484, 0.06007910537719727, 0.060104705810546874, 0.061291584014892576, 0.060910526275634765, 0.06029721450805664, 0.060080127716064455, 0.060333057403564455, 0.060047359466552735, 0.06012723159790039, 0.06020403289794922, 0.06009548950195313, 0.06017740631103516, 0.06015084838867187, 0.06011795043945312, 0.060171295166015625, 0.060176414489746095, 0.06022649765014648, 0.060063743591308595, 0.06020505523681641, 0.06025939178466797, 0.06018348693847656, 0.06049894332885742, 0.06037299346923828, 0.060227584838867185, 0.06006476974487305, 0.06196022415161133, 0.06052246475219727, 0.06003507232666016, 0.05996646499633789, 0.060111873626708986, 0.06007398223876953, 0.06016716766357422, 0.06012723159790039, 0.0600709114074707, 0.060077056884765626, 0.060047359466552735, 0.06002687835693359, 0.05998080062866211, 0.05997158432006836, 0.060020801544189456, 0.06005651092529297, 0.060028926849365234, 0.05994905471801758, 0.05987737655639649, 0.06006784057617188, 0.060055553436279295, 0.060009471893310545, 0.06000336074829102, 0.05989782333374023, 0.060572673797607425, 0.0606484489440918, 0.060281856536865235, 0.0601016960144043, 0.05993056106567383, 0.060265472412109375, 0.060668926239013675, 0.17047654724121095, 0.06005964660644531, 0.06021324920654297, 0.06000537490844727, 0.059859966278076174, 0.05996748733520508, 0.060028926849365234, 0.06006476974487305, 0.06010572814941406, 0.060333057403564455, 0.06016204833984375, 0.06027571105957031, 0.060951553344726565, 0.06043545532226562, 0.06069760131835938, 0.060633087158203126, 0.06048668670654297, 0.06047126388549805, 0.06047334289550781, 0.06032486343383789, 0.06001356887817383, 0.06045798492431641, 0.060009471893310545, 0.06001663970947266, 0.05995315170288086, 0.05999513626098633, 0.060249088287353515, 0.0600709114074707, 0.06002790451049805, 0.06007398223876953, 0.06011699295043945, 0.06021529769897461, 0.06023372650146484, 0.06009446334838867, 0.060061695098876954, 0.06016614532470703, 0.06054297637939453, 0.06076620864868164, 0.06059622573852539, 0.060075008392333984, 0.060022785186767576, 0.060042240142822265, 0.060028926849365234, 0.060142593383789064, 0.060120094299316404, 0.06017839813232422, 0.06066592025756836, 0.06103545761108398, 0.06056249618530273, 0.060310462951660156, 0.06016409683227539, 0.060246017456054686, 0.060115966796875, 0.060055553436279295, 0.060265472412109375, 0.06010265731811523, 0.06011603164672852, 0.0601701774597168, 0.06032998275756836, 0.06029312133789062, 0.06021222305297851, 0.06017433547973633, 0.06041497421264649, 0.17006285095214843, 0.06015999984741211, 0.06017433547973633, 0.06015590286254883, 0.060249088287353515, 0.06020198440551758, 0.060199935913085936, 0.060189697265625, 0.05984460830688477, 0.060247039794921874, 0.06025932693481445, 0.060037120819091794, 0.06019276809692383, 0.05994710540771484, 0.06027356719970703, 0.060109825134277345, 0.060034046173095705, 0.060265472412109375, 0.06053887939453125, 0.06002483367919922, 0.06007295989990234, 0.05989068984985352, 0.060042240142822265, 0.06014156723022461, 0.06014156723022461, 0.05990399932861328, 0.06002380752563476, 0.06008524703979492, 0.06018867111206055, 0.0601569938659668, 0.06012409591674805, 0.06041190338134766, 0.060224510192871096, 0.06030950546264648, 0.06030847930908203, 0.06073855972290039, 0.06051737594604492, 0.060063743591308595, 0.06013132858276367, 0.060243968963623044, 0.06011084747314453, 0.060028926849365234, 0.06010879898071289, 0.06014668655395508, 0.06014566421508789, 0.060561439514160154, 0.06110614395141602, 0.06026342391967773, 0.060352512359619144, 0.060194816589355465, 0.06028287887573242, 0.06029312133789062, 0.06018355178833008, 0.060142593383789064, 0.06020505523681641, 0.060900352478027345, 0.06121062469482422, 0.06106009674072266, 0.0612567024230957, 0.06125260925292969, 0.06122393417358398, 0.06104166412353516, 0.06120550537109375, 0.17150157165527344, 0.061088768005371094, 0.061110271453857425, 0.06109395217895508, 0.061724609375, 0.0612782096862793, 0.06109798431396484, 0.06099353790283203, 0.06109183883666992, 0.06110310363769531, 0.06113792037963867, 0.06126899337768555, 0.061300735473632816, 0.06097715377807617, 0.06128947067260742, 0.06111334228515625, 0.061257728576660155, 0.061067264556884764, 0.06129971313476563, 0.06122598266601562, 0.0609617919921875, 0.06116454315185547, 0.061058048248291016, 0.06101094436645508, 0.061093887329101565, 0.06106316757202149, 0.06106009674072266, 0.0616703987121582, 0.06128844833374023, 0.061018112182617185, 0.06104166412353516, 0.06119014358520508, 0.06163046264648438, 0.06039961624145508, 0.060224510192871096, 0.060115966796875, 0.06006579208374024, 0.06017433547973633, 0.05986406326293945, 0.06012211227416992, 0.05998387145996094, 0.060061695098876954, 0.06016614532470703, 0.06012313461303711, 0.06006784057617188, 0.06012723159790039, 0.060082176208496096, 0.06021017456054688, 0.060165119171142575, 0.060142593383789064, 0.06000332641601563, 0.060243968963623044, 0.06011084747314453, 0.059991039276123044, 0.06011084747314453, 0.060037120819091794, 0.06013644790649414, 0.06024192047119141, 0.06015795135498047, 0.060148735046386716, 0.06015078353881836, 0.05998591995239258, 0.06001152038574219]",tokens/s,16.122847697675788,,,main,False,False,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-160m,EleutherAI/pythia-160m,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1269.82144,1147.666432,0.0,501.219328,454.728704,s,17,0.16560361576080326,0.00974138916240019,0.0003419302422618868,0.009794143676757813,0.009896576118469239,0.010116653060913085,0.010716450424194335,"[0.010866399765014648, 0.009874815940856934, 0.009794143676757813, 0.009864735603332519, 0.009811296463012695, 0.00977888011932373, 0.009474687576293946, 0.009856096267700195, 0.009481216430664062, 0.009442208290100097, 0.00939072036743164, 0.009529088020324707, 0.009929216384887696, 0.009852191925048829, 0.009828543663024903, 0.009385631561279298, 0.009443743705749512]",tokens/s,26279.619439505477,kWh,1.156546586970968e-07,6.33611939991707e-08,2.5184462417612314e-07,4.3086047687239064e-07,tokens/kWh,594159858.5655847,MB,1269.82144,1147.666432,0.0,501.219328,469.748736,s,17,9.979045043945312,0.5870026496438419,0.015159569917142852,0.590857421875,0.5931046264648437,0.6021229736328125,0.6272746923828125,"[0.6335626220703126, 0.5920603637695312, 0.5942630615234376, 0.59169091796875, 0.5915021362304688, 0.5923323364257812, 0.573457763671875, 0.590857421875, 0.5667788696289062, 0.5659039306640625, 0.5670030517578125, 0.580317626953125, 0.5907263793945312, 0.590920654296875, 0.5911905517578125, 0.5862230834960938, 0.5802542724609375]",tokens/s,107.32489885390575,kWh,6.890079751853761e-06,3.77546111640819e-06,1.1825060462176718e-05,2.249060133043867e-05,tokens/kWh,2801170.100985077,,s,1070,10.123462645530696,0.009461180042552057,0.0012566724668822472,0.009364480018615723,0.009486495685577393,0.009889331531524659,0.01898519567489624,"[0.01040998363494873, 0.010371071815490723, 0.010267647743225097, 0.010099712371826173, 0.01011302375793457, 0.009925632476806641, 0.009852928161621094, 0.009810943603515625, 0.00993280029296875, 0.010011648178100586, 0.01002291202545166, 0.009890815734863282, 0.009880576133728027, 0.009975808143615723, 0.009991168022155762, 0.009988096237182617, 0.009889792442321778, 0.009927680015563965, 0.00994099235534668, 0.009776127815246583, 0.009834495544433594, 0.009888768196105957, 0.009887743949890136, 0.00984166431427002, 0.010090496063232422, 0.01010483169555664, 0.01003110408782959, 0.009906175613403321, 0.010007552146911621, 0.009877504348754883, 0.009948160171508789, 0.009822208404541016, 0.009880576133728027, 0.010025983810424804, 0.009837599754333496, 0.009852895736694336, 0.009864192008972168, 0.009901056289672852, 0.009809920310974121, 0.009978879928588867, 0.010066944122314453, 0.010034175872802734, 0.00984067153930664, 0.00988156795501709, 0.009943039894104003, 0.009866239547729493, 0.01104691219329834, 0.010270719528198241, 0.009727999687194825, 0.00955084800720215, 0.009523200035095216, 0.00984166431427002, 0.011508735656738281, 0.01165824031829834, 0.01061580753326416, 0.011687935829162598, 0.011284480094909668, 0.009621503829956055, 0.009678848266601562, 0.009586688041687011, 0.009547776222229003, 0.009473024368286133, 0.019834880828857423, 0.009332799911499023, 0.009333696365356445, 0.009372672080993653, 0.009364480018615723, 0.009366527557373047, 0.00941260814666748, 0.009354240417480468, 0.009460736274719238, 0.009407487869262696, 0.009365504264831542, 0.009396224021911622, 0.009401344299316406, 0.009382911682128906, 0.009369600296020507, 0.009370623588562011, 0.00934502410888672, 0.009379839897155762, 0.009379839897155762, 0.009377792358398437, 0.009392127990722657, 0.009364480018615723, 0.009326656341552735, 0.00941766357421875, 0.009390080451965332, 0.009381919860839845, 0.009389023780822753, 0.009385984420776367, 0.009350144386291503, 0.00940339183807373, 0.009341952323913574, 0.009337856292724609, 0.009399295806884766, 0.009510944366455078, 0.009471967697143554, 0.009379839897155762, 0.009382911682128906, 0.009439231872558594, 0.009506815910339356, 0.009374719619750976, 0.009431039810180664, 0.009382911682128906, 0.009403455734252929, 0.009410495758056641, 0.00941055965423584, 0.009359359741210938, 0.009400320053100587, 0.009359359741210938, 0.009440256118774413, 0.009442303657531738, 0.00942899227142334, 0.009417728424072265, 0.009372672080993653, 0.009384960174560546, 0.009376768112182618, 0.009360383987426758, 0.009349120140075684, 0.009373696327209472, 0.00937171173095703, 0.009406399726867676, 0.009356287956237793, 0.009431039810180664, 0.009347071647644043, 0.01962700843811035, 0.009444352149963378, 0.0094136323928833, 0.00939724826812744, 0.00940544033050537, 0.009391103744506836, 0.009391103744506836, 0.00940544033050537, 0.009379839897155762, 0.009391167640686034, 0.009384896278381348, 0.009365504264831542, 0.009373696327209472, 0.009381888389587402, 0.009370623588562011, 0.009382911682128906, 0.009411616325378418, 0.009337823867797851, 0.009372672080993653, 0.009417792320251465, 0.009358271598815918, 0.009393152236938476, 0.009379839897155762, 0.009362431526184082, 0.009388031959533692, 0.009370623588562011, 0.009399295806884766, 0.00939724826812744, 0.009556991577148437, 0.009451519966125489, 0.010083328247070313, 0.009886719703674317, 0.009499679565429688, 0.009508831977844238, 0.00939417552947998, 0.009485312461853027, 0.00940339183807373, 0.009708543777465821, 0.009431039810180664, 0.009372672080993653, 0.00942899227142334, 0.009367551803588867, 0.009384960174560546, 0.009487360000610352, 0.009434111595153808, 0.009375743865966797, 0.009424896240234374, 0.009370623588562011, 0.00940339183807373, 0.009409536361694336, 0.009374719619750976, 0.009374719619750976, 0.009369600296020507, 0.009371647834777832, 0.009445376396179199, 0.009388031959533692, 0.009369600296020507, 0.009415679931640625, 0.009458687782287598, 0.009393152236938476, 0.009397312164306641, 0.009393088340759278, 0.009368576049804688, 0.019726335525512697, 0.009376768112182618, 0.009389120101928711, 0.009404416084289552, 0.009401280403137207, 0.00939724826812744, 0.009329664230346679, 0.009393152236938476, 0.009383935928344727, 0.009356287956237793, 0.0094269437789917, 0.009371647834777832, 0.009389056205749511, 0.009379839897155762, 0.009352191925048828, 0.00933683204650879, 0.009388031959533692, 0.009338879585266113, 0.00936143970489502, 0.00936240005493164, 0.009362431526184082, 0.009338879585266113, 0.00941260814666748, 0.009327615737915039, 0.009340928077697755, 0.00940339183807373, 0.009339903831481934, 0.009370623588562011, 0.009341952323913574, 0.00933683204650879, 0.009364480018615723, 0.009422880172729492, 0.0093603515625, 0.00940236759185791, 0.009392127990722657, 0.009369600296020507, 0.009360383987426758, 0.009384960174560546, 0.009351167678833008, 0.009400320053100587, 0.009361408233642577, 0.009374719619750976, 0.009382911682128906, 0.00940544033050537, 0.009401344299316406, 0.009392127990722657, 0.00970751953125, 0.009538559913635255, 0.009404416084289552, 0.009356351852416992, 0.009378751754760743, 0.00941875171661377, 0.009343999862670899, 0.009363455772399902, 0.009387007713317871, 0.009332736015319825, 0.009400320053100587, 0.009358336448669433, 0.009470975875854493, 0.009387007713317871, 0.009381888389587402, 0.009364480018615723, 0.00942182445526123, 0.019689472198486328, 0.009407487869262696, 0.00939417552947998, 0.009387007713317871, 0.009376768112182618, 0.009339903831481934, 0.009249792098999024, 0.009203712463378906, 0.009241600036621094, 0.009272319793701172, 0.009758720397949219, 0.009401344299316406, 0.009479167938232422, 0.009360383987426758, 0.009388031959533692, 0.009391103744506836, 0.009375807762145996, 0.009349056243896484, 0.009374719619750976, 0.0094269437789917, 0.009398271560668945, 0.009355263710021973, 0.009378815650939941, 0.009368576049804688, 0.009358336448669433, 0.009386048316955567, 0.009404352188110351, 0.009215999603271484, 0.009220095634460449, 0.009265151977539063, 0.009340928077697755, 0.009351167678833008, 0.009369600296020507, 0.009379839897155762, 0.009369600296020507, 0.00941158390045166, 0.009393152236938476, 0.009475071907043458, 0.009393152236938476, 0.00942188835144043, 0.009366463661193848, 0.009409536361694336, 0.009384960174560546, 0.009401344299316406, 0.009406463623046875, 0.009415679931640625, 0.009360383987426758, 0.009401344299316406, 0.009382911682128906, 0.009371647834777832, 0.00939724826812744, 0.009375743865966797, 0.009390080451965332, 0.009351167678833008, 0.009340928077697755, 0.009390080451965332, 0.00931942367553711, 0.00934502410888672, 0.009902079582214356, 0.009380864143371583, 0.009363455772399902, 0.009380864143371583, 0.009390080451965332, 0.01964031982421875, 0.009349120140075684, 0.00942080020904541, 0.00974847984313965, 0.009385984420776367, 0.009469951629638672, 0.009352191925048828, 0.009401344299316406, 0.009387007713317871, 0.009693183898925782, 0.009470975875854493, 0.009378815650939941, 0.009356287956237793, 0.009372672080993653, 0.009366527557373047, 0.009349120140075684, 0.009377792358398437, 0.009350144386291503, 0.009374719619750976, 0.009860095977783203, 0.009532416343688965, 0.00939417552947998, 0.009387007713317871, 0.009355263710021973, 0.009365504264831542, 0.009392127990722657, 0.009376768112182618, 0.009350144386291503, 0.00939417552947998, 0.009422847747802734, 0.009380864143371583, 0.009390080451965332, 0.009370623588562011, 0.009486335754394531, 0.009446399688720703, 0.009348095893859864, 0.009401344299316406, 0.009354240417480468, 0.009379839897155762, 0.009384960174560546, 0.009356287956237793, 0.009315327644348144, 0.009362431526184082, 0.009329664230346679, 0.009364480018615723, 0.009335807800292969, 0.009325568199157714, 0.009365504264831542, 0.009368576049804688, 0.00935321617126465, 0.009370623588562011, 0.009341952323913574, 0.009351167678833008, 0.009379839897155762, 0.00934502410888672, 0.009347071647644043, 0.009355263710021973, 0.009409536361694336, 0.009339903831481934, 0.009360383987426758, 0.009341952323913574, 0.009325568199157714, 0.009376768112182618, 0.018949119567871094, 0.008985600471496581, 0.008961024284362793, 0.008967167854309082, 0.008960000038146973, 0.008980480194091797, 0.009013248443603515, 0.009037823677062988, 0.008987680435180664, 0.009005023956298827, 0.008960000038146973, 0.008998944282531738, 0.008995840072631836, 0.009016287803649903, 0.00900710391998291, 0.008993824005126954, 0.008983519554138184, 0.008988672256469727, 0.009013248443603515, 0.00901529598236084, 0.008938495635986327, 0.008996864318847657, 0.008948736190795899, 0.008976384162902832, 0.008951807975769043, 0.00900607967376709, 0.008976384162902832, 0.008986623764038085, 0.008973312377929688, 0.008972288131713867, 0.00902143955230713, 0.009049087524414063, 0.008977408409118653, 0.008980480194091797, 0.008994815826416015, 0.008985600471496581, 0.00897433567047119, 0.008983551979064941, 0.009022463798522949, 0.009003007888793945, 0.009003007888793945, 0.008957951545715333, 0.008994815826416015, 0.008965120315551758, 0.00900812816619873, 0.009101311683654785, 0.009379839897155762, 0.009409536361694336, 0.009389056205749511, 0.00941260814666748, 0.009385984420776367, 0.009371647834777832, 0.009340928077697755, 0.00921395206451416, 0.009368576049804688, 0.009374719619750976, 0.009351167678833008, 0.009366527557373047, 0.009437184333801269, 0.00941260814666748, 0.009396224021911622, 0.009379839897155762, 0.009405471801757813, 0.01970684814453125, 0.009371647834777832, 0.009393152236938476, 0.009367551803588867, 0.009379839897155762, 0.009564160346984863, 0.009482239723205567, 0.009423871994018555, 0.009370623588562011, 0.009372672080993653, 0.009392127990722657, 0.009374784469604493, 0.009373632431030274, 0.00941260814666748, 0.00941977596282959, 0.0094136323928833, 0.009382911682128906, 0.00934502410888672, 0.009341952323913574, 0.009377792358398437, 0.009351167678833008, 0.009275391578674316, 0.009232383728027344, 0.009341952323913574, 0.009398271560668945, 0.009422847747802734, 0.009340928077697755, 0.009363455772399902, 0.009352191925048828, 0.009351167678833008, 0.009331711769104004, 0.009360383987426758, 0.009349120140075684, 0.009341952323913574, 0.009358336448669433, 0.009425919532775879, 0.009355263710021973, 0.009378815650939941, 0.009375743865966797, 0.009360383987426758, 0.009338879585266113, 0.00932147216796875, 0.009408512115478516, 0.009374719619750976, 0.009333760261535644, 0.009356287956237793, 0.009340928077697755, 0.009338879585266113, 0.009370623588562011, 0.009360383987426758, 0.009340928077697755, 0.009454591751098633, 0.009385984420776367, 0.009365504264831542, 0.009375743865966797, 0.009352224349975586, 0.00934499168395996, 0.009378815650939941, 0.009366527557373047, 0.009324543952941895, 0.009437184333801269, 0.009372672080993653, 0.009362431526184082, 0.018964479446411133, 0.008985631942749024, 0.008995807647705079, 0.008987648010253906, 0.008963071823120117, 0.00897439956665039, 0.008997823715209961, 0.008969216346740723, 0.008957951545715333, 0.00901734447479248, 0.009065471649169921, 0.008988672256469727, 0.008999967575073242, 0.008996831893920898, 0.009012224197387696, 0.00903987216949463, 0.008954879760742187, 0.00898252773284912, 0.009026559829711914, 0.008980480194091797, 0.008936448097229004, 0.008995840072631836, 0.008975359916687011, 0.008957951545715333, 0.009029631614685058, 0.008988672256469727, 0.008988672256469727, 0.008993791580200196, 0.008973312377929688, 0.008992768287658692, 0.008994815826416015, 0.008986623764038085, 0.008971263885498047, 0.008968192100524902, 0.00903270435333252, 0.009022463798522949, 0.008956928253173829, 0.008977408409118653, 0.008995840072631836, 0.008967167854309082, 0.008964096069335938, 0.008967167854309082, 0.009034751892089844, 0.00902143955230713, 0.008980480194091797, 0.00900607967376709, 0.008999936103820801, 0.009010239601135254, 0.008994751930236817, 0.008969280242919922, 0.008985535621643066, 0.008965120315551758, 0.008973312377929688, 0.00897433567047119, 0.008953856468200684, 0.008988672256469727, 0.009009152412414552, 0.00898252773284912, 0.009068544387817384, 0.008986623764038085, 0.00898252773284912, 0.008983551979064941, 0.008952832221984864, 0.01887129592895508, 0.008979455947875976, 0.009058303833007812, 0.008978495597839356, 0.008967103958129883, 0.008999936103820801, 0.008977408409118653, 0.008953856468200684, 0.008947711944580078, 0.009055232048034668, 0.008979455947875976, 0.008973312377929688, 0.008960000038146973, 0.008983551979064941, 0.00899891185760498, 0.009026559829711914, 0.008968192100524902, 0.008986656188964843, 0.00895792007446289, 0.008979455947875976, 0.008967231750488282, 0.008977343559265136, 0.009019392013549805, 0.008944640159606934, 0.008957951545715333, 0.00899071979522705, 0.008976384162902832, 0.008958975791931152, 0.008958975791931152, 0.008946687698364257, 0.008950783729553222, 0.008958975791931152, 0.00894976043701172, 0.00899071979522705, 0.009088000297546387, 0.008975359916687011, 0.009014271736145019, 0.008970239639282226, 0.009020416259765626, 0.008957951545715333, 0.00897433567047119, 0.008943615913391113, 0.00900607967376709, 0.008944640159606934, 0.008857600212097168, 0.008885248184204102, 0.008935423851013183, 0.008869919776916504, 0.008879072189331054, 0.008975359916687011, 0.008981504440307618, 0.008989695549011231, 0.008986623764038085, 0.008971263885498047, 0.009027584075927735, 0.009011199951171875, 0.00897433567047119, 0.008956928253173829, 0.00899071979522705, 0.008970239639282226, 0.009068544387817384, 0.009001983642578124, 0.008975359916687011, 0.018934783935546876, 0.008968192100524902, 0.009001983642578124, 0.009048064231872559, 0.008964096069335938, 0.00897433567047119, 0.008996864318847657, 0.008988672256469727, 0.008976384162902832, 0.008957951545715333, 0.009033727645874023, 0.008985600471496581, 0.008970239639282226, 0.008995840072631836, 0.009034751892089844, 0.00901529598236084, 0.008989695549011231, 0.009009152412414552, 0.009079808235168458, 0.00901734447479248, 0.008951807975769043, 0.008973312377929688, 0.008988672256469727, 0.008958975791931152, 0.008976384162902832, 0.008994815826416015, 0.00900607967376709, 0.008968192100524902, 0.008973312377929688, 0.008965120315551758, 0.008953856468200684, 0.008976384162902832, 0.008973312377929688, 0.009054207801818847, 0.008936448097229004, 0.008976384162902832, 0.008948736190795899, 0.008980480194091797, 0.009052160263061524, 0.008994815826416015, 0.009056256294250489, 0.008963071823120117, 0.008979455947875976, 0.008989695549011231, 0.008995840072631836, 0.00897433567047119, 0.008964096069335938, 0.008980480194091797, 0.008981504440307618, 0.008975359916687011, 0.008996864318847657, 0.008973312377929688, 0.00897433567047119, 0.008977408409118653, 0.009073663711547851, 0.008992768287658692, 0.008957951545715333, 0.009019392013549805, 0.009067520141601563, 0.008991744041442871, 0.008983551979064941, 0.008993791580200196, 0.008991744041442871, 0.01900851249694824, 0.008993791580200196, 0.008988672256469727, 0.008981504440307618, 0.008943615913391113, 0.009000960350036622, 0.00899071979522705, 0.008963071823120117, 0.008965120315551758, 0.009022463798522949, 0.009022463798522949, 0.008965120315551758, 0.009044992446899413, 0.008978431701660156, 0.009014271736145019, 0.008988672256469727, 0.008992768287658692, 0.009072640419006347, 0.009010175704956054, 0.008988672256469727, 0.00897433567047119, 0.009000960350036622, 0.009005056381225587, 0.008977408409118653, 0.00899891185760498, 0.009026559829711914, 0.008967167854309082, 0.008948736190795899, 0.009001983642578124, 0.00899071979522705, 0.008981504440307618, 0.008956928253173829, 0.009060352325439454, 0.0095447359085083, 0.009396191596984863, 0.009393152236938476, 0.009379839897155762, 0.009399295806884766, 0.009360383987426758, 0.009372672080993653, 0.009352191925048828, 0.009451519966125489, 0.009434111595153808, 0.009425919532775879, 0.00934502410888672, 0.009373696327209472, 0.00934502410888672, 0.009361408233642577, 0.009406463623046875, 0.009363455772399902, 0.009407487869262696, 0.009369600296020507, 0.009381888389587402, 0.009387007713317871, 0.009373696327209472, 0.009367551803588867, 0.009612288475036621, 0.009530367851257325, 0.009515007972717286, 0.009977855682373048, 0.009527296066284179, 0.009473024368286133, 0.009439231872558594, 0.01985945510864258, 0.009339903831481934, 0.009377792358398437, 0.0093307523727417, 0.009375679969787597, 0.009393152236938476, 0.009367551803588867, 0.00940544033050537, 0.009559040069580077, 0.009425919532775879, 0.009371647834777832, 0.009376768112182618, 0.009348095893859864, 0.009370623588562011, 0.009374719619750976, 0.009364480018615723, 0.00942899227142334, 0.009348095893859864, 0.00939417552947998, 0.009432064056396485, 0.009380864143371583, 0.009362431526184082, 0.009381888389587402, 0.009375743865966797, 0.009376768112182618, 0.009486335754394531, 0.009354240417480468, 0.009395199775695801, 0.009348095893859864, 0.00934502410888672, 0.009365504264831542, 0.009373696327209472, 0.009383935928344727, 0.009338879585266113, 0.009371647834777832, 0.009346048355102539, 0.009342975616455078, 0.009399295806884766, 0.009369600296020507, 0.009366527557373047, 0.009239551544189453, 0.009192447662353515, 0.009172991752624511, 0.009211903572082519, 0.009249792098999024, 0.009202688217163087, 0.009272319793701172, 0.009417728424072265, 0.009392127990722657, 0.009377792358398437, 0.009757696151733398, 0.009370623588562011, 0.00939724826812744, 0.009341952323913574, 0.00933683204650879, 0.009376768112182618, 0.009364480018615723, 0.009346048355102539, 0.009431039810180664, 0.009391103744506836, 0.00941158390045166, 0.009374719619750976, 0.009361408233642577, 0.019706880569458008, 0.009331711769104004, 0.009422847747802734, 0.009383935928344727, 0.009447423934936524, 0.009379839897155762, 0.009360383987426758, 0.009392127990722657, 0.009404416084289552, 0.009452544212341308, 0.009377792358398437, 0.009367551803588867, 0.00941875171661377, 0.009384960174560546, 0.00942899227142334, 0.009391103744506836, 0.009387007713317871, 0.009362431526184082, 0.009357312202453612, 0.009495552062988282, 0.009391103744506836, 0.009372672080993653, 0.009431039810180664, 0.009388031959533692, 0.00939417552947998, 0.009389056205749511, 0.009348095893859864, 0.009384960174560546, 0.009379839897155762, 0.009363455772399902, 0.009368576049804688, 0.009407551765441894, 0.009354175567626953, 0.009358336448669433, 0.009364480018615723, 0.009348095893859864, 0.009348095893859864, 0.009362431526184082, 0.009362463951110839, 0.009398240089416504, 0.00939417552947998, 0.009351167678833008, 0.009415679931640625, 0.009370623588562011, 0.009372672080993653, 0.009369600296020507, 0.009360383987426758, 0.009356287956237793, 0.009368576049804688, 0.00942182445526123, 0.009385984420776367, 0.009273344039916993, 0.00921497631072998, 0.009294848442077636, 0.00923852825164795, 0.009234432220458985, 0.009343071937561035, 0.009414560317993164, 0.009401344299316406, 0.009338879585266113, 0.009373696327209472, 0.009363455772399902, 0.009356287956237793, 0.019703807830810546, 0.009461759567260742, 0.009390080451965332, 0.009352191925048828, 0.009364480018615723, 0.009346048355102539, 0.009447423934936524, 0.009399295806884766, 0.009359359741210938, 0.009406463623046875, 0.009395199775695801, 0.009339903831481934, 0.009360383987426758, 0.00940544033050537, 0.009380864143371583, 0.00939724826812744, 0.009372672080993653, 0.009334783554077148, 0.009433088302612304, 0.009385984420776367, 0.00932147216796875, 0.009388031959533692, 0.009335807800292969, 0.009325568199157714, 0.009361408233642577, 0.009371647834777832, 0.009327615737915039, 0.009338879585266113, 0.009379839897155762, 0.009380864143371583, 0.009395199775695801, 0.009388031959533692, 0.00940236759185791, 0.00934502410888672, 0.009341952323913574, 0.009369600296020507, 0.009391103744506836, 0.009352191925048828, 0.009370623588562011, 0.009443327903747559, 0.009361408233642577, 0.009338879585266113, 0.009364480018615723, 0.009339903831481934, 0.009395199775695801, 0.009384960174560546, 0.009341952323913574, 0.009366527557373047, 0.00942080020904541, 0.009370623588562011, 0.00942796802520752, 0.00932863998413086, 0.009354240417480468, 0.009360383987426758, 0.009390080451965332, 0.009333760261535644, 0.009400320053100587, 0.009369600296020507, 0.009364480018615723, 0.009368576049804688, 0.009360383987426758, 0.009443327903747559, 0.009326592445373535, 0.01901875114440918, 0.008952832221984864, 0.00899071979522705, 0.008967167854309082, 0.008939519882202148, 0.008987648010253906, 0.008966143608093263, 0.008965120315551758, 0.008954879760742187, 0.009019392013549805, 0.009109503746032714, 0.008960000038146973, 0.008952832221984864, 0.009220095634460449, 0.009861120223999023, 0.00948639965057373, 0.00944428825378418, 0.009376768112182618, 0.00932863998413086, 0.00941158390045166, 0.009361408233642577, 0.009380864143371583, 0.009380864143371583, 0.00933683204650879, 0.009382911682128906, 0.009383935928344727, 0.009369600296020507, 0.009364480018615723, 0.00942080020904541, 0.009350144386291503, 0.009341952323913574, 0.009348095893859864, 0.009343999862670899, 0.009343999862670899, 0.00932863998413086, 0.009337856292724609, 0.00932044792175293, 0.009356287956237793, 0.009367551803588867, 0.00934502410888672, 0.009392127990722657, 0.00929792022705078, 0.00931942367553711, 0.009338879585266113, 0.009375743865966797, 0.009399295806884766, 0.009381888389587402, 0.00935321617126465, 0.009390080451965332, 0.009349120140075684, 0.009335807800292969, 0.009380864143371583, 0.009374719619750976, 0.009377792358398437, 0.009381888389587402, 0.009395199775695801, 0.009673727989196777, 0.00941977596282959, 0.009387007713317871, 0.009326592445373535, 0.009343999862670899, 0.009375743865966797, 0.009371647834777832, 0.018974720001220705, 0.00899891185760498, 0.008985600471496581, 0.008976384162902832, 0.008964096069335938, 0.008963071823120117, 0.00900710391998291, 0.008983551979064941, 0.008967167854309082, 0.008986623764038085, 0.00899788761138916, 0.008945664405822755, 0.008963071823120117, 0.008977408409118653, 0.00902451229095459, 0.009000960350036622, 0.009065471649169921, 0.008977408409118653, 0.009055232048034668, 0.009011199951171875, 0.008955904006958008, 0.009010175704956054, 0.009030655860900879, 0.008962047576904298, 0.008970239639282226, 0.009018367767333984, 0.008996864318847657, 0.008975359916687011, 0.009067520141601563, 0.009383935928344727, 0.009449472427368164, 0.009466879844665528, 0.009407487869262696, 0.009392127990722657, 0.009387007713317871, 0.009497599601745605, 0.009343999862670899, 0.00940339183807373, 0.009377792358398437, 0.009355263710021973, 0.009339903831481934, 0.009409536361694336, 0.009327615737915039, 0.009347071647644043, 0.009401344299316406, 0.009355263710021973, 0.0093307523727417, 0.009383872032165527, 0.009399295806884766, 0.009380864143371583, 0.009392127990722657, 0.009382911682128906, 0.009409536361694336, 0.009367551803588867, 0.009356287956237793, 0.009332736015319825, 0.009358336448669433, 0.009363455772399902, 0.009395199775695801, 0.009376768112182618, 0.009335807800292969, 0.009364480018615723, 0.009350144386291503]",tokens/s,105.69506081719804,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-6.7b,,cuda,0,42,,,,,True,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,217063f5c507ed7cc255df7e1f64c4333a0b4dfe,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,,MB,1280.765952,15125.184512,0.0,14478.737408,14237.625344,s,10,0.7475148468017578,0.07475148468017578,0.000259700632932575,0.07466180801391602,0.07482695083618164,0.07517518653869629,0.075453775100708,"[0.07552342224121093, 0.07474956512451172, 0.07465363311767578, 0.07465500640869141, 0.0746383056640625, 0.07464227294921875, 0.07466860961914062, 0.07467030334472656, 0.0746146240234375, 0.07469910430908203]",tokens/s,3424.6811430608495,kWh,8.849784571238814e-07,4.849176747850821e-07,4.5043361076120084e-06,5.874232239520972e-06,tokens/kWh,43580163.255662516,MB,1280.765952,15125.184512,0.0,14478.737408,14440.977408,s,10,21.599800048828126,2.1599800048828124,0.003288007360278083,2.1589039306640627,2.160863452148438,2.1653070922851563,2.1688620043945313,"[2.169750732421875, 2.159219482421875, 2.158219970703125, 2.158555419921875, 2.1598759765625, 2.158583740234375, 2.1588251953125, 2.158982666015625, 2.158514404296875, 2.1592724609375]",tokens/s,29.166936664961398,kWh,2.54379255402444e-05,1.3940759493216047e-05,0.00011370850846258598,0.0001530871934960464,tokens/kWh,411530.1780721914,,s,629,22.236759033203107,0.035352558081404015,0.008907020984191575,0.034277374267578126,0.03439431686401367,0.03449446487426758,0.10923319305419922,"[0.03484364700317383, 0.03491123199462891, 0.03477503967285156, 0.03484262466430664, 0.03492764663696289, 0.03489379119873047, 0.03489894485473633, 0.034857982635498046, 0.03505254364013672, 0.034895870208740236, 0.03494400024414063, 0.03482112121582031, 0.034786304473876956, 0.03454771041870117, 0.03423539352416992, 0.03422208023071289, 0.034301952362060545, 0.034326526641845705, 0.03417292785644531, 0.034239486694335936, 0.03416883087158203, 0.03415961456298828, 0.03419443130493164, 0.034206718444824216, 0.03422412872314453, 0.0342108154296875, 0.03422719955444336, 0.03420569610595703, 0.034260990142822266, 0.03424358367919922, 0.034269184112548826, 0.034260990142822266, 0.0342845458984375, 0.03430092620849609, 0.0343040657043457, 0.03431622314453125, 0.034269184112548826, 0.034328575134277346, 0.03425075149536133, 0.03425996780395508, 0.0342999038696289, 0.03428659057617187, 0.03431423950195313, 0.03431628799438476, 0.03445248031616211, 0.03434905624389648, 0.03437363052368164, 0.03434700775146484, 0.034304000854492187, 0.03437158584594727, 0.034331649780273435, 0.03437363052368164, 0.03433779144287109, 0.03443711853027344, 0.03450368118286133, 0.034371646881103515, 0.034324417114257814, 0.03433779144287109, 0.03438079833984375, 0.0344279670715332, 0.03435103988647461, 0.03438387298583984, 0.10949734497070313, 0.034098175048828124, 0.034119678497314454, 0.034127872467041014, 0.034119678497314454, 0.03415961456298828, 0.0341473274230957, 0.034149375915527344, 0.03415347290039063, 0.03410739135742188, 0.03418931198120117, 0.03413913726806641, 0.034184192657470705, 0.03417190551757812, 0.03415347290039063, 0.03418316650390625, 0.03415961456298828, 0.034135040283203126, 0.03418828964233398, 0.03415552139282227, 0.03419340896606445, 0.03426303863525391, 0.034184192657470705, 0.034239486694335936, 0.03422719955444336, 0.03423129653930664, 0.03420774459838867, 0.03429580688476563, 0.03430604934692383, 0.034223102569580076, 0.034277374267578126, 0.034253822326660154, 0.034290687561035156, 0.034283519744873044, 0.03424563217163086, 0.03425689697265625, 0.034277374267578126, 0.03430809783935547, 0.03433779144287109, 0.03437977600097656, 0.03430297470092773, 0.03441766357421875, 0.034576385498046876, 0.0342999038696289, 0.034336769104003906, 0.034310142517089845, 0.034339839935302735, 0.03437363052368164, 0.03433884811401367, 0.0343592643737793, 0.03433881759643555, 0.03434086227416992, 0.03431731033325195, 0.034352191925048826, 0.034393024444580075, 0.034358272552490236, 0.034402305603027344, 0.034440193176269535, 0.03434086227416992, 0.03433062362670898, 0.034344959259033206, 0.03437977600097656, 0.034377727508544925, 0.10925465393066407, 0.03410128021240234, 0.03414934539794922, 0.034146305084228515, 0.03410739135742188, 0.03412582397460937, 0.034097152709960936, 0.03413302230834961, 0.034131935119628906, 0.03416883087158203, 0.0341258544921875, 0.034161632537841796, 0.03415654373168946, 0.034179073333740234, 0.03414527893066406, 0.034187263488769534, 0.03417804718017578, 0.03420569610595703, 0.03415859222412109, 0.0342108154296875, 0.03414527893066406, 0.03425689697265625, 0.03423027038574219, 0.03420159912109375, 0.03419750213623047, 0.034252799987792966, 0.03422719955444336, 0.03422515106201172, 0.034239486694335936, 0.03424563217163086, 0.03421804809570313, 0.03430495834350586, 0.034277374267578126, 0.03427020645141601, 0.03420876693725586, 0.03426508712768555, 0.03425484848022461, 0.034290687561035156, 0.0342845458984375, 0.03427328109741211, 0.03429478454589844, 0.03430912017822266, 0.034253822326660154, 0.03430912017822266, 0.03429683303833008, 0.034339839935302735, 0.0342999038696289, 0.03440639877319336, 0.03439206314086914, 0.03434188842773438, 0.03433168029785156, 0.03437257766723633, 0.03434598541259765, 0.03435417556762695, 0.034336769104003906, 0.034372608184814454, 0.03435520172119141, 0.03435724639892578, 0.03435212707519531, 0.034344959259033206, 0.03438284683227539, 0.0344002571105957, 0.03437055969238281, 0.10925772857666016, 0.03405516815185547, 0.03412275314331055, 0.034149375915527344, 0.03407257461547852, 0.0341739501953125, 0.03413094329833984, 0.03417292785644531, 0.03415654373168946, 0.034184192657470705, 0.03414425659179687, 0.03417497634887695, 0.034149375915527344, 0.0341104621887207, 0.034157569885253904, 0.03418828964233398, 0.034170879364013675, 0.03418316650390625, 0.03415859222412109, 0.03421491241455078, 0.03420774459838867, 0.03424563217163086, 0.034162689208984375, 0.03419750213623047, 0.03425075149536133, 0.03430297470092773, 0.03423542404174805, 0.03424764633178711, 0.0342210578918457, 0.0342927360534668, 0.03427328109741211, 0.03421286392211914, 0.034239486694335936, 0.03428147125244141, 0.03426508712768555, 0.03433267211914062, 0.03425075149536133, 0.034272254943847655, 0.03429785537719727, 0.03430092620849609, 0.03426611328125, 0.034318336486816405, 0.034282497406005856, 0.0342999038696289, 0.03435007858276367, 0.034305023193359374, 0.03428966522216797, 0.03432550430297852, 0.034369537353515625, 0.03428966522216797, 0.03436550521850586, 0.03433158493041992, 0.03431731033325195, 0.03437158584594727, 0.034318336486816405, 0.03435520172119141, 0.034405376434326174, 0.03436544036865234, 0.03432447814941406, 0.03438079833984375, 0.03449856185913086, 0.034459648132324217, 0.03441254425048828, 0.10924339294433594, 0.03409100723266602, 0.03411763381958008, 0.034100223541259765, 0.034113536834716796, 0.03410230255126953, 0.03408995056152344, 0.03415961456298828, 0.03414220809936523, 0.03415961456298828, 0.03418316650390625, 0.034103294372558594, 0.034164737701416016, 0.034200576782226565, 0.03439513778686523, 0.03429683303833008, 0.03422003173828125, 0.03417497634887695, 0.034249729156494144, 0.034216960906982424, 0.034377727508544925, 0.03458150482177735, 0.03419852828979492, 0.034200576782226565, 0.0342210578918457, 0.034203647613525394, 0.034336769104003906, 0.034202625274658206, 0.034301952362060545, 0.034229248046875, 0.03423231887817383, 0.03418521499633789, 0.03428761672973633, 0.034290687561035156, 0.034301952362060545, 0.03427840042114258, 0.034334720611572264, 0.034282497406005856, 0.034304000854492187, 0.03425894546508789, 0.03433574295043945, 0.03424665451049805, 0.034334720611572264, 0.03429584121704102, 0.0343408317565918, 0.03429580688476563, 0.03433574295043945, 0.034543617248535156, 0.03470131301879883, 0.03439411163330078, 0.03433779144287109, 0.03429478454589844, 0.03436236953735351, 0.03438489532470703, 0.03437158584594727, 0.03430809783935547, 0.034348033905029295, 0.034377727508544925, 0.03435212707519531, 0.03435520172119141, 0.03437875366210937, 0.034339839935302735, 0.0344002571105957, 0.10922803497314452, 0.03407360076904297, 0.03411251068115234, 0.03411558532714844, 0.034118656158447266, 0.03413913726806641, 0.03412275314331055, 0.03416064071655273, 0.034132991790771484, 0.03417292785644531, 0.03413913726806641, 0.034162689208984375, 0.03419443130493164, 0.03419647979736328, 0.034200576782226565, 0.03415039825439453, 0.03416579055786133, 0.03418211364746094, 0.03421798324584961, 0.03418316650390625, 0.034233345031738284, 0.034239486694335936, 0.03422719955444336, 0.03421286392211914, 0.03424051284790039, 0.034200576782226565, 0.03427020645141601, 0.03426303863525391, 0.03424563217163086, 0.03425689697265625, 0.03426201629638672, 0.03427328109741211, 0.034277374267578126, 0.03427328109741211, 0.03429785537719727, 0.03423235321044922, 0.03430499267578125, 0.03429580688476563, 0.034293758392333985, 0.03426508712768555, 0.034310142517089845, 0.03430604934692383, 0.034298881530761716, 0.034323455810546875, 0.034328575134277346, 0.03432447814941406, 0.03431731033325195, 0.034339839935302735, 0.03434188842773438, 0.03436544036865234, 0.034328575134277346, 0.0343111686706543, 0.034339839935302735, 0.0343818244934082, 0.03433881759643555, 0.034336769104003906, 0.034356224060058595, 0.03438079833984375, 0.03439513778686523, 0.03433779144287109, 0.03442176055908203, 0.03434188842773438, 0.03440127944946289, 0.10923519897460937, 0.03415244674682617, 0.03411049652099609, 0.03409814453125, 0.03412998580932617, 0.03419027328491211, 0.03410636901855469, 0.03419033432006836, 0.0344637451171875, 0.03425075149536133, 0.034164737701416016, 0.03414425659179687, 0.03419136047363281, 0.03421798324584961, 0.03420774459838867, 0.034146305084228515, 0.03416371154785156, 0.03421491241455078, 0.034111488342285154, 0.03423436737060547, 0.03419033432006836, 0.034239486694335936, 0.034184192657470705, 0.03421798324584961, 0.034202625274658206, 0.03422515106201172, 0.034181121826171876, 0.03425791931152344, 0.03425689697265625, 0.03424051284790039, 0.03420569610595703, 0.03425177764892578, 0.034252799987792966, 0.03426508712768555, 0.034226207733154296, 0.03426095962524414, 0.03429171371459961, 0.034318336486816405, 0.03422412872314453, 0.03430297470092773, 0.034282497406005856, 0.03431935882568359, 0.03426611328125, 0.034320384979248046, 0.034288639068603514, 0.03436851119995117, 0.034377761840820316, 0.034314208984375, 0.0343296012878418, 0.034385921478271485, 0.034304000854492187, 0.034374656677246096, 0.03436236953735351, 0.03439513778686523, 0.03433267211914062, 0.03433267211914062, 0.034358272552490236, 0.03436441421508789, 0.03431731033325195, 0.03436441421508789, 0.034369537353515625, 0.034423809051513675, 0.034418689727783204, 0.10927206420898437, 0.034062335968017575, 0.034129920959472655, 0.03410124969482422, 0.03409612655639648, 0.03416985702514649, 0.03413094329833984, 0.03412070465087891, 0.03414425659179687, 0.034088958740234376, 0.03414527893066406, 0.03418521499633789, 0.034181121826171876, 0.034148353576660156, 0.03416371154785156, 0.034187263488769534, 0.03418521499633789, 0.03425996780395508, 0.03419647979736328, 0.03418009567260742, 0.03420979309082031, 0.03440332794189453, 0.034533374786376955, 0.0341923828125, 0.03419443130493164, 0.03425996780395508, 0.03437670516967774, 0.03431423950195313, 0.03422208023071289, 0.03428761672973633, 0.034249729156494144, 0.03426406478881836, 0.03428659057617187, 0.03425894546508789, 0.03421593475341797, 0.034285568237304685, 0.03426611328125, 0.034237438201904294, 0.03430092620849609, 0.034321407318115234, 0.034260990142822266, 0.03431731033325195, 0.03428966522216797, 0.03427328109741211, 0.034269184112548826, 0.034307071685791016, 0.0343818244934082, 0.03433779144287109, 0.03434188842773438, 0.034334720611572264, 0.03437055969238281, 0.034334720611572264, 0.034323455810546875, 0.034339839935302735, 0.034342910766601564, 0.034348033905029295, 0.03448831939697266, 0.03439206314086914, 0.034339839935302735, 0.0343633918762207, 0.03435007858276367, 0.034356224060058595, 0.03442892837524414, 0.10916352081298829, 0.03413401412963867, 0.034078720092773435, 0.034121726989746096, 0.03408486557006836, 0.034146305084228515, 0.034103294372558594, 0.03413913726806641, 0.034135040283203126, 0.0341401596069336, 0.03416371154785156, 0.03412684631347656, 0.03415961456298828, 0.034119678497314454, 0.03416166305541992, 0.03416169738769531, 0.034210784912109375, 0.034141185760498044, 0.03420159912109375, 0.03414527893066406, 0.03422412872314453, 0.034195457458496094, 0.03421184158325195, 0.034195457458496094, 0.034249729156494144, 0.034216960906982424, 0.034344959259033206, 0.034405376434326174, 0.03427942276000977, 0.03424563217163086, 0.034290687561035156, 0.03421388626098633, 0.03424256134033203, 0.034277374267578126, 0.034310142517089845, 0.034245697021484375, 0.034280384063720706, 0.03423027038574219, 0.0342845458984375, 0.03423436737060547, 0.034299934387207034, 0.034287582397460936, 0.034326526641845705, 0.03426508712768555, 0.03439308929443359, 0.03458662414550781, 0.034525184631347655, 0.03434905624389648, 0.034339839935302735, 0.0342999038696289, 0.03434188842773438, 0.03433267211914062, 0.03434086227416992, 0.034339839935302735, 0.034359294891357424, 0.034307071685791016, 0.034348033905029295, 0.034318336486816405, 0.034348033905029295, 0.03433881759643555, 0.0344002571105957, 0.03432550430297852, 0.034375679016113284, 0.10929971313476562, 0.03405414581298828, 0.03410432052612305, 0.03409612655639648, 0.034135040283203126, 0.03409920120239258, 0.034154495239257815, 0.03415859222412109, 0.03416985702514649, 0.03415654373168946, 0.03416678237915039, 0.03412684631347656, 0.03419955062866211, 0.03427123260498047, 0.0343818244934082, 0.03427020645141601, 0.034170879364013675, 0.03421491241455078, 0.03416678237915039, 0.0342108154296875, 0.03421491241455078, 0.03421798324584961, 0.03422208023071289, 0.03418931198120117, 0.03422617721557617, 0.03422003173828125, 0.03427123260498047, 0.03423539352416992, 0.03428659057617187, 0.03430198287963867, 0.03428860855102539, 0.03420979309082031, 0.034282497406005856, 0.03431423950195313, 0.034305023193359374, 0.03427328109741211, 0.034255870819091795, 0.03426816177368164, 0.034274303436279296, 0.03429171371459961, 0.034290687561035156, 0.034260990142822266, 0.03443097686767578, 0.034353153228759765, 0.034328575134277346, 0.03427328109741211, 0.03432447814941406, 0.03437363052368164, 0.03430604934692383, 0.034323455810546875, 0.03433779144287109, 0.03434092712402344, 0.03447187042236328, 0.03434393692016602, 0.03442073440551758, 0.03433881759643555, 0.03439513778686523, 0.03445452880859375, 0.034353153228759765, 0.03435212707519531, 0.03438387298583984, 0.03437977600097656, 0.03439923095703125]",tokens/s,28.286496204811137,,,main,False,False,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-13b,facebook/opt-13b,cuda,0,42,,,True,,,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",opt,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-6.7b,,cuda,0,42,,,,,True,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,d35829e539df8480b726c647eeabf91e41eae047,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,,MB,1192.706048,14382.792704,0.0,13736.3456,13487.531008,s,10,0.6736084442138672,0.06736084442138672,0.0002525542761618945,0.06724551773071288,0.06760586013793946,0.06782964820861817,0.06800867866516114,"[0.06805343627929687, 0.06725020599365235, 0.06722144317626953, 0.06722035217285156, 0.06722418975830079, 0.0672086410522461, 0.06755612945556641, 0.06724082946777343, 0.06725161743164063, 0.06738159942626953]",tokens/s,3800.4274174259212,kWh,7.965395170125742e-07,4.3641060099177397e-07,4.3883946367381145e-06,5.621344754742463e-06,tokens/kWh,45540704.43447271,MB,1192.706048,14382.792704,0.0,13736.3456,13661.254656,s,10,19.870003173828128,1.9870003173828124,0.001308161958882297,1.9866483154296875,1.9882964599609374,1.9892427124023437,1.9899997143554689,"[1.99018896484375, 1.988086181640625, 1.98498486328125, 1.98658154296875, 1.9861776123046875, 1.986427734375, 1.9865947265625, 1.98747412109375, 1.986701904296875, 1.9867855224609374]",tokens/s,31.706084517883102,kWh,2.3405527987313928e-05,1.282685471910084e-05,0.00010697394445306076,0.00014320632715947553,tokens/kWh,439924.6964126297,,s,629,20.443699218749988,0.03250190654809221,0.008036461391132918,0.03153817558288574,0.031657990264892584,0.03169198112487793,0.09906618194580079,"[0.03139174461364746, 0.03147878456115723, 0.03156172752380371, 0.03142348861694336, 0.031454208374023435, 0.031442943572998046, 0.03150131225585937, 0.031455232620239255, 0.03145011138916016, 0.03141228866577148, 0.03145619201660156, 0.03143168067932129, 0.03148598480224609, 0.031468511581420895, 0.0316549129486084, 0.03184639930725098, 0.03200921630859375, 0.0318474235534668, 0.031547391891479495, 0.031475711822509765, 0.031494144439697266, 0.03146956825256347, 0.03148492813110351, 0.03150233650207519, 0.03157401657104492, 0.03155046463012695, 0.03151263999938965, 0.03151865577697754, 0.03153817558288574, 0.03150233650207519, 0.03153510475158691, 0.03154742431640625, 0.03157705688476563, 0.031559680938720705, 0.03156070327758789, 0.03156480026245117, 0.031591424942016604, 0.03154332733154297, 0.03157910346984863, 0.031579135894775394, 0.03159552001953125, 0.03159040069580078, 0.03161087989807129, 0.0316753921508789, 0.031692800521850584, 0.03162732887268067, 0.031632320404052734, 0.03161497688293457, 0.031628288269042966, 0.03162112045288086, 0.03166924858093262, 0.03167033576965332, 0.03166201591491699, 0.03164262390136719, 0.031747072219848634, 0.031752191543579104, 0.03167334365844727, 0.03168767929077149, 0.03163852882385254, 0.03165798377990723, 0.03169075202941894, 0.03166521644592285, 0.09908729553222656, 0.03136614418029785, 0.03138252830505371, 0.03139174461364746, 0.03137126350402832, 0.03139788818359375, 0.03159756851196289, 0.03200307083129883, 0.031575040817260744, 0.031497215270996096, 0.03148492813110351, 0.031461408615112305, 0.03139785575866699, 0.031455232620239255, 0.03140201568603516, 0.031416288375854494, 0.031455232620239255, 0.031460351943969726, 0.03146137619018555, 0.031437824249267575, 0.03146342468261719, 0.031453184127807614, 0.031441919326782225, 0.03158732795715332, 0.03161907196044922, 0.03153919982910156, 0.03149516868591309, 0.031509504318237305, 0.03149516868591309, 0.031475711822509765, 0.03150028800964356, 0.03147776031494141, 0.03151667213439941, 0.031541248321533204, 0.03151974487304687, 0.03149926376342774, 0.031531007766723636, 0.031543296813964845, 0.03154841613769531, 0.03161907196044922, 0.03155251121520996, 0.03156684875488281, 0.031601696014404296, 0.03161596870422363, 0.03155251121520996, 0.03158016014099121, 0.03159654426574707, 0.03159040069580078, 0.03159552001953125, 0.03159555244445801, 0.031607776641845706, 0.031604736328125, 0.031665151596069335, 0.03163033676147461, 0.03159859275817871, 0.03167027282714844, 0.03164262390136719, 0.031623231887817384, 0.03163852882385254, 0.031638463973999026, 0.03161702346801758, 0.03202969741821289, 0.03202252960205078, 0.09912422180175781, 0.03140096092224121, 0.031352832794189454, 0.03136614418029785, 0.03136511993408203, 0.0313753604888916, 0.0313856315612793, 0.03141423988342285, 0.031388671875, 0.03137945556640625, 0.03140300750732422, 0.03141529655456543, 0.03138252830505371, 0.03142348861694336, 0.031421440124511715, 0.03143270492553711, 0.031415327072143555, 0.03142243194580078, 0.03143577575683594, 0.031425535202026365, 0.03141529655456543, 0.03144601631164551, 0.03142758369445801, 0.03144806480407715, 0.031441919326782225, 0.031455232620239255, 0.031476736068725586, 0.031492095947265625, 0.03147980880737305, 0.031508480072021484, 0.03156377601623535, 0.031664127349853514, 0.03156684875488281, 0.031508480072021484, 0.031514623641967776, 0.031575040817260744, 0.031529983520507815, 0.03155353546142578, 0.031541248321533204, 0.0315228157043457, 0.031525888442993165, 0.03153203201293946, 0.03153408050537109, 0.031529983520507815, 0.031591424942016604, 0.03157401657104492, 0.03159654426574707, 0.03160166358947754, 0.0315863037109375, 0.03157811164855957, 0.0315729923248291, 0.03157606315612793, 0.0315729923248291, 0.03157708740234375, 0.03160681533813477, 0.03164975929260254, 0.03162521553039551, 0.03162009620666504, 0.03159859275817871, 0.03159449577331543, 0.03158425521850586, 0.03159654426574707, 0.03161087989807129, 0.09906175994873047, 0.0313702392578125, 0.03139993667602539, 0.031325183868408206, 0.03136000061035156, 0.03138662338256836, 0.031409151077270506, 0.03139891242980957, 0.03141119956970215, 0.03138457679748535, 0.03141836738586426, 0.03139174461364746, 0.03139481544494629, 0.031442943572998046, 0.031409151077270506, 0.03146342468261719, 0.03142963218688965, 0.031442943572998046, 0.031457279205322264, 0.03146444892883301, 0.03148595237731933, 0.031497215270996096, 0.03148185539245606, 0.03143168067932129, 0.0314521598815918, 0.031494144439697266, 0.031453184127807614, 0.031494144439697266, 0.031498239517211916, 0.031509504318237305, 0.03150233650207519, 0.031512575149536134, 0.03153919982910156, 0.03155251121520996, 0.03150441551208496, 0.03152892875671387, 0.03157708740234375, 0.03157811164855957, 0.03159756851196289, 0.03159449577331543, 0.03163443183898926, 0.03154944038391113, 0.03155046463012695, 0.03158937644958496, 0.03158527946472168, 0.03158937644958496, 0.03158732795715332, 0.03159244728088379, 0.03159859275817871, 0.03162521553039551, 0.03164262390136719, 0.03161395263671875, 0.031666175842285156, 0.031734783172607424, 0.031627264022827145, 0.03164262390136719, 0.03179827117919922, 0.03181875228881836, 0.031697919845581055, 0.03163852882385254, 0.03163750457763672, 0.03167231941223145, 0.031627264022827145, 0.09906790161132813, 0.031389696121215824, 0.03135078430175781, 0.03138662338256836, 0.03139379119873047, 0.03141324806213379, 0.031547391891479495, 0.03158937644958496, 0.03138150405883789, 0.031410175323486327, 0.03142963218688965, 0.03141119956970215, 0.03142560005187988, 0.031449024200439456, 0.03146956825256347, 0.0314450569152832, 0.03142854309082031, 0.03147776031494141, 0.03141119956970215, 0.031480831146240236, 0.03143475151062012, 0.03144806480407715, 0.031474687576293944, 0.031456256866455076, 0.03144499206542969, 0.03148902320861816, 0.0314726390838623, 0.031488000869750975, 0.03144908714294434, 0.031503360748291014, 0.031508480072021484, 0.03150028800964356, 0.031525888442993165, 0.03153715133666992, 0.03158016014099121, 0.03167334365844727, 0.03153919982910156, 0.031575040817260744, 0.031575040817260744, 0.03156076812744141, 0.03154323196411133, 0.03155353546142578, 0.03156377601623535, 0.03156582450866699, 0.03157196807861328, 0.03156588745117188, 0.031603647232055665, 0.03156582450866699, 0.03158732795715332, 0.03161907196044922, 0.03158323287963867, 0.03164262390136719, 0.03158432006835937, 0.031652799606323244, 0.03164159965515137, 0.0316231689453125, 0.03161087989807129, 0.031665151596069335, 0.031615999221801756, 0.03159756851196289, 0.03160576057434082, 0.03161292839050293, 0.031666175842285156, 0.09904742431640624, 0.03136409568786621, 0.03138252830505371, 0.031339519500732424, 0.03135385513305664, 0.03144806480407715, 0.03139583969116211, 0.03142457580566406, 0.03139065551757812, 0.03141324806213379, 0.03146240043640137, 0.031441919326782225, 0.03142860794067383, 0.0314204158782959, 0.03144806480407715, 0.031437824249267575, 0.03143065643310547, 0.03145011138916016, 0.03143475151062012, 0.03148287963867188, 0.031458303451538085, 0.0315043830871582, 0.031456256866455076, 0.03144499206542969, 0.03142963218688965, 0.031509504318237305, 0.03147372817993164, 0.03150943946838379, 0.031508480072021484, 0.031514623641967776, 0.0314839038848877, 0.03146854400634766, 0.03151872062683105, 0.031535135269165036, 0.031506399154663085, 0.03158016014099121, 0.03155046463012695, 0.03163443183898926, 0.03157094383239746, 0.03156172752380371, 0.03168358421325684, 0.03168870353698731, 0.03152383995056152, 0.03157094383239746, 0.03155558395385742, 0.0315863037109375, 0.03160166358947754, 0.03159756851196289, 0.031591424942016604, 0.03159756851196289, 0.03162214469909668, 0.0316231689453125, 0.03158323287963867, 0.03162931251525879, 0.0315996150970459, 0.03165286445617676, 0.03163340759277344, 0.03167027282714844, 0.031663103103637694, 0.031632383346557616, 0.03163552093505859, 0.03165996742248535, 0.03170816040039062, 0.09976012420654297, 0.031454208374023435, 0.03139379119873047, 0.03139276885986328, 0.03138355255126953, 0.03136511993408203, 0.03139891242980957, 0.03139174461364746, 0.031399999618530273, 0.03145209693908691, 0.031406080245971676, 0.03146240043640137, 0.03140096092224121, 0.03144499206542969, 0.03143577575683594, 0.03143065643310547, 0.031470592498779294, 0.03145011138916016, 0.03150131225585937, 0.031442943572998046, 0.03143475151062012, 0.0314521598815918, 0.03148595237731933, 0.031471616744995115, 0.031455232620239255, 0.03149516868591309, 0.031457279205322264, 0.031475711822509765, 0.031470592498779294, 0.03151769638061523, 0.031503360748291014, 0.031527936935424807, 0.031513599395751955, 0.03155459213256836, 0.031517663955688474, 0.03159449577331543, 0.03156172752380371, 0.031547391891479495, 0.031562751770019534, 0.03156070327758789, 0.03158220863342285, 0.031558656692504884, 0.0315361270904541, 0.0315228157043457, 0.03155046463012695, 0.03160780715942383, 0.03159040069580078, 0.03178291130065918, 0.03182080078125, 0.03165801620483399, 0.031572959899902345, 0.03167436790466309, 0.031611904144287106, 0.03163852882385254, 0.031647743225097655, 0.0315996150970459, 0.03160063934326172, 0.03161907196044922, 0.03160780715942383, 0.031643648147583005, 0.03161087989807129, 0.031649791717529296, 0.031632383346557616, 0.09910886383056641, 0.03138150405883789, 0.0314071044921875, 0.03135897636413574, 0.03136716842651367, 0.03142963218688965, 0.03143475151062012, 0.031421440124511715, 0.03140403175354004, 0.03141222381591797, 0.03143577575683594, 0.031389696121215824, 0.03139686393737793, 0.0314521598815918, 0.031426559448242186, 0.031510528564453126, 0.03149004745483398, 0.031457279205322264, 0.03144601631164551, 0.03152076721191406, 0.031457279205322264, 0.03157708740234375, 0.03172761535644531, 0.03187711906433106, 0.031529983520507815, 0.03153510475158691, 0.03146342468261719, 0.03148287963867188, 0.03147980880737305, 0.031541248321533204, 0.03152076721191406, 0.03151667213439941, 0.03154944038391113, 0.03156684875488281, 0.03152383995056152, 0.03153715133666992, 0.03155251121520996, 0.03159756851196289, 0.0315729923248291, 0.03158527946472168, 0.03160783958435059, 0.031556575775146485, 0.031541248321533204, 0.03158835220336914, 0.031558656692504884, 0.03158323287963867, 0.03159654426574707, 0.031661056518554685, 0.03171327972412109, 0.03164262390136719, 0.03163340759277344, 0.03160985565185547, 0.03162419128417969, 0.031628288269042966, 0.031604736328125, 0.03163750457763672, 0.03163955116271973, 0.03164159965515137, 0.03167647933959961, 0.031638463973999026, 0.031616031646728514, 0.03169487953186035, 0.031621055603027345, 0.09908428955078125, 0.03140918350219726, 0.031331296920776366, 0.03136819267272949, 0.03139481544494629, 0.03138764762878418, 0.03138457679748535, 0.03139174461364746, 0.0314019832611084, 0.031445024490356445, 0.0314050235748291, 0.03144908714294434, 0.03142451286315918, 0.03150028800964356, 0.03148595237731933, 0.03143270492553711, 0.03142860794067383, 0.03146751976013184, 0.03142451286315918, 0.03145011138916016, 0.0314337272644043, 0.031441919326782225, 0.03146137619018555, 0.031475711822509765, 0.03148595237731933, 0.03146444892883301, 0.031474687576293944, 0.0314726390838623, 0.031480831146240236, 0.03151872062683105, 0.0315729923248291, 0.03156787109375, 0.03154431915283203, 0.03152179145812988, 0.03156172752380371, 0.03151155281066895, 0.03150643157958984, 0.031562751770019534, 0.03154841613769531, 0.03154841613769531, 0.0315228157043457, 0.03158323287963867, 0.031559680938720705, 0.03156070327758789, 0.03158016014099121, 0.031645696640014646, 0.03160166358947754, 0.03160166358947754, 0.0317890567779541, 0.03191500854492187, 0.03163033676147461, 0.03160678482055664, 0.03164672088623047, 0.031628288269042966, 0.03163955116271973, 0.031662080764770506, 0.03161497688293457, 0.03163852882385254, 0.03161087989807129, 0.03161907196044922, 0.031628288269042966, 0.031664127349853514, 0.031714303970336914, 0.09944064331054688, 0.03139481544494629, 0.03144396781921387, 0.031352832794189454, 0.0313702392578125, 0.03138047981262207, 0.03140812873840332, 0.03138665580749512, 0.03141321563720703, 0.03141324806213379, 0.031466527938842775, 0.03147772789001465, 0.03141939163208008, 0.03143168067932129, 0.03142758369445801, 0.0314768009185791, 0.03142035293579101, 0.03143680000305176, 0.03144806480407715, 0.031477792739868164, 0.03144803237915039, 0.03148902320861816, 0.03144499206542969, 0.031525888442993165, 0.03147878456115723, 0.03148287963867188, 0.0314654712677002, 0.031528959274291994, 0.031498239517211916, 0.0315545597076416, 0.03154431915283203, 0.031546367645263675, 0.03153715133666992, 0.03155660820007324, 0.03152179145812988, 0.03159654426574707, 0.031591424942016604, 0.031575040817260744, 0.03157196807861328, 0.03157606315612793, 0.031581216812133786, 0.031552480697631835, 0.0315996150970459, 0.03160166358947754, 0.03155353546142578, 0.03160883140563965, 0.03160268783569336, 0.03161395263671875, 0.03159859275817871, 0.03161087989807129, 0.0316753921508789, 0.031649791717529296, 0.03160268783569336, 0.031627264022827145, 0.031665151596069335, 0.031649791717529296, 0.03161087989807129, 0.031648767471313476, 0.03165798377990723, 0.031665151596069335, 0.03160678482055664, 0.03167231941223145, 0.03162419128417969]",tokens/s,30.767425859166956,,,main,False,False,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,microsoft/rho-math-1b-v0.1,microsoft/rho-math-1b-v0.1,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1281.41312,3041.394688,0.0,2394.947584,2279.417856,s,10,0.23879766464233396,0.023879766464233395,0.0019186688840044529,0.023459343910217284,0.024200454521179197,0.026859299564361564,0.02898637559890747,"[0.029518144607543945, 0.02255036735534668, 0.023508031845092772, 0.022488000869750977, 0.023500608444213866, 0.02360960006713867, 0.023418079376220702, 0.023249856948852538, 0.023357599258422852, 0.023597375869750976]",tokens/s,10720.372847172997,kWh,2.755448700726511e-07,1.5098588436768086e-07,1.0324639184335585e-06,1.4589946728738905e-06,tokens/kWh,175463286.30229864,MB,1281.41312,3041.394688,0.0,2394.947584,2279.420416,s,10,14.219128295898438,1.4219128295898438,0.017921205502425246,1.4310108642578125,1.4332750732421875,1.4344568725585938,1.4354023120117188,"[1.432612060546875, 1.3843092041015626, 1.435638671875, 1.42259716796875, 1.433012451171875, 1.42940966796875, 1.389419921875, 1.4264703369140626, 1.43274169921875, 1.4329171142578125]",tokens/s,44.30651351403348,kWh,1.681395282529389e-05,9.213966185877955e-06,3.6267677032765595e-05,6.229559604393743e-05,tokens/kWh,1011307.4438771843,,s,629,14.40284059143066,0.022897997760621087,0.002853396107912129,0.022707199096679686,0.022890904998779295,0.023071335220336914,0.04606173294067386,"[0.023362560272216795, 0.02249113655090332, 0.022731775283813475, 0.022437887191772463, 0.02246143913269043, 0.022460416793823244, 0.022551551818847656, 0.022419456481933595, 0.022502399444580077, 0.022548479080200197, 0.022486015319824217, 0.022541311264038084, 0.023034879684448242, 0.02268876838684082, 0.02254240036010742, 0.02255558395385742, 0.02253209686279297, 0.022450176239013672, 0.022443008422851563, 0.022403072357177735, 0.022277151107788086, 0.021957599639892578, 0.02185523223876953, 0.02195769691467285, 0.023056320190429688, 0.023190528869628906, 0.02307276725769043, 0.022871040344238282, 0.022906879425048828, 0.022733823776245117, 0.022961151123046874, 0.02283417510986328, 0.022779903411865234, 0.022882303237915038, 0.022823936462402345, 0.022898687362670898, 0.022755327224731444, 0.023018495559692383, 0.022730752944946288, 0.022622207641601562, 0.022715391159057616, 0.022692895889282225, 0.022666208267211913, 0.022689792633056642, 0.022798336029052735, 0.02269491195678711, 0.02272153663635254, 0.02275328063964844, 0.022786048889160155, 0.022626304626464845, 0.022771711349487304, 0.0228351993560791, 0.022767616271972657, 0.02273689651489258, 0.022757375717163086, 0.022278144836425783, 0.022647808074951172, 0.022813695907592774, 0.022773759841918945, 0.023772159576416017, 0.02286591911315918, 0.022733823776245117, 0.04492390441894531, 0.02185420799255371, 0.02185420799255371, 0.021777408599853516, 0.02188390350341797, 0.02148044776916504, 0.02164735984802246, 0.021833728790283204, 0.021857280731201172, 0.021805055618286134, 0.02187571144104004, 0.021824512481689453, 0.021768192291259765, 0.02187059211730957, 0.021825536727905274, 0.021836799621582033, 0.021857280731201172, 0.02187264060974121, 0.021778432846069336, 0.02170675277709961, 0.02149478340148926, 0.02150297546386719, 0.02149580764770508, 0.021712896347045898, 0.02182143974304199, 0.02192793655395508, 0.021801984786987305, 0.02184806442260742, 0.021817344665527344, 0.02186240005493164, 0.021901376724243166, 0.021828544616699218, 0.02186649513244629, 0.02186751937866211, 0.02165350341796875, 0.021812223434448243, 0.021592063903808592, 0.021549055099487305, 0.021523456573486328, 0.021542911529541017, 0.02184499168395996, 0.021808128356933593, 0.02187468719482422, 0.02227507209777832, 0.0218470401763916, 0.02185318374633789, 0.021798912048339843, 0.02185318374633789, 0.021798912048339843, 0.022038528442382813, 0.022045759201049803, 0.021880767822265626, 0.02189619255065918, 0.021935104370117187, 0.022813695907592774, 0.02350796890258789, 0.02306662368774414, 0.022940671920776368, 0.02307072067260742, 0.02286489677429199, 0.022813695907592774, 0.02271027183532715, 0.022791168212890626, 0.046709758758544925, 0.022777856826782225, 0.022684671401977538, 0.02352639961242676, 0.022934560775756837, 0.022803424835205078, 0.02241433525085449, 0.022648832321166993, 0.022815744400024415, 0.022734848022460938, 0.022767616271972657, 0.02286796760559082, 0.022768640518188478, 0.02286899185180664, 0.02251366424560547, 0.022581247329711913, 0.02272972869873047, 0.02287718391418457, 0.022759424209594727, 0.022723583221435546, 0.022758399963378906, 0.022784000396728517, 0.02268262481689453, 0.023182336807250976, 0.022725631713867187, 0.0228853759765625, 0.02284851264953613, 0.022384639739990234, 0.02244095993041992, 0.022840320587158205, 0.022906879425048828, 0.022793216705322264, 0.022791168212890626, 0.022610944747924806, 0.022487039566040038, 0.0228351993560791, 0.022799360275268556, 0.022820863723754883, 0.02281062316894531, 0.022930431365966796, 0.022794240951538085, 0.022779903411865234, 0.022725631713867187, 0.02287308883666992, 0.022794240951538085, 0.022777856826782225, 0.02283622360229492, 0.02282803153991699, 0.022697984695434572, 0.02234982490539551, 0.022738943099975584, 0.022774784088134766, 0.02285260772705078, 0.022773759841918945, 0.022821887969970703, 0.023173120498657225, 0.022930431365966796, 0.022800384521484376, 0.02271232032775879, 0.02287615966796875, 0.022839296340942384, 0.022839296340942384, 0.022759424209594727, 0.04511948776245117, 0.022977535247802734, 0.02328371238708496, 0.022558719635009765, 0.022576128005981445, 0.022527999877929687, 0.022632448196411133, 0.02246451187133789, 0.022587423324584962, 0.02348441505432129, 0.022849599838256837, 0.022773664474487306, 0.02265395164489746, 0.022495231628417968, 0.022129663467407225, 0.022198272705078126, 0.022073375701904298, 0.022494304656982423, 0.02254323196411133, 0.022435840606689454, 0.022429695129394533, 0.022487039566040038, 0.022543359756469726, 0.022477888107299806, 0.02256278419494629, 0.022457311630249024, 0.02239897537231445, 0.02244607925415039, 0.022386688232421875, 0.022451200485229493, 0.022406143188476564, 0.022302879333496093, 0.02272751998901367, 0.022502399444580077, 0.022467584609985353, 0.021970943450927736, 0.021485567092895508, 0.022544384002685547, 0.02253107261657715, 0.02247475242614746, 0.021906431198120118, 0.022520832061767578, 0.022528032302856445, 0.022576095581054688, 0.022777856826782225, 0.022726655960083008, 0.022755327224731444, 0.023173120498657225, 0.022930431365966796, 0.022750207901000977, 0.02273689651489258, 0.022737920761108397, 0.022761472702026365, 0.02249830436706543, 0.022364160537719727, 0.022716415405273437, 0.022766592025756836, 0.022773759841918945, 0.022760448455810548, 0.022846464157104493, 0.02275328063964844, 0.023029760360717775, 0.022812671661376953, 0.04679884719848633, 0.022707199096679686, 0.022794240951538085, 0.02272870445251465, 0.02272051239013672, 0.022742015838623047, 0.022366207122802736, 0.022768640518188478, 0.022740991592407226, 0.022757375717163086, 0.02272051239013672, 0.022764543533325195, 0.02268262481689453, 0.02282803153991699, 0.022773759841918945, 0.022920223236083986, 0.023918560028076172, 0.023290880203247072, 0.023020544052124024, 0.02292633628845215, 0.022974464416503908, 0.022872064590454103, 0.02271334457397461, 0.02267852783203125, 0.022708223342895507, 0.022812671661376953, 0.02279225540161133, 0.023296991348266603, 0.022676479339599608, 0.022615007400512695, 0.022527999877929687, 0.022212608337402344, 0.02231091117858887, 0.022579200744628908, 0.022795328140258787, 0.022712255477905275, 0.022737920761108397, 0.02272153663635254, 0.022750207901000977, 0.022370304107666016, 0.022379520416259766, 0.022626304626464845, 0.022742015838623047, 0.022814720153808594, 0.022648832321166993, 0.022609920501708985, 0.022768640518188478, 0.02243891143798828, 0.02292633628845215, 0.022381568908691408, 0.022460416793823244, 0.022855680465698244, 0.022772768020629882, 0.02273583984375, 0.02274508857727051, 0.022743040084838868, 0.022421503067016603, 0.022591487884521484, 0.02288844871520996, 0.022746112823486327, 0.02274406433105469, 0.022837247848510742, 0.02285670471191406, 0.04677836990356445, 0.0227194881439209, 0.022603776931762694, 0.022898687362670898, 0.022821887969970703, 0.02274406433105469, 0.022702112197875976, 0.023112672805786133, 0.022724607467651366, 0.022762496948242186, 0.022746112823486327, 0.02273587226867676, 0.02281062316894531, 0.02287820816040039, 0.022768640518188478, 0.022734848022460938, 0.022358015060424806, 0.022375423431396483, 0.022350847244262697, 0.022346752166748047, 0.022593536376953126, 0.02286591911315918, 0.022774784088134766, 0.022815744400024415, 0.02267852783203125, 0.022743072509765625, 0.02273072052001953, 0.02265497589111328, 0.022746112823486327, 0.02272051239013672, 0.02270310401916504, 0.022689792633056642, 0.02274406433105469, 0.022789119720458984, 0.0227061767578125, 0.022796287536621093, 0.02275542449951172, 0.02264259147644043, 0.02262118339538574, 0.022563840866088865, 0.022441984176635742, 0.022408191680908202, 0.022353055953979493, 0.022639455795288085, 0.022743040084838868, 0.0227194881439209, 0.022687744140625, 0.022760448455810548, 0.022658048629760744, 0.022676479339599608, 0.022755327224731444, 0.023144447326660156, 0.02266214370727539, 0.022600704193115235, 0.022657024383544923, 0.02233241653442383, 0.022551551818847656, 0.022384639739990234, 0.022495231628417968, 0.022747135162353514, 0.02271232032775879, 0.022805503845214844, 0.022817792892456053, 0.046637054443359374, 0.02272051239013672, 0.02269081687927246, 0.0227194881439209, 0.02274406433105469, 0.022716415405273437, 0.022730752944946288, 0.022673408508300782, 0.022730752944946288, 0.022718463897705078, 0.022718463897705078, 0.022790143966674805, 0.022756351470947265, 0.02268671989440918, 0.02275225639343262, 0.02271129608154297, 0.022725631713867187, 0.02285260772705078, 0.02287308883666992, 0.02265395164489746, 0.02184499168395996, 0.021757951736450197, 0.021745664596557617, 0.021933055877685546, 0.021767168045043944, 0.021810176849365235, 0.021711872100830077, 0.021535743713378908, 0.021582847595214845, 0.021521408081054686, 0.02145075225830078, 0.02187468719482422, 0.02143436813354492, 0.02145484733581543, 0.021525503158569336, 0.021622783660888673, 0.021687295913696288, 0.021629951477050782, 0.02147737693786621, 0.021622783660888673, 0.021679103851318358, 0.021388288497924804, 0.02166783905029297, 0.0214835205078125, 0.02166988754272461, 0.021483615875244142, 0.021405664443969727, 0.021476287841796875, 0.021429376602172853, 0.021533567428588866, 0.02146201515197754, 0.02153267288208008, 0.0214517765045166, 0.021493824005126953, 0.02159916877746582, 0.021519359588623048, 0.021604352951049805, 0.021626880645751953, 0.022758399963378906, 0.02306355285644531, 0.022715391159057616, 0.022569984436035157, 0.02266214370727539, 0.04642816162109375, 0.02273587226867676, 0.022692863464355468, 0.022771711349487304, 0.022775808334350587, 0.02277894401550293, 0.02258937644958496, 0.0226693115234375, 0.022746112823486327, 0.022718463897705078, 0.022742015838623047, 0.022755327224731444, 0.022738943099975584, 0.023169023513793945, 0.022759424209594727, 0.022755327224731444, 0.02285670471191406, 0.022759424209594727, 0.022822912216186524, 0.022793216705322264, 0.022733823776245117, 0.022709247589111328, 0.022754304885864256, 0.022792192459106447, 0.022759424209594727, 0.022779903411865234, 0.02269900894165039, 0.02264575958251953, 0.022820863723754883, 0.02268262481689453, 0.02250547218322754, 0.022483968734741212, 0.022610944747924806, 0.022633472442626954, 0.022542335510253905, 0.022031360626220704, 0.022311935424804686, 0.02267955207824707, 0.022475776672363282, 0.02250649642944336, 0.022452224731445314, 0.022567935943603516, 0.0224849910736084, 0.022397951126098634, 0.02250351905822754, 0.022749088287353517, 0.022381568908691408, 0.02243174362182617, 0.022571008682250978, 0.02248089599609375, 0.022640640258789063, 0.022564863204956053, 0.02251263999938965, 0.022549503326416014, 0.022584320068359375, 0.022519807815551757, 0.022467584609985353, 0.02282700729370117, 0.022468639373779298, 0.022391935348510743, 0.02254332733154297, 0.02267430305480957, 0.022750207901000977, 0.04662988662719727, 0.022409215927124023, 0.022370367050170897, 0.02230076789855957, 0.02207027244567871, 0.02236911964416504, 0.022517759323120116, 0.022584352493286133, 0.0226856632232666, 0.02247987174987793, 0.02240716743469238, 0.02245631980895996, 0.022478847503662108, 0.02252288055419922, 0.022935552597045897, 0.02267238426208496, 0.02272870445251465, 0.022811647415161132, 0.02268671989440918, 0.022700031280517577, 0.022730752944946288, 0.022831104278564454, 0.022714368820190428, 0.022751232147216797, 0.02275328063964844, 0.02272972869873047, 0.02289254379272461, 0.022788095474243163, 0.02294790458679199, 0.022905792236328125, 0.022761472702026365, 0.022801408767700194, 0.022777856826782225, 0.02285977554321289, 0.022767616271972657, 0.022790143966674805, 0.022833152770996092, 0.02310553550720215, 0.022748159408569335, 0.02265395164489746, 0.022647808074951172, 0.02270515251159668, 0.02272051239013672, 0.022722560882568358, 0.022766592025756836, 0.0226375675201416, 0.022649856567382814, 0.022717439651489257, 0.02285158348083496, 0.02291814422607422, 0.024172544479370117, 0.023150592803955077, 0.022890495300292968, 0.02286591911315918, 0.02267852783203125, 0.022701055526733398, 0.022684671401977538, 0.022740991592407226, 0.02281062316894531, 0.022697984695434572, 0.02268262481689453, 0.022972415924072266, 0.022773759841918945, 0.04684697723388672, 0.022648832321166993, 0.02265907287597656, 0.022709247589111328, 0.022737920761108397, 0.022675455093383787, 0.02269593620300293, 0.02263039970397949, 0.022725664138793945, 0.022663135528564454, 0.022797311782836914, 0.02272870445251465, 0.022543359756469726, 0.022708223342895507, 0.02269388771057129, 0.02305638313293457, 0.023341056823730468, 0.02262835121154785, 0.022754304885864256, 0.02268262481689453, 0.022762496948242186, 0.02271129608154297, 0.02268262481689453, 0.022586368560791017, 0.022822912216186524, 0.022692863464355468, 0.022929407119750975, 0.022788095474243163, 0.022575103759765625, 0.0227194881439209, 0.022781951904296875, 0.02287308883666992, 0.022782976150512696, 0.02267955207824707, 0.022747135162353514, 0.022739967346191405, 0.02265497589111328, 0.02277894401550293, 0.022729663848876952, 0.022737920761108397, 0.0226375675201416, 0.02307174491882324, 0.02280243110656738, 0.022692863464355468, 0.0226693115234375, 0.022649856567382814, 0.022598655700683593, 0.022725631713867187, 0.022765567779541016, 0.022796287536621093, 0.02269696044921875, 0.02262835121154785, 0.022590463638305663, 0.022832128524780275, 0.022590463638305663, 0.022388736724853517, 0.02285260772705078, 0.022785024642944338, 0.02274508857727051, 0.022776832580566408, 0.022692863464355468, 0.023166976928710937, 0.0227194881439209]",tokens/s,43.67194068468962,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-2.7b,,cuda,0,42,,,,,True,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,217063f5c507ed7cc255df7e1f64c4333a0b4dfe,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,,MB,1278.517248,6797.39392,0.0,6150.946816,5988.307968,s,10,0.30529612541198725,0.030529612541198726,0.00029578028387582366,0.0304357271194458,0.030562937927246094,0.030988860702514647,0.03132959892272949,"[0.031414783477783204, 0.030425983428955077, 0.030416704177856444, 0.030448543548583985, 0.030395135879516602, 0.030442911148071287, 0.03040652847290039, 0.030428543090820312, 0.03044870376586914, 0.03046828842163086]",tokens/s,8385.301308836995,kWh,3.621152406884342e-07,1.983538829961684e-07,1.7911295334801126e-06,2.351598657164715e-06,tokens/kWh,108862113.5328658,MB,1278.517248,6797.39392,0.0,6150.946816,6117.056512,s,10,13.401929077148438,1.3401929077148438,0.0071886328834873696,1.3391688232421877,1.348996240234375,1.3506322143554688,1.3519409936523439,"[1.3486326904296875, 1.3366287841796876, 1.33829296875, 1.3279244384765625, 1.3390897216796875, 1.331222900390625, 1.3392479248046876, 1.3414462890625, 1.3522681884765626, 1.3471751708984374]",tokens/s,47.0081580325783,kWh,1.6584539956587576e-05,9.086562614075947e-06,5.5923666638320975e-05,8.15947692089845e-05,tokens/kWh,772108.317858481,,s,629,13.652671457290648,0.021705360027489108,0.0036889444810717584,0.021188608169555666,0.021647769546508787,0.021866086959838868,0.052123201904296874,"[0.02315158462524414, 0.021507072448730468, 0.021169151306152344, 0.021189632415771483, 0.02108723258972168, 0.021118976593017577, 0.02105958366394043, 0.021150720596313476, 0.02145996856689453, 0.02214297676086426, 0.021568511962890623, 0.021142528533935546, 0.021331968307495116, 0.02123263931274414, 0.021134336471557616, 0.021182464599609374, 0.021183551788330077, 0.02118137550354004, 0.021149696350097655, 0.021133312225341795, 0.021165056228637694, 0.02110361671447754, 0.021190656661987304, 0.021172224044799806, 0.021176319122314453, 0.02168934440612793, 0.02230784034729004, 0.022359039306640623, 0.022817792892456053, 0.021707775115966797, 0.021389312744140625, 0.021339136123657225, 0.021132287979125978, 0.021615615844726564, 0.021440511703491212, 0.021191743850708007, 0.021265344619750978, 0.021350400924682617, 0.02145484733581543, 0.021112831115722656, 0.021153791427612305, 0.02129100799560547, 0.021210111618041993, 0.021176319122314453, 0.020975616455078124, 0.021134336471557616, 0.02128179168701172, 0.021424127578735352, 0.021164031982421876, 0.02147123146057129, 0.021137407302856445, 0.021136383056640624, 0.021198911666870116, 0.021197759628295898, 0.021564416885375977, 0.021275648117065428, 0.021738496780395508, 0.02145587158203125, 0.021147647857666017, 0.021404672622680664, 0.021176319122314453, 0.021209087371826172, 0.05206220626831055, 0.021217279434204102, 0.021114879608154297, 0.021175296783447265, 0.021149696350097655, 0.021136383056640624, 0.02111692810058594, 0.021132287979125978, 0.020974592208862306, 0.021183488845825195, 0.021378047943115236, 0.021217279434204102, 0.021171199798583985, 0.021161983489990235, 0.02151219177246094, 0.021953535079956055, 0.021575679779052736, 0.021283840179443358, 0.021097471237182617, 0.021204992294311522, 0.021110784530639647, 0.02124595260620117, 0.02128179168701172, 0.02112512016296387, 0.021160959243774414, 0.021156864166259767, 0.021206016540527343, 0.021168127059936523, 0.020938751220703124, 0.020883455276489257, 0.021525503158569336, 0.021367839813232422, 0.021205984115600585, 0.02125619125366211, 0.02105036735534668, 0.02100223922729492, 0.02109644889831543, 0.021209087371826172, 0.021560319900512694, 0.021194751739501954, 0.021209087371826172, 0.02127667236328125, 0.02122035217285156, 0.02151628875732422, 0.021166080474853514, 0.02123366355895996, 0.02181427192687988, 0.021175296783447265, 0.021175359725952147, 0.021204927444458007, 0.021198848724365234, 0.021195775985717775, 0.02119987106323242, 0.02116409683227539, 0.021214143753051758, 0.021202943801879884, 0.021249088287353515, 0.021211103439331056, 0.020974559783935545, 0.020967424392700194, 0.020986879348754883, 0.020982784271240236, 0.020980735778808594, 0.05229363250732422, 0.021223424911499023, 0.0212674560546875, 0.021357568740844726, 0.02119987106323242, 0.021359615325927735, 0.021277727127075194, 0.021219295501708986, 0.021334016799926758, 0.020937728881835937, 0.020953088760375976, 0.02096335983276367, 0.020942815780639647, 0.021489664077758788, 0.021152767181396484, 0.020988927841186524, 0.02105548858642578, 0.021020671844482423, 0.021046272277832033, 0.020988927841186524, 0.021013504028320314, 0.021296159744262695, 0.020981727600097658, 0.020942848205566408, 0.02104217529296875, 0.021198848724365234, 0.021210111618041993, 0.021167104721069335, 0.021308416366577147, 0.021634048461914062, 0.021225536346435547, 0.021304256439208986, 0.021321727752685548, 0.021328895568847657, 0.021288991928100586, 0.0211507511138916, 0.02122438430786133, 0.02168627166748047, 0.02142617607116699, 0.02145587158203125, 0.02143948745727539, 0.0212674560546875, 0.021184511184692383, 0.021235712051391603, 0.021432319641113282, 0.021646335601806642, 0.02123263931274414, 0.02124185562133789, 0.021122047424316406, 0.021097471237182617, 0.021448703765869142, 0.021118976593017577, 0.021316608428955077, 0.021188608169555666, 0.021110784530639647, 0.021191680908203125, 0.021106687545776368, 0.02122035217285156, 0.02162073516845703, 0.021327871322631836, 0.0216494083404541, 0.021144575119018554, 0.021132287979125978, 0.0521278076171875, 0.02114348793029785, 0.02109337615966797, 0.021195775985717775, 0.021113855361938477, 0.021104639053344726, 0.021105663299560547, 0.02108518409729004, 0.02085478401184082, 0.020959264755249025, 0.02088956832885742, 0.021190656661987304, 0.021115903854370118, 0.021073919296264648, 0.021061632156372072, 0.020925439834594727, 0.020864000320434572, 0.02088140869140625, 0.020875328063964842, 0.0211342716217041, 0.021313535690307618, 0.02089369583129883, 0.02084966468811035, 0.021094400405883788, 0.02087116813659668, 0.020932607650756836, 0.020891679763793945, 0.02094076728820801, 0.020950048446655274, 0.02113225555419922, 0.020961280822753905, 0.020908031463623047, 0.02084659194946289, 0.021204992294311522, 0.02112512016296387, 0.021218303680419923, 0.021189632415771483, 0.02124083137512207, 0.021145599365234375, 0.021166080474853514, 0.021150720596313476, 0.021180416107177736, 0.021139455795288087, 0.021213184356689452, 0.021121023178100586, 0.020936767578125, 0.020938688278198243, 0.020944896697998046, 0.020924415588378906, 0.021013504028320314, 0.021109760284423826, 0.021324800491333007, 0.02121625518798828, 0.021170175552368165, 0.021173248291015623, 0.021180448532104493, 0.021215200424194336, 0.021215232849121093, 0.021196800231933592, 0.021197824478149413, 0.021211135864257814, 0.020909055709838868, 0.02105958366394043, 0.05219225692749024, 0.021195775985717775, 0.021131263732910157, 0.021156864166259767, 0.021120000839233398, 0.02104934310913086, 0.02109644889831543, 0.021144575119018554, 0.021117952346801756, 0.021128192901611328, 0.021160959243774414, 0.021140512466430665, 0.021258207321166993, 0.02167500877380371, 0.02167091178894043, 0.021138431549072266, 0.021149696350097655, 0.02148863983154297, 0.02122956848144531, 0.021140544891357423, 0.021104576110839844, 0.021189632415771483, 0.02123161506652832, 0.02110054397583008, 0.021124095916748048, 0.021136383056640624, 0.021395456314086913, 0.021560319900512694, 0.021188608169555666, 0.021139455795288087, 0.021227519989013673, 0.0218787841796875, 0.02181427192687988, 0.021749759674072267, 0.022569984436035157, 0.021801984786987305, 0.021146623611450196, 0.021138431549072266, 0.021151744842529296, 0.021173248291015623, 0.02120806312561035, 0.021485567092895508, 0.021615615844726564, 0.02122547149658203, 0.021111808776855468, 0.021169151306152344, 0.021153791427612305, 0.021166080474853514, 0.02109443283081055, 0.02115171241760254, 0.021118976593017577, 0.021123071670532227, 0.021133312225341795, 0.021165056228637694, 0.021214208602905273, 0.020917280197143555, 0.020961248397827148, 0.021195775985717775, 0.020981760025024415, 0.021275648117065428, 0.020948991775512696, 0.020929536819458007, 0.020956159591674805, 0.05220454406738281, 0.021105663299560547, 0.021098495483398438, 0.021003263473510742, 0.020945920944213867, 0.020917247772216797, 0.02105753517150879, 0.02113747215270996, 0.021320640563964845, 0.02128998374938965, 0.021177343368530274, 0.02109644889831543, 0.021081087112426757, 0.021127168655395507, 0.02088652801513672, 0.020917247772216797, 0.021206016540527343, 0.021145599365234375, 0.02118662452697754, 0.02097862434387207, 0.021014528274536134, 0.020933631896972657, 0.02092748832702637, 0.021242879867553712, 0.021271551132202148, 0.020993024826049804, 0.021128192901611328, 0.021403648376464843, 0.02124799919128418, 0.02107187271118164, 0.020920320510864256, 0.020946943283081054, 0.02092748832702637, 0.020980735778808594, 0.020956159591674805, 0.021134336471557616, 0.021124095916748048, 0.02126233673095703, 0.021165056228637694, 0.021515264511108398, 0.0212490234375, 0.020935680389404295, 0.02106470489501953, 0.020941823959350587, 0.020913152694702147, 0.020934656143188478, 0.02102681541442871, 0.021110784530639647, 0.021444608688354492, 0.021179391860961915, 0.02127462387084961, 0.021159936904907226, 0.0212490234375, 0.021129215240478515, 0.021122047424316406, 0.021401599884033205, 0.021362688064575194, 0.021210111618041993, 0.021203968048095705, 0.021150720596313476, 0.021230592727661132, 0.021167104721069335, 0.021178367614746094, 0.052111358642578126, 0.021118976593017577, 0.020980735778808594, 0.0208855037689209, 0.02104425621032715, 0.021075935363769532, 0.021207040786743164, 0.021715967178344727, 0.021116992950439454, 0.021187519073486327, 0.021124095916748048, 0.021139488220214844, 0.021084192276000977, 0.020877248764038087, 0.020890623092651366, 0.020932607650756836, 0.02086911964416504, 0.021407743453979493, 0.02165862464904785, 0.02149171257019043, 0.02147327995300293, 0.02104012870788574, 0.02102374458312988, 0.02147942352294922, 0.021346303939819337, 0.02164735984802246, 0.021155839920043946, 0.021391359329223633, 0.02229350471496582, 0.021433343887329103, 0.021144575119018554, 0.021138431549072266, 0.021134336471557616, 0.021259263992309572, 0.021135360717773437, 0.0210831356048584, 0.021127168655395507, 0.021142528533935546, 0.021374975204467773, 0.021348352432250976, 0.021121023178100586, 0.021213247299194337, 0.021127103805541992, 0.021192703247070312, 0.021142528533935546, 0.021227519989013673, 0.02147942352294922, 0.021687295913696288, 0.021188608169555666, 0.021158912658691405, 0.021123071670532227, 0.021176319122314453, 0.021159936904907226, 0.021210111618041993, 0.021604352951049805, 0.02123161506652832, 0.021191680908203125, 0.02123980712890625, 0.02152038383483887, 0.02187980842590332, 0.02136479949951172, 0.02115782356262207, 0.021164031982421876, 0.05229363250732422, 0.02128691291809082, 0.021124160766601563, 0.02119980812072754, 0.021124095916748048, 0.021359615325927735, 0.02163199996948242, 0.02129100799560547, 0.021139455795288087, 0.021135360717773437, 0.021113855361938477, 0.0211015682220459, 0.021213184356689452, 0.02147331237792969, 0.021088224411010742, 0.021138431549072266, 0.02122550392150879, 0.02112406349182129, 0.021171199798583985, 0.021111808776855468, 0.021135360717773437, 0.0214835205078125, 0.021629951477050782, 0.021530624389648437, 0.021300224304199217, 0.02108518409729004, 0.02119987106323242, 0.021140480041503908, 0.021122047424316406, 0.02162892723083496, 0.021175296783447265, 0.02108415985107422, 0.021073919296264648, 0.021102592468261717, 0.021137407302856445, 0.02111692810058594, 0.021203968048095705, 0.021312511444091797, 0.021389312744140625, 0.021161983489990235, 0.021180416107177736, 0.021223424911499023, 0.021179391860961915, 0.021172224044799806, 0.021183488845825195, 0.021178367614746094, 0.021189632415771483, 0.02119990348815918, 0.021215200424194336, 0.022297599792480468, 0.021877824783325197, 0.021234624862670897, 0.021341184616088867, 0.021515264511108398, 0.02147430419921875, 0.021747711181640626, 0.021585920333862304, 0.021187583923339845, 0.021350400924682617, 0.021182464599609374, 0.021379072189331053, 0.02148249626159668, 0.021196800231933592, 0.05266636657714844, 0.021631040573120118, 0.021608383178710937, 0.02187468719482422, 0.021363712310791014, 0.021097471237182617, 0.02206822395324707, 0.021594112396240234, 0.02127667236328125, 0.021204992294311522, 0.021975040435791016, 0.021763071060180664, 0.021630975723266603, 0.02165247917175293, 0.021614591598510743, 0.02164531135559082, 0.021626880645751953, 0.02165043258666992, 0.021156864166259767, 0.021204992294311522, 0.021514240264892577, 0.021634048461914062, 0.021575679779052736, 0.02127462387084961, 0.02108415985107422, 0.021165056228637694, 0.021135360717773437, 0.021395456314086913, 0.021733375549316408, 0.02247065544128418, 0.02149478340148926, 0.021177343368530274, 0.021144575119018554, 0.021160959243774414, 0.021287935256958008, 0.021737472534179687, 0.022750207901000977, 0.02188800048828125, 0.021184576034545897, 0.021054399490356444, 0.021436416625976562, 0.021634048461914062, 0.02187264060974121, 0.021333055496215822, 0.021095359802246094, 0.02122035217285156, 0.021159999847412108, 0.02117420768737793, 0.021149696350097655, 0.02167705535888672, 0.021754880905151368, 0.02129817581176758, 0.0211015682220459, 0.021127168655395507, 0.021140480041503908, 0.02105548858642578, 0.02185625648498535, 0.021533695220947266, 0.021393407821655275, 0.02123776054382324, 0.021102592468261717, 0.021171199798583985, 0.021081087112426757, 0.052157440185546876, 0.021506048202514647, 0.021296127319335938, 0.02109644889831543, 0.02110361671447754, 0.021959680557250977, 0.021820415496826173, 0.021610496520996093, 0.021731327056884766, 0.022018047332763673, 0.022477823257446287, 0.021622783660888673, 0.021420032501220702, 0.02158799934387207, 0.021698528289794922, 0.021127168655395507, 0.02166169548034668, 0.021218303680419923, 0.021341184616088867, 0.02129206466674805, 0.021461984634399415, 0.021774335861206053, 0.02163302421569824, 0.021148672103881837, 0.02122035217285156, 0.021359615325927735, 0.02150297546386719, 0.02135865592956543, 0.021761983871459962, 0.02112512016296387, 0.02146816062927246, 0.021209087371826172, 0.021767168045043944, 0.021223424911499023, 0.021212160110473634, 0.021576736450195314, 0.02110358428955078, 0.021202943801879884, 0.021251071929931642, 0.021410816192626952, 0.021275648117065428, 0.021175296783447265, 0.021154815673828126, 0.021187583923339845, 0.021127168655395507, 0.021149696350097655, 0.02115385627746582, 0.021178304672241213, 0.021600255966186522, 0.021304319381713867, 0.021184511184692383, 0.021212160110473634, 0.021193727493286133, 0.020982847213745118, 0.020898752212524414, 0.020959232330322267, 0.020960256576538085, 0.020980735778808594, 0.021381120681762695, 0.02190540885925293, 0.021721088409423828, 0.02125823974609375, 0.021195775985717775]",tokens/s,46.07156936044984,,,main,False,False,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen2-beta-72B,Qwen/Qwen2-beta-72B,cuda,0,42,,,True,,,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",qwen2,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-0.5B,,cuda,0,42,,,,,True,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,217063f5c507ed7cc255df7e1f64c4333a0b4dfe,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,,MB,1280.438272,2227.699712,0.0,1581.252608,1426.269184,s,10,0.22319069099426267,0.02231906909942627,0.0006853761571708607,0.022013392448425293,0.023180089569091795,0.023655692863464355,0.024036175498962403,"[0.024131296157836914, 0.02194451141357422, 0.02196284866333008, 0.02203647994995117, 0.021943328857421875, 0.023074399948120116, 0.022088191986083985, 0.021981056213378907, 0.022038272857666016, 0.021990304946899415]",tokens/s,11470.012430159137,kWh,2.634063857199384e-07,1.443334596052776e-07,6.662186720489865e-07,1.0739585173742026e-06,tokens/kWh,238370473.21521556,MB,1280.438272,2229.796864,0.0,1583.34976,1426.271744,s,10,13.361648193359375,1.3361648193359374,0.008261447064200395,1.3341007690429687,1.3417626953125,1.3503688354492187,1.3572537475585937,"[1.3589749755859375, 1.3335765380859375, 1.3398502197265625, 1.334625, 1.334904541015625, 1.337093505859375, 1.331390380859375, 1.3334971923828125, 1.3282730712890625, 1.3294627685546876]",tokens/s,47.14987184837756,kWh,1.588435125045628e-05,8.704408624641345e-06,2.7742581832751557e-05,5.23313417078492e-05,tokens/kWh,1203867.4710790112,,s,629,13.537337352752663,0.02152199897098996,0.002695537719474812,0.021122047424316406,0.021710694885253908,0.02202910804748535,0.04357207275390625,"[0.022559743881225586, 0.022222848892211915, 0.021828607559204103, 0.022128639221191407, 0.021942272186279296, 0.022040576934814454, 0.021809152603149414, 0.021838911056518556, 0.02180089569091797, 0.02190540885925293, 0.022153215408325197, 0.02190745544433594, 0.021831680297851562, 0.02187980842590332, 0.022090751647949217, 0.021777408599853516, 0.022014976501464844, 0.02206105613708496, 0.02186751937866211, 0.021571584701538086, 0.02170572853088379, 0.021514240264892577, 0.022281280517578127, 0.022376415252685546, 0.022209503173828124, 0.02208460807800293, 0.02190028762817383, 0.02165760040283203, 0.021153791427612305, 0.021094400405883788, 0.02109132766723633, 0.02122652816772461, 0.021211103439331056, 0.02106879997253418, 0.022016000747680665, 0.02165657615661621, 0.021625856399536132, 0.021544960021972655, 0.02109132766723633, 0.02106368064880371, 0.021104639053344726, 0.021174272537231444, 0.02103193664550781, 0.021572608947753907, 0.02148454475402832, 0.02109337615966797, 0.021117952346801756, 0.021147647857666017, 0.021151744842529296, 0.021167104721069335, 0.021140480041503908, 0.020942848205566408, 0.020892671585083008, 0.020931583404541015, 0.020880384445190428, 0.021172256469726564, 0.0212838077545166, 0.02102783966064453, 0.020950016021728517, 0.02107187271118164, 0.021102592468261717, 0.02106889533996582, 0.04357519912719727, 0.021029888153076173, 0.021086208343505858, 0.02107084846496582, 0.021061632156372072, 0.020987903594970703, 0.02081996726989746, 0.020817920684814452, 0.021014528274536134, 0.021014528274536134, 0.021102592468261717, 0.02104729652404785, 0.021130239486694336, 0.0210565128326416, 0.02104115104675293, 0.02102374458312988, 0.021134336471557616, 0.02084351921081543, 0.020912128448486327, 0.02229452705383301, 0.021174272537231444, 0.021123071670532227, 0.02110771179199219, 0.021111808776855468, 0.02108518409729004, 0.02105548858642578, 0.021115903854370118, 0.02153267288208008, 0.021576736450195314, 0.021346271514892576, 0.021134336471557616, 0.02107084846496582, 0.021159936904907226, 0.02116921615600586, 0.021016511917114258, 0.021104639053344726, 0.021127168655395507, 0.021179391860961915, 0.021037055969238282, 0.021122047424316406, 0.021517311096191406, 0.02105958366394043, 0.021352479934692383, 0.02109334373474121, 0.020898815155029296, 0.0208855037689209, 0.021604352951049805, 0.02150092887878418, 0.021053440093994142, 0.021114879608154297, 0.021158912658691405, 0.021109760284423826, 0.021325824737548828, 0.021412864685058593, 0.021230592727661132, 0.021134336471557616, 0.021133344650268556, 0.021173215866088866, 0.02144767951965332, 0.021160959243774414, 0.02116511917114258, 0.021230560302734375, 0.021363679885864257, 0.04356403350830078, 0.02108006477355957, 0.02227916717529297, 0.02147942352294922, 0.02121116828918457, 0.021158880233764648, 0.021124095916748048, 0.021201919555664063, 0.021148672103881837, 0.021153791427612305, 0.021186559677124024, 0.021238784790039062, 0.021124095916748048, 0.021213184356689452, 0.0209039363861084, 0.020892671585083008, 0.021001216888427734, 0.02170572853088379, 0.021378047943115236, 0.02142310333251953, 0.021405759811401366, 0.02119161605834961, 0.021226495742797852, 0.021180416107177736, 0.021174272537231444, 0.021198848724365234, 0.021202943801879884, 0.021255168914794922, 0.021155839920043946, 0.021127168655395507, 0.02111907196044922, 0.02119055938720703, 0.021159936904907226, 0.021092384338378907, 0.021170143127441407, 0.020935680389404295, 0.021115903854370118, 0.021138431549072266, 0.021085216522216798, 0.021734367370605467, 0.02228121566772461, 0.021347328186035155, 0.02122137641906738, 0.021197824478149413, 0.021284896850585936, 0.02128483200073242, 0.021189632415771483, 0.021184511184692383, 0.021128192901611328, 0.022082559585571288, 0.021767168045043944, 0.021749759674072267, 0.02148863983154297, 0.021188608169555666, 0.021191680908203125, 0.021154815673828126, 0.021122047424316406, 0.021078016281127928, 0.02112006378173828, 0.021082048416137696, 0.021128192901611328, 0.02107494354248047, 0.021176319122314453, 0.043676673889160154, 0.021167104721069335, 0.021152767181396484, 0.02110873603820801, 0.021159936904907226, 0.021137407302856445, 0.021102592468261717, 0.02107084846496582, 0.021082111358642578, 0.021538816452026367, 0.021122047424316406, 0.021157888412475585, 0.021147647857666017, 0.021188608169555666, 0.021200895309448242, 0.021165056228637694, 0.021129215240478515, 0.021120000839233398, 0.021104639053344726, 0.02107904052734375, 0.021126144409179686, 0.02160540771484375, 0.021596128463745118, 0.021311487197875977, 0.021238784790039062, 0.02110771179199219, 0.021174272537231444, 0.02110054397583008, 0.021141504287719725, 0.021173280715942384, 0.021195743560791017, 0.021143552780151367, 0.02208051109313965, 0.021738496780395508, 0.021741567611694337, 0.021716991424560548, 0.021517311096191406, 0.021174272537231444, 0.02102681541442871, 0.02104832077026367, 0.02110054397583008, 0.02104115104675293, 0.021138431549072266, 0.02109235191345215, 0.021046272277832033, 0.02107494354248047, 0.02108006477355957, 0.021141504287719725, 0.02110678482055664, 0.02106972885131836, 0.021073919296264648, 0.021090303421020508, 0.0210513916015625, 0.02101862335205078, 0.02106470489501953, 0.021129215240478515, 0.021090303421020508, 0.02088960075378418, 0.02084966468811035, 0.020898815155029296, 0.021061632156372072, 0.021102592468261717, 0.020914176940917968, 0.043529216766357424, 0.021110815048217775, 0.021018592834472657, 0.021086208343505858, 0.020965375900268556, 0.02084249687194824, 0.020921344757080077, 0.02110873603820801, 0.021105663299560547, 0.021037055969238282, 0.020925439834594727, 0.020892671585083008, 0.021015552520751952, 0.02109644889831543, 0.02105958366394043, 0.02086502456665039, 0.022183935165405275, 0.021518335342407227, 0.02244095993041992, 0.021497856140136717, 0.02102783966064453, 0.021007360458374022, 0.020953088760375976, 0.02083839988708496, 0.02107699203491211, 0.021109760284423826, 0.020916223526000977, 0.02103091239929199, 0.021152767181396484, 0.021078016281127928, 0.02105763244628906, 0.02106563186645508, 0.02108415985107422, 0.02107811164855957, 0.021054367065429687, 0.02109235191345215, 0.021321727752685548, 0.021570560455322265, 0.021163007736206055, 0.021053440093994142, 0.02104729652404785, 0.021081087112426757, 0.021071935653686525, 0.021014463424682616, 0.021124095916748048, 0.020974592208862306, 0.020892671585083008, 0.021836799621582033, 0.02294272041320801, 0.021943296432495117, 0.021601280212402343, 0.02105958366394043, 0.021121023178100586, 0.021038080215454103, 0.021217279434204102, 0.021200895309448242, 0.02116409683227539, 0.02114348793029785, 0.021153791427612305, 0.021046272277832033, 0.02106879997253418, 0.02104422378540039, 0.021135360717773437, 0.04473040008544922, 0.021055456161499023, 0.02105753517150879, 0.021007423400878907, 0.021079999923706055, 0.02103193664550781, 0.02104422378540039, 0.02107084846496582, 0.021123071670532227, 0.021135360717773437, 0.021032960891723632, 0.021090303421020508, 0.02104115104675293, 0.02106879997253418, 0.0210882568359375, 0.021016607284545897, 0.02089571189880371, 0.02082508850097656, 0.020988927841186524, 0.02104319953918457, 0.02106675148010254, 0.021055551528930665, 0.021128128051757813, 0.021126144409179686, 0.021122047424316406, 0.021163007736206055, 0.021169151306152344, 0.021222400665283202, 0.02171392059326172, 0.021219327926635743, 0.021159936904907226, 0.021243904113769533, 0.021362688064575194, 0.020934656143188478, 0.02120806312561035, 0.021178367614746094, 0.02105753517150879, 0.02110361671447754, 0.021407743453979493, 0.022149120330810547, 0.021997568130493163, 0.02170675277709961, 0.02164019203186035, 0.021610496520996093, 0.02128281593322754, 0.021121023178100586, 0.022024192810058595, 0.021898239135742188, 0.021695487976074217, 0.021709888458251954, 0.021155775070190428, 0.021135360717773437, 0.0210882568359375, 0.021134336471557616, 0.021148672103881837, 0.021169151306152344, 0.02108006477355957, 0.021147647857666017, 0.02109951972961426, 0.020916223526000977, 0.020914176940917968, 0.021113855361938477, 0.021191680908203125, 0.043676673889160154, 0.0210882568359375, 0.021136383056640624, 0.021152767181396484, 0.021209087371826172, 0.02103398323059082, 0.020950016021728517, 0.020884479522705078, 0.020907007217407226, 0.021012479782104493, 0.020821056365966796, 0.020826047897338867, 0.021190656661987304, 0.02109132766723633, 0.021167104721069335, 0.02117843246459961, 0.02111788749694824, 0.021020671844482423, 0.0209039363861084, 0.020968479156494142, 0.021054431915283202, 0.021183488845825195, 0.021086208343505858, 0.02109132766723633, 0.02104832077026367, 0.020999168395996092, 0.02082611274719238, 0.02106470489501953, 0.021045248031616212, 0.021534719467163087, 0.021780479431152345, 0.021824512481689453, 0.021428224563598632, 0.020934656143188478, 0.02107187271118164, 0.021207040786743164, 0.021165056228637694, 0.021154815673828126, 0.021163007736206055, 0.02085273551940918, 0.02112512016296387, 0.021128192901611328, 0.021147647857666017, 0.02110054397583008, 0.021145599365234375, 0.021089279174804687, 0.020888576507568358, 0.02102681541442871, 0.021037055969238282, 0.021155839920043946, 0.021120000839233398, 0.021195775985717775, 0.021061632156372072, 0.021045248031616212, 0.021137407302856445, 0.02109644889831543, 0.021135360717773437, 0.021342208862304687, 0.021772287368774415, 0.02126848030090332, 0.021121023178100586, 0.02124492835998535, 0.021255168914794922, 0.04366643142700195, 0.021107744216918946, 0.02111996841430664, 0.020930559158325195, 0.02103398323059082, 0.021124160766601563, 0.021094335556030273, 0.021147647857666017, 0.021131263732910157, 0.02110361671447754, 0.021155839920043946, 0.021187583923339845, 0.0210831356048584, 0.02226585578918457, 0.021432319641113282, 0.021159936904907226, 0.02123980712890625, 0.02123980712890625, 0.021129215240478515, 0.021172224044799806, 0.021148672103881837, 0.021146623611450196, 0.021111839294433592, 0.021207008361816406, 0.021174272537231444, 0.021160959243774414, 0.021169151306152344, 0.021135360717773437, 0.021179391860961915, 0.0211691837310791, 0.021205984115600585, 0.021185535430908203, 0.021179391860961915, 0.021157888412475585, 0.021142528533935546, 0.021149696350097655, 0.021218303680419923, 0.02124185562133789, 0.02110771179199219, 0.02086911964416504, 0.02109337615966797, 0.02107699203491211, 0.020989952087402345, 0.020930559158325195, 0.020890623092651366, 0.020916223526000977, 0.020987903594970703, 0.020907007217407226, 0.020984832763671874, 0.020993024826049804, 0.020907007217407226, 0.020964351654052735, 0.021131263732910157, 0.02123673629760742, 0.021191680908203125, 0.021154815673828126, 0.021164031982421876, 0.02135353660583496, 0.02117318344116211, 0.021138431549072266, 0.022032384872436524, 0.02146918487548828, 0.021255168914794922, 0.04367769622802734, 0.021151744842529296, 0.021104639053344726, 0.021408767700195314, 0.021440511703491212, 0.021179391860961915, 0.021037055969238282, 0.021102592468261717, 0.021013504028320314, 0.021038080215454103, 0.021053440093994142, 0.020899839401245117, 0.020784128189086915, 0.020792320251464845, 0.021090303421020508, 0.021003263473510742, 0.020948991775512696, 0.020789247512817383, 0.020766719818115235, 0.020760576248168947, 0.0208353271484375, 0.020757503509521484, 0.02081279945373535, 0.020853759765625, 0.020981760025024415, 0.02079641532897949, 0.021102592468261717, 0.021061632156372072, 0.02109337615966797, 0.02107904052734375, 0.021001216888427734, 0.021097471237182617, 0.02107699203491211, 0.02112518310546875, 0.0211711368560791, 0.021019647598266602, 0.02107494354248047, 0.02104012870788574, 0.021625856399536132, 0.021117952346801756, 0.021113855361938477, 0.021131263732910157, 0.021104639053344726, 0.02107084846496582, 0.022344703674316405, 0.02166886329650879, 0.021098495483398438, 0.021123071670532227, 0.02110873603820801, 0.021141504287719725, 0.021181440353393553, 0.02105753517150879, 0.021142528533935546, 0.021120000839233398, 0.020959232330322267, 0.0209039363861084, 0.020940799713134766, 0.02083843231201172, 0.02113430404663086, 0.021120031356811522, 0.021047264099121093, 0.021104639053344726, 0.021118976593017577, 0.043650047302246094, 0.02107596778869629, 0.021155839920043946, 0.02109951972961426, 0.021115903854370118, 0.021140480041503908, 0.021109760284423826, 0.021144575119018554, 0.02108518409729004, 0.021135360717773437, 0.021137407302856445, 0.02106879997253418, 0.021078016281127928, 0.02106879997253418, 0.021137407302856445, 0.021020671844482423, 0.02107699203491211, 0.020948991775512696, 0.021157888412475585, 0.020884479522705078, 0.021053440093994142, 0.021028863906860353, 0.02109337615966797, 0.02106982421875, 0.02110361671447754, 0.02105753517150879, 0.021110784530639647, 0.022001663208007814, 0.02185830307006836, 0.02109337615966797, 0.021020671844482423, 0.021137407302856445, 0.021181440353393553, 0.02108723258972168, 0.021128192901611328, 0.02105753517150879, 0.021103647232055663, 0.02116815948486328, 0.02114758491516113, 0.020839424133300782, 0.020840448379516603, 0.020925439834594727, 0.02105446434020996, 0.021123071670532227, 0.02128998374938965, 0.02103398323059082, 0.02103398323059082, 0.021012479782104493, 0.021150720596313476, 0.02104729652404785, 0.02110873603820801, 0.021008384704589843, 0.021084224700927735, 0.021048255920410156, 0.0211015682220459, 0.02079644775390625, 0.020956127166748047, 0.0208353271484375, 0.020936704635620116, 0.021168127059936523, 0.021029888153076173, 0.02107494354248047, 0.021189632415771483]",tokens/s,46.464085485178444,,,main,False,False,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-1.8B,,cuda,0,42,,,,,True,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,217063f5c507ed7cc255df7e1f64c4333a0b4dfe,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,,MB,1280.77824,5222.432768,0.0,4575.985664,4408.404992,s,10,0.2996208972930908,0.02996208972930909,0.0006406597139741031,0.029801631927490232,0.03034066619873047,0.031043501091003418,0.03160576900482178,"[0.03174633598327637, 0.03018448066711426, 0.02932364845275879, 0.029996639251708986, 0.029662208557128908, 0.029786144256591797, 0.02980108833312988, 0.029802175521850587, 0.029428159713745117, 0.029890016555786134]",tokens/s,8544.130343137558,kWh,3.491530141764803e-07,1.9131910481615884e-07,1.3412940917168101e-06,1.8817662107094493e-06,tokens/kWh,136042404.49374676,MB,1280.77824,5222.432768,0.0,4575.985664,4408.407552,s,10,14.066130737304686,1.4066130737304687,0.0129631800192974,1.4080363159179687,1.4199089721679687,1.4203238708496093,1.4206557897949217,"[1.4068297119140625, 1.398411865234375, 1.4057283935546876, 1.409242919921875, 1.42073876953125, 1.4197501220703126, 1.4104517822265625, 1.3749361572265626, 1.4002242431640626, 1.4198167724609374]",tokens/s,44.78843626336995,kWh,1.6028081063939408e-05,8.783234179226324e-06,3.890990477588561e-05,6.372122001905135e-05,tokens/kWh,988681.6351156534,,s,629,14.310509576797493,0.022751207594272634,0.003601284367660396,0.02248294448852539,0.022652845001220702,0.02277171211242676,0.051737850646972665,"[0.022715391159057616, 0.022351871490478514, 0.02245427131652832, 0.022495231628417968, 0.022346752166748047, 0.022502399444580077, 0.02242355155944824, 0.02225663948059082, 0.022139904022216796, 0.02265190315246582, 0.022495231628417968, 0.023756799697875978, 0.02314348793029785, 0.022495168685913086, 0.02230784034729004, 0.022492160797119142, 0.022768640518188478, 0.02249728012084961, 0.022336511611938475, 0.02209280014038086, 0.021994495391845705, 0.022839296340942384, 0.023358463287353515, 0.023001087188720702, 0.022676479339599608, 0.021539840698242187, 0.021612543106079102, 0.021598207473754884, 0.021549055099487305, 0.021599231719970705, 0.021565439224243164, 0.021593088150024413, 0.02163609504699707, 0.021554176330566405, 0.021563392639160156, 0.021565439224243164, 0.021630048751831055, 0.021596063613891603, 0.021597183227539063, 0.02162892723083496, 0.022115327835083007, 0.022598655700683593, 0.02246348762512207, 0.02247270393371582, 0.02253004837036133, 0.022410240173339844, 0.02246553611755371, 0.022561792373657227, 0.0224532470703125, 0.022406143188476564, 0.0225167350769043, 0.02246451187133789, 0.022452224731445314, 0.02249318313598633, 0.022460416793823244, 0.022445056915283205, 0.022608896255493165, 0.02263039970397949, 0.02243891143798828, 0.022518783569335937, 0.02253004837036133, 0.022520864486694336, 0.05312918472290039, 0.022384639739990234, 0.022483968734741212, 0.022458368301391602, 0.0224716796875, 0.02241433525085449, 0.022475776672363282, 0.0224901123046875, 0.022459392547607423, 0.022458368301391602, 0.02234783935546875, 0.022313919067382813, 0.021922815322875978, 0.02142617607116699, 0.021525503158569336, 0.02150399971008301, 0.021515264511108398, 0.021531648635864258, 0.02147225570678711, 0.021498880386352538, 0.02150912094116211, 0.021501951217651367, 0.02233241653442383, 0.022379520416259766, 0.022388736724853517, 0.022346752166748047, 0.02247987174987793, 0.022435840606689454, 0.021498880386352538, 0.02150092887878418, 0.02147532844543457, 0.02142720031738281, 0.021531648635864258, 0.021498880386352538, 0.021550079345703126, 0.021772287368774415, 0.021788703918457032, 0.0216790714263916, 0.0225218563079834, 0.022510591506958007, 0.022525951385498046, 0.02251468849182129, 0.022633472442626954, 0.022552576065063477, 0.02254643249511719, 0.0226375675201416, 0.022676479339599608, 0.022518783569335937, 0.02254643249511719, 0.022541311264038084, 0.022568960189819336, 0.02247987174987793, 0.0224768009185791, 0.02249625587463379, 0.022509567260742186, 0.022494207382202147, 0.02250752067565918, 0.022533119201660155, 0.022451200485229493, 0.022432832717895507, 0.022519775390625, 0.02253206443786621, 0.022553600311279298, 0.05137510299682617, 0.021578752517700195, 0.021590015411376954, 0.021659648895263672, 0.021695487976074217, 0.021901311874389647, 0.021865535736083984, 0.02178144073486328, 0.021763071060180664, 0.02208665657043457, 0.02170163154602051, 0.02166988754272461, 0.02164531135559082, 0.022152191162109376, 0.022773759841918945, 0.022584320068359375, 0.02250752067565918, 0.02251366424560547, 0.02282700729370117, 0.022526975631713866, 0.022533119201660155, 0.022525951385498046, 0.02251571273803711, 0.02247372817993164, 0.02252288055419922, 0.022568960189819336, 0.022542335510253905, 0.02254745674133301, 0.022608896255493165, 0.02253107261657715, 0.022634496688842775, 0.022675455093383787, 0.022602752685546876, 0.022689855575561524, 0.022552511215209962, 0.022611967086791994, 0.022607872009277344, 0.022584320068359375, 0.022517759323120116, 0.022552576065063477, 0.02249830436706543, 0.02254643249511719, 0.022540288925170897, 0.022569984436035157, 0.02244915199279785, 0.022549503326416014, 0.021626880645751953, 0.02206105613708496, 0.021769216537475586, 0.02165760040283203, 0.02163711929321289, 0.02168115234375, 0.021651456832885742, 0.022657024383544923, 0.02271232032775879, 0.022576128005981445, 0.02254745674133301, 0.022581247329711913, 0.022580223083496095, 0.02249728012084961, 0.022625280380249024, 0.02285158348083496, 0.022625280380249024, 0.05304217529296875, 0.022627328872680662, 0.022492160797119142, 0.022543359756469726, 0.022617088317871094, 0.022536191940307617, 0.02262015914916992, 0.022578176498413087, 0.022579200744628908, 0.022510591506958007, 0.022639616012573242, 0.02249932861328125, 0.022580223083496095, 0.022502399444580077, 0.022529024124145508, 0.022566911697387695, 0.02250752067565918, 0.02248192024230957, 0.02253926467895508, 0.023560192108154295, 0.02304921531677246, 0.02271129608154297, 0.022500352859497072, 0.02246143913269043, 0.02242355155944824, 0.022562816619873048, 0.0225218563079834, 0.022591487884521484, 0.022467584609985353, 0.021570560455322265, 0.02163711929321289, 0.021541887283325196, 0.021567487716674806, 0.021560319900512694, 0.021580799102783203, 0.021576704025268553, 0.02165350341796875, 0.021573631286621094, 0.021580799102783203, 0.021710847854614256, 0.021651456832885742, 0.021577728271484374, 0.021606399536132814, 0.021575712203979493, 0.022620128631591796, 0.022480928421020507, 0.02253206443786621, 0.022524927139282228, 0.022495231628417968, 0.02251366424560547, 0.022786048889160155, 0.02392678451538086, 0.022965248107910157, 0.022518783569335937, 0.022549503326416014, 0.02246451187133789, 0.022519807815551757, 0.02248192024230957, 0.02249830436706543, 0.022529024124145508, 0.022533119201660155, 0.022047744750976563, 0.022481983184814452, 0.052620223999023434, 0.02247987174987793, 0.02248192024230957, 0.022459392547607423, 0.022503423690795898, 0.022504512786865234, 0.022506431579589845, 0.02242460823059082, 0.022255584716796874, 0.02225971221923828, 0.02249113655090332, 0.02248908805847168, 0.022501375198364256, 0.022509567260742186, 0.022540288925170897, 0.02274406433105469, 0.023601152420043944, 0.022785024642944338, 0.02262015914916992, 0.022579200744628908, 0.022608896255493165, 0.022592512130737305, 0.02249932861328125, 0.022609920501708985, 0.02267750358581543, 0.022486015319824217, 0.022588415145874022, 0.022556671142578123, 0.02271027183532715, 0.022520832061767578, 0.022458368301391602, 0.022740991592407226, 0.022557695388793944, 0.022495231628417968, 0.022519807815551757, 0.022494207382202147, 0.022460416793823244, 0.022584320068359375, 0.02250752067565918, 0.022674432754516603, 0.022386688232421875, 0.022236160278320313, 0.022572032928466795, 0.02251571273803711, 0.02246963119506836, 0.022549503326416014, 0.022500415802001954, 0.022565824508666992, 0.022542335510253905, 0.02253926467895508, 0.02253107261657715, 0.02244607925415039, 0.022544384002685547, 0.0225218563079834, 0.02247372817993164, 0.022424575805664062, 0.022803455352783202, 0.022466560363769532, 0.02248089599609375, 0.022543359756469726, 0.02264678382873535, 0.022483968734741212, 0.02248806381225586, 0.052775936126708986, 0.02246963119506836, 0.023412736892700195, 0.023617536544799804, 0.02287001609802246, 0.022684671401977538, 0.022562816619873048, 0.022571008682250978, 0.02253926467895508, 0.022508544921875, 0.022507551193237305, 0.022638559341430664, 0.022658048629760744, 0.022542335510253905, 0.022508544921875, 0.022587392807006838, 0.022213632583618165, 0.022483968734741212, 0.02226278305053711, 0.02226688003540039, 0.02227097511291504, 0.02223411178588867, 0.022232063293457033, 0.022487039566040038, 0.022510591506958007, 0.02255462455749512, 0.022458368301391602, 0.022537216186523438, 0.022559743881225586, 0.022441984176635742, 0.02246553611755371, 0.02243891143798828, 0.022490144729614258, 0.02251158332824707, 0.02244812774658203, 0.0224768009185791, 0.022497312545776367, 0.02251158332824707, 0.0224849910736084, 0.022502399444580077, 0.022534143447875975, 0.02253004837036133, 0.022214656829833986, 0.022436864852905275, 0.02261299133300781, 0.022565887451171874, 0.022510591506958007, 0.022508544921875, 0.022526975631713866, 0.02264575958251953, 0.02251366424560547, 0.022622207641601562, 0.022556671142578123, 0.02244710350036621, 0.02250444793701172, 0.022550527572631835, 0.02243071937561035, 0.02259971237182617, 0.022450143814086915, 0.022478847503662108, 0.022536191940307617, 0.022534143447875975, 0.02247372817993164, 0.05276774215698242, 0.02241535949707031, 0.022396928787231447, 0.022406240463256837, 0.022459295272827147, 0.022486015319824217, 0.02248294448852539, 0.022467584609985353, 0.02251366424560547, 0.022411264419555665, 0.022486080169677736, 0.02238355255126953, 0.022477823257446287, 0.022492160797119142, 0.022486015319824217, 0.022487039566040038, 0.022437887191772463, 0.02259660720825195, 0.022445056915283205, 0.022445056915283205, 0.022509567260742186, 0.02253107261657715, 0.022537216186523438, 0.02247987174987793, 0.02245734405517578, 0.022509567260742186, 0.021600255966186522, 0.021563392639160156, 0.02322640037536621, 0.022596576690673827, 0.02248806381225586, 0.022571008682250978, 0.0225218563079834, 0.022540288925170897, 0.022395904541015626, 0.02248192024230957, 0.021597183227539063, 0.021573631286621094, 0.021533695220947266, 0.021567487716674806, 0.022444032669067384, 0.02249625587463379, 0.022990848541259764, 0.0224768009185791, 0.021507072448730468, 0.021772287368774415, 0.02249113655090332, 0.022445056915283205, 0.02253004837036133, 0.022420480728149415, 0.02247372817993164, 0.022343679428100584, 0.02240716743469238, 0.022391807556152343, 0.02240716743469238, 0.022418432235717774, 0.022673408508300782, 0.02244710350036621, 0.022495231628417968, 0.02244812774658203, 0.02244095993041992, 0.02246963119506836, 0.022459392547607423, 0.05181958389282226, 0.021584831237792968, 0.021586944580078125, 0.021565439224243164, 0.021608448028564452, 0.021562368392944335, 0.021711872100830077, 0.021618688583374023, 0.021625856399536132, 0.021588991165161133, 0.02165247917175293, 0.021811199188232423, 0.022715391159057616, 0.022563840866088865, 0.02248806381225586, 0.02245020866394043, 0.022469600677490233, 0.022502399444580077, 0.02254745674133301, 0.02249830436706543, 0.02250649642944336, 0.022583328247070312, 0.022413280487060545, 0.022537216186523438, 0.022483968734741212, 0.02167807960510254, 0.021610496520996093, 0.021600255966186522, 0.02168934440612793, 0.02161664009094238, 0.021560319900512694, 0.021545984268188476, 0.021599231719970705, 0.021543935775756837, 0.021642240524291992, 0.02152448081970215, 0.021622783660888673, 0.021827583312988282, 0.021625856399536132, 0.021646335601806642, 0.02166169548034668, 0.021590015411376954, 0.021613567352294923, 0.021577728271484374, 0.021573631286621094, 0.021572608947753907, 0.021604352951049805, 0.021775360107421874, 0.021618688583374023, 0.02163199996948242, 0.021568511962890623, 0.021604352951049805, 0.021630975723266603, 0.02142720031738281, 0.021769216537475586, 0.0216944637298584, 0.021609472274780273, 0.02166783905029297, 0.021393407821655275, 0.02142207908630371, 0.02172313690185547, 0.021596160888671875, 0.02246553611755371, 0.05152767944335938, 0.021556224822998047, 0.021604352951049805, 0.021659648895263672, 0.021533695220947266, 0.021557247161865235, 0.021612607955932617, 0.021518272399902345, 0.021511167526245118, 0.021368831634521485, 0.02129817581176758, 0.021336063385009766, 0.0212674560546875, 0.02126233673095703, 0.021362688064575194, 0.02148044776916504, 0.021552127838134767, 0.022743040084838868, 0.022487039566040038, 0.022403135299682617, 0.02242246437072754, 0.022452224731445314, 0.022380544662475587, 0.022477823257446287, 0.022443008422851563, 0.022433792114257813, 0.022435840606689454, 0.0224768009185791, 0.02250444793701172, 0.02243177604675293, 0.022405088424682616, 0.022466560363769532, 0.02245631980895996, 0.02244915199279785, 0.022468608856201173, 0.02246246337890625, 0.022459392547607423, 0.0226693115234375, 0.022466560363769532, 0.022569984436035157, 0.02230067253112793, 0.022441984176635742, 0.02247270393371582, 0.02245734405517578, 0.02241433525085449, 0.02251468849182129, 0.022466560363769532, 0.022387712478637696, 0.022434816360473633, 0.022178815841674804, 0.022377471923828125, 0.022726655960083008, 0.02271753692626953, 0.022612895965576172, 0.022610944747924806, 0.022582271575927734, 0.022536191940307617, 0.022627328872680662, 0.022610944747924806, 0.022758399963378906, 0.022577152252197266, 0.022500352859497072, 0.022563840866088865, 0.052776958465576174, 0.02245734405517578, 0.022587392807006838, 0.02244710350036621, 0.02254745674133301, 0.02248089599609375, 0.022640640258789063, 0.022603776931762694, 0.02230169677734375, 0.022387712478637696, 0.022592512130737305, 0.02228428840637207, 0.022072320938110353, 0.022564863204956053, 0.022470720291137697, 0.022454208374023437, 0.022749183654785156, 0.022617088317871094, 0.02260479927062988, 0.02259660720825195, 0.02270412826538086, 0.022557695388793944, 0.022668319702148436, 0.022660064697265624, 0.02268262481689453, 0.022648832321166993, 0.02249932861328125, 0.022681600570678712, 0.022684671401977538, 0.02259660720825195, 0.022603776931762694, 0.02263043212890625, 0.022645727157592773, 0.02255462455749512, 0.022381568908691408, 0.022427648544311524, 0.02244915199279785, 0.022444032669067384, 0.022208511352539064, 0.022519807815551757, 0.022527999877929687, 0.022549503326416014, 0.022468608856201173, 0.02226278305053711, 0.022576128005981445, 0.022622207641601562, 0.022564863204956053, 0.022208511352539064, 0.022409215927124023, 0.022624256134033204, 0.022239231109619142, 0.02232524871826172, 0.02230988883972168, 0.022397951126098634, 0.022588415145874022, 0.02254745674133301, 0.02303385543823242, 0.02294169616699219, 0.022591583251953123, 0.02265283203125, 0.022559743881225586, 0.02259971237182617, 0.022652896881103515]",tokens/s,43.953710846176755,,,main,False,False,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-70m,,cuda,0,42,,,,,True,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,217063f5c507ed7cc255df7e1f64c4333a0b4dfe,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,,MB,1250.996224,885.522432,0.0,239.075328,215.486464,s,33,0.1769663357734681,0.0053626162355596375,0.0002570235649559515,0.00526259183883667,0.005536326313018799,0.005578227043151855,0.006252962608337402,"[0.0065446081161499026, 0.005397280216217041, 0.005633215904235839, 0.005541567802429199, 0.005203455924987793, 0.005216959953308105, 0.0052919678688049315, 0.005539455890655518, 0.005497536182403564, 0.0055116481781005856, 0.005508543968200684, 0.005171135902404785, 0.005177504062652588, 0.005198751926422119, 0.005169536113739014, 0.005292768001556396, 0.005238239765167236, 0.005190879821777344, 0.005200255870819092, 0.00551526403427124, 0.005498079776763916, 0.005522528171539307, 0.005209440231323242, 0.005263679981231689, 0.005157504081726074, 0.00525708818435669, 0.00526259183883667, 0.005179391860961914, 0.005231840133666992, 0.005466207981109619, 0.005160255908966065, 0.0051933441162109375, 0.0055238080024719236]",tokens/s,47737.89299007783,kWh,6.423990868072943e-08,3.51987831490572e-08,1.1726808366394197e-07,2.167067754937286e-07,tokens/kWh,1181319778.3814032,MB,1250.996224,885.522432,0.0,239.075328,220.896256,s,33,10.074363525390625,0.3052837431936553,0.008342521184346705,0.30772052001953126,0.31419124145507815,0.31469938964843747,0.31493685302734375,"[0.3145581665039062, 0.30772052001953126, 0.3149112243652344, 0.312678955078125, 0.29504302978515623, 0.31121762084960936, 0.3128113098144531, 0.313947998046875, 0.3149489135742187, 0.31358535766601564, 0.3130281982421875, 0.29540716552734375, 0.2955582275390625, 0.295082275390625, 0.29649365234375, 0.3095794982910156, 0.295734130859375, 0.2946538696289063, 0.3122529907226563, 0.3142242431640625, 0.31405923461914065, 0.3140561218261719, 0.30607077026367185, 0.29541567993164064, 0.2955962829589844, 0.29640576171875, 0.29568914794921874, 0.304894287109375, 0.2955860290527344, 0.31265093994140625, 0.29499560546875, 0.30248748779296875, 0.31301882934570313]",tokens/s,206.36539417703696,kWh,3.599791314735198e-06,1.9725323771527164e-06,5.574361048760254e-06,1.1146684740648168e-05,tokens/kWh,5651904.711206232,,s,2078,10.24010649585725,0.004927866456139191,0.0007231453694833516,0.004890640020370483,0.005005289602279663,0.005042329406738281,0.01032829954147339,"[0.005499904155731201, 0.005431295871734619, 0.00538316822052002, 0.005511231899261474, 0.005427135944366455, 0.005411839962005615, 0.00535756778717041, 0.005387263774871826, 0.005327871799468994, 0.005294079780578613, 0.005265408039093018, 0.005604351997375488, 0.00603545618057251, 0.005058559894561767, 0.004873216152191162, 0.004924416065216064, 0.004873216152191162, 0.004868095874786377, 0.004929535865783692, 0.0048895998001098635, 0.004844543933868409, 0.004865024089813232, 0.004894720077514648, 0.004839424133300781, 0.004888576030731201, 0.004899839878082276, 0.0048568320274353025, 0.004831232070922851, 0.004867167949676514, 0.004913055896759033, 0.004837376117706299, 0.00481382417678833, 0.004835328102111816, 0.004921343803405762, 0.004843520164489746, 0.004834303855895996, 0.00486297607421875, 0.004884479999542236, 0.004859903812408447, 0.0048220157623291016, 0.004880383968353271, 0.004827136039733886, 0.004858880043029785, 0.004835328102111816, 0.004918272018432617, 0.004798463821411133, 0.004741119861602783, 0.0048220157623291016, 0.004884479999542236, 0.004828159809112549, 0.0047226881980895995, 0.0048455681800842285, 0.004901887893676758, 0.004798463821411133, 0.00481279993057251, 0.0048527359962463375, 0.004838399887084961, 0.0048189439773559575, 0.004831232070922851, 0.0048527359962463375, 0.004820991992950439, 0.004820991992950439, 0.0107325439453125, 0.004819968223571777, 0.004838399887084961, 0.004832255840301514, 0.0048527359962463375, 0.0048189439773559575, 0.0048056960105895994, 0.004883391857147217, 0.004858880043029785, 0.00481382417678833, 0.00481279993057251, 0.004873216152191162, 0.00485478401184082, 0.0048189439773559575, 0.004799488067626953, 0.00487014389038086, 0.0048148479461669925, 0.004811776161193848, 0.004828159809112549, 0.0048261117935180665, 0.00487120008468628, 0.004923359870910644, 0.004980735778808594, 0.004897856235504151, 0.004856768131256103, 0.004845664024353027, 0.004908959865570068, 0.0048455681800842285, 0.004867072105407715, 0.004957183837890625, 0.0050728960037231445, 0.0049530878067016604, 0.004930560111999512, 0.00481279993057251, 0.004874239921569825, 0.005253119945526123, 0.004974592208862305, 0.004901887893676758, 0.004824063777923584, 0.00485478401184082, 0.004835328102111816, 0.004834303855895996, 0.004819968223571777, 0.0048527359962463375, 0.0048230400085449215, 0.0048496642112731934, 0.0048220157623291016, 0.004788224220275879, 0.004817920207977295, 0.004724800109863281, 0.0047134079933166505, 0.004698112010955811, 0.004704256057739258, 0.004717567920684815, 0.004741119861602783, 0.005054463863372802, 0.005022719860076904, 0.0050360321998596195, 0.005044223785400391, 0.0051036162376403805, 0.005037055969238281, 0.005049344062805176, 0.005074944019317627, 0.01113804817199707, 0.00499507188796997, 0.0050032639503479, 0.004965375900268554, 0.005058559894561767, 0.00501145601272583, 0.005102591991424561, 0.004989952087402344, 0.005040128231048584, 0.005040192127227784, 0.0049642882347106934, 0.004977663993835449, 0.005035007953643799, 0.0049725441932678225, 0.004965375900268554, 0.00501964807510376, 0.0049797120094299315, 0.004978687763214112, 0.004946944236755371, 0.0049725441932678225, 0.00506982421875, 0.004982880115509033, 0.004983712196350098, 0.004969471931457519, 0.004978687763214112, 0.004992000102996826, 0.00501043176651001, 0.004976640224456787, 0.004992000102996826, 0.004963327884674072, 0.004977727890014648, 0.004983744144439697, 0.0049613118171691895, 0.005017568111419677, 0.004976640224456787, 0.004963327884674072, 0.004977663993835449, 0.004998144149780274, 0.004988927841186524, 0.004988927841186524, 0.004952064037322998, 0.0053432960510253905, 0.004991936206817627, 0.004948031902313232, 0.00498476791381836, 0.00495411205291748, 0.004992000102996826, 0.004959231853485108, 0.004944896221160889, 0.004976640224456787, 0.0049572157859802245, 0.004960224151611328, 0.004968448162078858, 0.004994048118591309, 0.00496230411529541, 0.004978687763214112, 0.004952064037322998, 0.004977663993835449, 0.004971519947052002, 0.004963327884674072, 0.0049500160217285155, 0.005015552043914795, 0.004989952087402344, 0.01095372772216797, 0.004964352130889893, 0.004955135822296143, 0.00496127986907959, 0.004971519947052002, 0.004981760025024414, 0.004963327884674072, 0.004943871974945068, 0.004965375900268554, 0.004980735778808594, 0.004976704120635987, 0.004959167957305909, 0.004990015983581543, 0.004945856094360351, 0.0049827837944030765, 0.00496127986907959, 0.004895743846893311, 0.004903935909271241, 0.00491315221786499, 0.004877312183380127, 0.004922368049621582, 0.004915200233459473, 0.0049192957878112795, 0.004889632225036621, 0.004949984073638916, 0.004987904071807861, 0.0049459199905395506, 0.004987904071807861, 0.004999167919158935, 0.004998144149780274, 0.004980735778808594, 0.0049827837944030765, 0.004975615978240967, 0.00499507188796997, 0.004978687763214112, 0.0049725441932678225, 0.00496230411529541, 0.004986879825592041, 0.004988927841186524, 0.004956160068511963, 0.004947968006134033, 0.00496127986907959, 0.004981760025024414, 0.004971519947052002, 0.004959231853485108, 0.004956160068511963, 0.0049909758567810054, 0.004968448162078858, 0.004952064037322998, 0.005001215934753418, 0.004998144149780274, 0.004981791973114014, 0.004954080104827881, 0.00497049617767334, 0.0050145602226257325, 0.004914144039154053, 0.0049192957878112795, 0.004902912139892578, 0.0049500160217285155, 0.004883456230163574, 0.004921343803405762, 0.004907008171081543, 0.004892672061920166, 0.010329088211059571, 0.0046694397926330565, 0.004678656101226806, 0.004690944194793701, 0.00467251205444336, 0.004685823917388916, 0.004670464038848877, 0.004663296222686767, 0.004666368007659912, 0.00469708776473999, 0.00466534423828125, 0.004675583839416504, 0.00467251205444336, 0.004696063995361328, 0.004682752132415771, 0.00469708776473999, 0.004679679870605469, 0.004683775901794434, 0.004663296222686767, 0.00467251205444336, 0.004712448120117187, 0.004685823917388916, 0.00465715217590332, 0.004668416023254395, 0.004675583839416504, 0.004677631855010986, 0.004734975814819336, 0.004694015979766846, 0.004708384037017822, 0.0046878399848937985, 0.004675615787506104, 0.004689888000488281, 0.004668447971343994, 0.0047144641876220705, 0.004701183795928955, 0.0046694397926330565, 0.004753407955169678, 0.0046694397926330565, 0.004701183795928955, 0.004677631855010986, 0.0046561279296875, 0.004666368007659912, 0.0046254081726074215, 0.004626431941986084, 0.004654079914093018, 0.004618239879608154, 0.004628511905670166, 0.004619232177734375, 0.004628511905670166, 0.004636640071868896, 0.004675583839416504, 0.004675583839416504, 0.004689919948577881, 0.004667391777038574, 0.00466534423828125, 0.004677631855010986, 0.0046888961791992185, 0.004709375858306885, 0.0046929922103881834, 0.004690944194793701, 0.004680704116821289, 0.004674560070037842, 0.00470630407333374, 0.01032806396484375, 0.004671487808227539, 0.0046581759452819825, 0.004675583839416504, 0.004671487808227539, 0.004712448120117187, 0.0046888961791992185, 0.004668416023254395, 0.00466534423828125, 0.0046929922103881834, 0.004682752132415771, 0.0046991357803344725, 0.004853759765625, 0.004770815849304199, 0.004801536083221435, 0.004729856014251709, 0.005141503810882568, 0.005125120162963868, 0.0050657281875610355, 0.004988927841186524, 0.004998208045959473, 0.004996064186096191, 0.005023712158203125, 0.004985856056213379, 0.004974592208862305, 0.005014527797698975, 0.005037055969238281, 0.005047296047210693, 0.004992000102996826, 0.004994048118591309, 0.005000192165374756, 0.0050032639503479, 0.004958208084106445, 0.004984831809997559, 0.005001215934753418, 0.004998144149780274, 0.004946944236755371, 0.004952064037322998, 0.005041152000427246, 0.005030911922454834, 0.004992000102996826, 0.0049511041641235355, 0.004987840175628662, 0.005055488109588623, 0.004964352130889893, 0.004957183837890625, 0.004955135822296143, 0.004985856056213379, 0.004942880153656006, 0.004987872123718261, 0.004988927841186524, 0.00497049617767334, 0.004969471931457519, 0.005315584182739258, 0.005313536167144775, 0.00501043176651001, 0.004973567962646484, 0.004968448162078858, 0.005018623828887939, 0.004956160068511963, 0.0049530878067016604, 0.004958208084106445, 0.0050022401809692385, 0.01041817569732666, 0.004678656101226806, 0.004678656101226806, 0.0046622719764709475, 0.004676608085632325, 0.004665408134460449, 0.00502675199508667, 0.004980735778808594, 0.004988927841186524, 0.004963327884674072, 0.004971519947052002, 0.0050432000160217285, 0.005030911922454834, 0.004988927841186524, 0.005026815891265869, 0.0050421757698059086, 0.005039167881011963, 0.005014463901519776, 0.005008416175842285, 0.005006303787231445, 0.004968448162078858, 0.004975615978240967, 0.0049530878067016604, 0.0050165758132934574, 0.004974624156951904, 0.004963295936584473, 0.005091328144073487, 0.005024767875671386, 0.004981760025024414, 0.0049725441932678225, 0.0049807682037353515, 0.004980703830718994, 0.004975615978240967, 0.004967423915863037, 0.0049797120094299315, 0.004977663993835449, 0.004977663993835449, 0.004996096134185791, 0.0049530878067016604, 0.00501964807510376, 0.004983808040618896, 0.0049725441932678225, 0.0049500160217285155, 0.005023744106292725, 0.004988927841186524, 0.00496230411529541, 0.0049797120094299315, 0.00502784013748169, 0.004963327884674072, 0.005005311965942383, 0.004965375900268554, 0.004958208084106445, 0.005007359981536865, 0.004986879825592041, 0.004930560111999512, 0.004980735778808594, 0.004975615978240967, 0.00496230411529541, 0.004965375900268554, 0.004967423915863037, 0.005044223785400391, 0.0049725441932678225, 0.004932608127593994, 0.010950655937194824, 0.004984831809997559, 0.004969471931457519, 0.005023744106292725, 0.004992000102996826, 0.0050032639503479, 0.004968448162078858, 0.004993023872375488, 0.004994048118591309, 0.004964352130889893, 0.004968448162078858, 0.005007359981536865, 0.00496230411529541, 0.004939839839935302, 0.004944831848144531, 0.004971519947052002, 0.0050432319641113285, 0.004955103874206543, 0.0049489917755126955, 0.0049827837944030765, 0.004974656105041504, 0.005074880123138428, 0.004981760025024414, 0.0049541440010070804, 0.004952032089233399, 0.004978752136230469, 0.004955103874206543, 0.00502678394317627, 0.0050063362121582035, 0.004977663993835449, 0.004968448162078858, 0.005014527797698975, 0.00495411205291748, 0.0049725441932678225, 0.0049489917755126955, 0.0049489917755126955, 0.005033984184265137, 0.00496230411529541, 0.004973567962646484, 0.005026815891265869, 0.004977663993835449, 0.004976640224456787, 0.004994080066680908, 0.004943840026855469, 0.004973567962646484, 0.004937727928161621, 0.004959231853485108, 0.004947968006134033, 0.004973567962646484, 0.00495411205291748, 0.004951039791107178, 0.004965375900268554, 0.005014527797698975, 0.004964352130889893, 0.004963327884674072, 0.004951039791107178, 0.004991040229797363, 0.005002175807952881, 0.0049530878067016604, 0.004936704158782959, 0.004977663993835449, 0.004977663993835449, 0.004968480110168457, 0.010910688400268554, 0.004977663993835449, 0.004957248210906982, 0.0050001277923583986, 0.004966400146484375, 0.004984831809997559, 0.004963327884674072, 0.004947968006134033, 0.005006368160247803, 0.00495305585861206, 0.004962399959564209, 0.004974495887756347, 0.005032959938049316, 0.005007359981536865, 0.004974592208862305, 0.005009439945220948, 0.005253087997436523, 0.005024767875671386, 0.005025792121887207, 0.005073919773101807, 0.005023744106292725, 0.00501145601272583, 0.0050063362121582035, 0.00501964807510376, 0.0050032639503479, 0.004946944236755371, 0.004964352130889893, 0.004985856056213379, 0.004966432094573974, 0.005014495849609375, 0.004976640224456787, 0.005004288196563721, 0.004988927841186524, 0.005054463863372802, 0.00496230411529541, 0.004983808040618896, 0.00501964807510376, 0.004975615978240967, 0.004967423915863037, 0.0050124797821044925, 0.004989952087402344, 0.004968448162078858, 0.004947968006134033, 0.004963327884674072, 0.0050206718444824215, 0.0049797120094299315, 0.00496127986907959, 0.004973567962646484, 0.005055488109588623, 0.004993023872375488, 0.004943871974945068, 0.004975615978240967, 0.005015552043914795, 0.004975615978240967, 0.005039103984832764, 0.004957183837890625, 0.004994048118591309, 0.004956160068511963, 0.004981760025024414, 0.0049530878067016604, 0.004999167919158935, 0.0050032639503479, 0.004947968006134033, 0.010899456024169921, 0.004956160068511963, 0.005339136123657226, 0.004977663993835449, 0.004944896221160889, 0.0049459199905395506, 0.004975615978240967, 0.0049500160217285155, 0.004955135822296143, 0.0049469761848449706, 0.004966368198394775, 0.00496127986907959, 0.004967423915863037, 0.004941823959350586, 0.0049387521743774416, 0.004975615978240967, 0.004955135822296143, 0.004935679912567138, 0.004940800189971924, 0.004968448162078858, 0.004942912101745606, 0.004944831848144531, 0.004957183837890625, 0.004935679912567138, 0.004963327884674072, 0.004973567962646484, 0.004981760025024414, 0.004947968006134033, 0.004965375900268554, 0.004944896221160889, 0.0049725441932678225, 0.004968448162078858, 0.0049459199905395506, 0.004939775943756103, 0.0049387521743774416, 0.004959231853485108, 0.004952064037322998, 0.004964352130889893, 0.004922368049621582, 0.004946944236755371, 0.004955135822296143, 0.004947968006134033, 0.004941823959350586, 0.004968448162078858, 0.004951039791107178, 0.004951039791107178, 0.004958208084106445, 0.005014527797698975, 0.005008384227752686, 0.0053053760528564455, 0.00501961612701416, 0.004959231853485108, 0.004993023872375488, 0.00495411205291748, 0.0049530878067016604, 0.0049530878067016604, 0.004937727928161621, 0.004996096134185791, 0.004956192016601563, 0.0050052800178527835, 0.004957183837890625, 0.004971519947052002, 0.004952064037322998, 0.010943488121032715, 0.004969471931457519, 0.004958208084106445, 0.004952064037322998, 0.004971519947052002, 0.004953120231628418, 0.004945888042449951, 0.004965375900268554, 0.004978687763214112, 0.0049909758567810054, 0.004960256099700928, 0.004955135822296143, 0.004973567962646484, 0.004984831809997559, 0.004951039791107178, 0.004955135822296143, 0.00496127986907959, 0.0050022401809692385, 0.004971519947052002, 0.0049500160217285155, 0.004969471931457519, 0.004958208084106445, 0.004959231853485108, 0.004963327884674072, 0.004947968006134033, 0.004978687763214112, 0.0049530878067016604, 0.004960256099700928, 0.004988927841186524, 0.0049725441932678225, 0.005004288196563721, 0.004957183837890625, 0.004946944236755371, 0.004976640224456787, 0.004969471931457519, 0.004966400146484375, 0.0049530878067016604, 0.004976640224456787, 0.004966400146484375, 0.004958208084106445, 0.004967423915863037, 0.004927487850189209, 0.004940800189971924, 0.004959231853485108, 0.004929535865783692, 0.004959231853485108, 0.004932608127593994, 0.004964416027069092, 0.004933568000793457, 0.004967423915863037, 0.00495411205291748, 0.004966400146484375, 0.004933631896972656, 0.004965407848358154, 0.004966368198394775, 0.0049428482055664065, 0.004939775943756103, 0.004945951938629151, 0.0050124478340148925, 0.004957183837890625, 0.004965375900268554, 0.004958208084106445, 0.004975615978240967, 0.010309632301330567, 0.004661248207092285, 0.004689919948577881, 0.004666368007659912, 0.004683775901794434, 0.004651072025299072, 0.004675519943237305, 0.004666368007659912, 0.004701183795928955, 0.0047288317680358885, 0.004684800148010254, 0.004674560070037842, 0.004679679870605469, 0.004649983882904053, 0.004682752132415771, 0.004661248207092285, 0.00466431999206543, 0.004660223960876465, 0.004668416023254395, 0.004673535823822021, 0.004698112010955811, 0.004671487808227539, 0.004660223960876465, 0.00464793586730957, 0.004661248207092285, 0.004670464038848877, 0.004675583839416504, 0.004677696228027344, 0.004669375896453857, 0.004677631855010986, 0.004670464038848877, 0.004671487808227539, 0.004698112010955811, 0.004674560070037842, 0.0046929922103881834, 0.0046694397926330565, 0.004743167877197266, 0.004678656101226806, 0.004701183795928955, 0.004677696228027344, 0.004694975852966308, 0.004681727886199951, 0.004681727886199951, 0.004690944194793701, 0.004739071846008301, 0.004680704116821289, 0.004676608085632325, 0.0046694397926330565, 0.004678656101226806, 0.004708352088928222, 0.004675583839416504, 0.0046991357803344725, 0.004668416023254395, 0.004682752132415771, 0.004674560070037842, 0.004695040225982666, 0.004667391777038574, 0.0046929922103881834, 0.004695040225982666, 0.004689919948577881, 0.004726784229278564, 0.004747263908386231, 0.004675583839416504, 0.010272768020629883, 0.004667391777038574, 0.004685823917388916, 0.004677631855010986, 0.004673535823822021, 0.004708352088928222, 0.004666368007659912, 0.004690944194793701, 0.004684800148010254, 0.0046694397926330565, 0.004674560070037842, 0.004666368007659912, 0.004674560070037842, 0.005073919773101807, 0.004741119861602783, 0.004696063995361328, 0.004681727886199951, 0.004684800148010254, 0.00469708776473999, 0.004673535823822021, 0.004673535823822021, 0.004674560070037842, 0.004663296222686767, 0.004666368007659912, 0.004695040225982666, 0.004681727886199951, 0.004686848163604736, 0.00467251205444336, 0.0046622719764709475, 0.004681727886199951, 0.004689919948577881, 0.004676608085632325, 0.004676608085632325, 0.004682752132415771, 0.004679679870605469, 0.004696063995361328, 0.0046991357803344725, 0.004671487808227539, 0.004667391777038574, 0.004671487808227539, 0.004666368007659912, 0.004683775901794434, 0.00471449613571167, 0.004660223960876465, 0.004687871932983399, 0.0046592001914978025, 0.0046622719764709475, 0.004686848163604736, 0.0046694397926330565, 0.004675615787506104, 0.0047103681564331056, 0.004660287857055664, 0.004666304111480713, 0.004675583839416504, 0.004678656101226806, 0.00465715217590332, 0.004686848163604736, 0.004663296222686767, 0.004663296222686767, 0.004667391777038574, 0.004703231811523437, 0.0046622719764709475, 0.004679679870605469, 0.010308608055114746, 0.004680704116821289, 0.004694015979766846, 0.00467251205444336, 0.004668416023254395, 0.004666368007659912, 0.004689919948577881, 0.004663296222686767, 0.0046888961791992185, 0.004654111862182617, 0.00467145586013794, 0.00466431999206543, 0.0046991357803344725, 0.004660223960876465, 0.00465715217590332, 0.0046592001914978025, 0.004660223960876465, 0.004723711967468262, 0.004698112010955811, 0.004667391777038574, 0.004667391777038574, 0.004695040225982666, 0.004678656101226806, 0.004667391777038574, 0.004678656101226806, 0.00468179178237915, 0.0046724481582641605, 0.004663296222686767, 0.004666368007659912, 0.004708352088928222, 0.0046888961791992185, 0.004686848163604736, 0.00466534423828125, 0.004670464038848877, 0.004702208042144776, 0.004663296222686767, 0.004685823917388916, 0.004660223960876465, 0.004684800148010254, 0.0046694397926330565, 0.00466431999206543, 0.004683775901794434, 0.004723711967468262, 0.004679679870605469, 0.0046694397926330565, 0.004674560070037842, 0.00466534423828125, 0.004664383888244629, 0.004686848163604736, 0.004668352127075196, 0.004679679870605469, 0.0046919679641723635, 0.004680704116821289, 0.0046888961791992185, 0.00467251205444336, 0.004683775901794434, 0.00466431999206543, 0.004671487808227539, 0.004675583839416504, 0.004724736213684082, 0.0046592001914978025, 0.0046694397926330565, 0.00466534423828125, 0.010238975524902343, 0.004681727886199951, 0.004683775901794434, 0.004646912097930909, 0.0046694397926330565, 0.004671487808227539, 0.004682752132415771, 0.00466431999206543, 0.004667391777038574, 0.00467251205444336, 0.004666431903839112, 0.0046908798217773435, 0.004677631855010986, 0.004666368007659912, 0.004679679870605469, 0.0046581759452819825, 0.004660223960876465, 0.004676608085632325, 0.004689919948577881, 0.004679679870605469, 0.004685855865478515, 0.004684768199920654, 0.00467251205444336, 0.004689919948577881, 0.004683775901794434, 0.00469708776473999, 0.004671487808227539, 0.00468179178237915, 0.004666304111480713, 0.004685823917388916, 0.004700160026550293, 0.004696063995361328, 0.004673535823822021, 0.00467251205444336, 0.004687871932983399, 0.004684800148010254, 0.004671487808227539, 0.0046888961791992185, 0.004687871932983399, 0.004675583839416504, 0.004683775901794434, 0.00470527982711792, 0.004673535823822021, 0.004683775901794434, 0.004690976142883301, 0.0046806721687316895, 0.004701183795928955, 0.004927487850189209, 0.004762688159942627, 0.004704192161560059, 0.0047226881980895995, 0.0046991357803344725, 0.004758528232574463, 0.0046929922103881834, 0.0046694397926330565, 0.004680704116821289, 0.00467251205444336, 0.0047523841857910155, 0.004759552001953125, 0.0047185921669006346, 0.004741119861602783, 0.0048865280151367185, 0.004988927841186524, 0.010466303825378418, 0.004708352088928222, 0.004667391777038574, 0.004715519905090332, 0.004671487808227539, 0.004667391777038574, 0.004675583839416504, 0.004675583839416504, 0.004668416023254395, 0.004681727886199951, 0.004679679870605469, 0.004694015979766846, 0.004682752132415771, 0.004671487808227539, 0.004673535823822021, 0.00471449613571167, 0.004663296222686767, 0.005022719860076904, 0.00501964807510376, 0.005008384227752686, 0.004973567962646484, 0.004971519947052002, 0.004971519947052002, 0.0049725441932678225, 0.005005311965942383, 0.004981760025024414, 0.004964352130889893, 0.004997119903564453, 0.005005311965942383, 0.00499507188796997, 0.004981760025024414, 0.004976640224456787, 0.005396480083465576, 0.004939839839935302, 0.004965312004089355, 0.005051392078399658, 0.004981760025024414, 0.004957183837890625, 0.004968448162078858, 0.0050063362121582035, 0.004973567962646484, 0.004967423915863037, 0.00497049617767334, 0.004996096134185791, 0.00496230411529541, 0.004964352130889893, 0.0049797120094299315, 0.004967423915863037, 0.00496127986907959, 0.004981760025024414, 0.004943871974945068, 0.0049623680114746094, 0.004978623867034912, 0.004967423915863037, 0.004967423915863037, 0.004952127933502197, 0.005006271839141846, 0.0050032639503479, 0.004969471931457519, 0.004956192016601563, 0.005013472080230713, 0.004978687763214112, 0.005000192165374756, 0.010371071815490723, 0.004695040225982666, 0.0047185921669006346, 0.004679711818695068, 0.004683743953704834, 0.004731904029846192, 0.004683775901794434, 0.004679679870605469, 0.004703231811523437, 0.004663296222686767, 0.004685823917388916, 0.0046991357803344725, 0.004680704116821289, 0.004716544151306152, 0.004674560070037842, 0.004694015979766846, 0.00471347188949585, 0.004679679870605469, 0.0046888961791992185, 0.004683775901794434, 0.0046888961791992185, 0.0046919679641723635, 0.004694015979766846, 0.0046919679641723635, 0.004703231811523437, 0.004685823917388916, 0.004701183795928955, 0.004687903881072998, 0.004727776050567627, 0.0046888961791992185, 0.0046888961791992185, 0.004685823917388916, 0.004681727886199951, 0.004695040225982666, 0.004703231811523437, 0.004724768161773682, 0.004707295894622803, 0.004676608085632325, 0.00466534423828125, 0.004696063995361328, 0.004670464038848877, 0.004690944194793701, 0.0046694397926330565, 0.004678656101226806, 0.004678656101226806, 0.004709375858306885, 0.004681727886199951, 0.004682752132415771, 0.004676608085632325, 0.004670464038848877, 0.00465715217590332, 0.004695040225982666, 0.004677631855010986, 0.0046694397926330565, 0.0046622719764709475, 0.0046622719764709475, 0.004685823917388916, 0.00471449613571167, 0.004674560070037842, 0.004676608085632325, 0.004685823917388916, 0.00467251205444336, 0.00466431999206543, 0.01028819179534912, 0.004674496173858643, 0.004673535823822021, 0.004695040225982666, 0.0046888961791992185, 0.004679679870605469, 0.004674560070037842, 0.004660223960876465, 0.004665408134460449, 0.004706240177154541, 0.004668416023254395, 0.004679679870605469, 0.0046561279296875, 0.0046561279296875, 0.00466534423828125, 0.004700160026550293, 0.004658239841461182, 0.004682688236236572, 0.004687871932983399, 0.004645952224731445, 0.0046663999557495115, 0.004707231998443604, 0.004675648212432861, 0.004669375896453857, 0.004670464038848877, 0.004675583839416504, 0.004661248207092285, 0.00469814395904541, 0.004681695938110351, 0.0046929922103881834, 0.004711423873901367, 0.004680704116821289, 0.004666368007659912, 0.004678656101226806, 0.004668416023254395, 0.004682752132415771, 0.004803584098815918, 0.004682752132415771, 0.004653056144714355, 0.004696063995361328, 0.0046919679641723635, 0.004673535823822021, 0.00466534423828125, 0.00463974380493164, 0.004631552219390869, 0.004644864082336426, 0.004636672019958496, 0.0046284799575805665, 0.004631552219390869, 0.004626431941986084, 0.0046356477737426755, 0.004688928127288819, 0.0046560959815979, 0.00469708776473999, 0.004623360157012939, 0.004632575988769531, 0.0046254081726074215, 0.00463974380493164, 0.0046694397926330565, 0.004702208042144776, 0.004667391777038574, 0.004676608085632325, 0.004671487808227539, 0.010304512023925782, 0.004685823917388916, 0.004689919948577881, 0.004676608085632325, 0.004686848163604736, 0.004702208042144776, 0.004680704116821289, 0.004667391777038574, 0.0046561279296875, 0.0046561279296875, 0.004683775901794434, 0.0047226881980895995, 0.005391359806060791, 0.005395455837249756, 0.005496831893920898, 0.005044223785400391, 0.004983808040618896, 0.005035039901733399, 0.004988895893096924, 0.0049827837944030765, 0.004996096134185791, 0.005032959938049316, 0.005000192165374756, 0.004996096134185791, 0.004989952087402344, 0.004996096134185791, 0.004955135822296143, 0.0049725441932678225, 0.004958208084106445, 0.004997119903564453, 0.004957183837890625, 0.005006464004516602, 0.004940671920776367, 0.00496127986907959, 0.005092351913452148, 0.004964352130889893, 0.004960256099700928, 0.004983808040618896, 0.004971519947052002, 0.004966400146484375, 0.004955135822296143, 0.004987904071807861, 0.0049500160217285155, 0.004951039791107178, 0.004985856056213379, 0.004974688053131103, 0.005358496189117432, 0.004964352130889893, 0.004943871974945068, 0.0049827837944030765, 0.00496127986907959, 0.0049459199905395506, 0.004946944236755371, 0.0050135040283203125, 0.004973567962646484, 0.004967423915863037, 0.004963327884674072, 0.004971519947052002, 0.004955135822296143, 0.004975679874420166, 0.00497657585144043, 0.004976640224456787, 0.0049500160217285155, 0.010941439628601075, 0.00497049617767334, 0.004964352130889893, 0.0049500160217285155, 0.004947008132934571, 0.005011392116546631, 0.005238783836364746, 0.00501964807510376, 0.0050135040283203125, 0.004999167919158935, 0.0050135040283203125, 0.0050022401809692385, 0.004996096134185791, 0.0049500160217285155, 0.004967423915863037, 0.004964352130889893, 0.00497049617767334, 0.004973567962646484, 0.005004288196563721, 0.004947968006134033, 0.004965375900268554, 0.00501964807510376, 0.004977663993835449, 0.004956160068511963, 0.0049428482055664065, 0.004980735778808594, 0.0049489917755126955, 0.004966432094573974, 0.00496124792098999, 0.004993023872375488, 0.004963327884674072, 0.004966400146484375, 0.004988927841186524, 0.004967423915863037, 0.004971519947052002, 0.004966400146484375, 0.004939775943756103, 0.004988927841186524, 0.004994048118591309, 0.004947968006134033, 0.0049469761848449706, 0.005011424064636231, 0.0050022401809692385, 0.0049725441932678225, 0.004961343765258789, 0.004974527835845947, 0.004957183837890625, 0.004996096134185791, 0.004956160068511963, 0.004975615978240967, 0.004956160068511963, 0.004981823921203613, 0.004953023910522461, 0.0049489917755126955, 0.005021696090698242, 0.004991040229797363, 0.005002175807952881, 0.004964352130889893, 0.005024767875671386, 0.004981760025024414, 0.004977663993835449, 0.004973567962646484, 0.004992000102996826, 0.010962944030761718, 0.005007359981536865, 0.00497049617767334, 0.004943871974945068, 0.004962336063385009, 0.00498582410812378, 0.0049428482055664065, 0.0049459199905395506, 0.004980735778808594, 0.0049725441932678225, 0.004994048118591309, 0.004947968006134033, 0.004956160068511963, 0.004969471931457519, 0.004966400146484375, 0.0049387521743774416, 0.004966400146484375, 0.0049827837944030765, 0.00494598388671875, 0.004952000141143799, 0.004947968006134033, 0.004983808040618896, 0.004944896221160889, 0.004949088096618652, 0.0049622077941894535, 0.0049500160217285155, 0.00499510383605957, 0.004972511768341064, 0.004959231853485108, 0.004958208084106445, 0.004980735778808594, 0.004956223964691162, 0.004955071926116944, 0.0049428482055664065, 0.004969471931457519, 0.0049827837944030765, 0.004958208084106445, 0.004981760025024414, 0.00496127986907959, 0.004999167919158935, 0.005646336078643799, 0.005000192165374756, 0.0049500160217285155, 0.0049725441932678225, 0.0049725441932678225, 0.004955135822296143, 0.004999167919158935, 0.00496127986907959, 0.004963327884674072, 0.004971519947052002, 0.005002336025238037, 0.0049407038688659664, 0.004978687763214112, 0.004967423915863037, 0.004951039791107178, 0.0050022401809692385, 0.00497049617767334, 0.004958208084106445, 0.004984831809997559, 0.00496230411529541, 0.004958208084106445, 0.004964352130889893, 0.004960256099700928, 0.01093120002746582, 0.004985856056213379, 0.004956160068511963, 0.004959231853485108, 0.004964352130889893, 0.0049909758567810054, 0.0049500160217285155, 0.005033984184265137, 0.004980735778808594, 0.00517632007598877, 0.0050135040283203125, 0.004992000102996826, 0.00497049617767334, 0.00502784013748169, 0.00501964807510376, 0.004993023872375488, 0.00496127986907959, 0.005008384227752686, 0.004987904071807861, 0.004984831809997559, 0.0049316477775573735, 0.004972479820251465, 0.004956160068511963, 0.004951039791107178, 0.004960256099700928, 0.004994048118591309, 0.00499513578414917, 0.004973504066467285, 0.0049459199905395506, 0.00495411205291748, 0.004994048118591309, 0.0049797120094299315, 0.0049428482055664065, 0.004959231853485108, 0.004924416065216064, 0.00491212797164917, 0.004874239921569825, 0.004888576030731201, 0.00491212797164917, 0.004946944236755371, 0.004894720077514648, 0.004904960155487061, 0.005119999885559082, 0.004955135822296143, 0.004970560073852539, 0.004945856094360351, 0.004975615978240967, 0.004952064037322998, 0.004958208084106445, 0.004966400146484375, 0.004952064037322998, 0.0050063362121582035, 0.004992000102996826, 0.004965375900268554, 0.004986879825592041, 0.004980735778808594, 0.004977663993835449, 0.004964352130889893, 0.00495411205291748, 0.005389311790466309, 0.004977663993835449, 0.0049879360198974605, 0.0049786558151245115, 0.01033625602722168, 0.004702208042144776, 0.004678656101226806, 0.004690944194793701, 0.004661248207092285, 0.004683775901794434, 0.004661248207092285, 0.0046592001914978025, 0.004661248207092285, 0.004703231811523437, 0.004674560070037842, 0.004671487808227539, 0.00466534423828125, 0.004683775901794434, 0.0046694397926330565, 0.004723711967468262, 0.004724736213684082, 0.004767744064331054, 0.004716544151306152, 0.0047185921669006346, 0.004746240139007568, 0.004702271938323975, 0.004668384075164795, 0.004682720184326172, 0.004678656101226806, 0.0046561279296875, 0.004690944194793701, 0.00489577579498291, 0.004986847877502441, 0.004984831809997559, 0.005001215934753418, 0.0049530878067016604, 0.004968448162078858, 0.0049725441932678225, 0.0050063362121582035, 0.004997119903564453, 0.0049909758567810054, 0.004937727928161621, 0.004955135822296143, 0.004980735778808594, 0.0049797120094299315, 0.004963391780853271, 0.004947904109954834, 0.0049797120094299315, 0.004960256099700928, 0.004955135822296143, 0.004951039791107178, 0.004967423915863037, 0.004956160068511963, 0.004985856056213379, 0.004937727928161621, 0.00496127986907959, 0.004935679912567138, 0.004963327884674072, 0.004936704158782959, 0.004987904071807861, 0.0049695358276367185, 0.004972479820251465, 0.005102591991424561, 0.005063680171966553, 0.004999167919158935, 0.004966400146484375, 0.004966400146484375, 0.010396736145019531, 0.004684735774993896, 0.004666368007659912, 0.00466431999206543, 0.004674560070037842, 0.004703231811523437, 0.0046919679641723635, 0.00466534423828125, 0.004677631855010986, 0.00466534423828125, 0.004708352088928222, 0.0046888961791992185, 0.0046622719764709475, 0.0046581759452819825, 0.004683775901794434, 0.0046581759452819825, 0.0046929922103881834, 0.004671487808227539, 0.004666431903839112, 0.00467142391204834, 0.004673535823822021, 0.004682752132415771, 0.004686848163604736, 0.004682752132415771, 0.004679679870605469, 0.004689919948577881, 0.004709375858306885, 0.004702208042144776, 0.004690944194793701, 0.004687871932983399, 0.004689919948577881, 0.00467251205444336, 0.004668416023254395, 0.004680704116821289, 0.004684800148010254, 0.004668416023254395, 0.004678688049316406, 0.004682720184326172, 0.004690944194793701, 0.004685823917388916, 0.004717567920684815, 0.00467251205444336, 0.004682752132415771, 0.004677696228027344, 0.004673471927642822, 0.004681727886199951, 0.0046919679641723635, 0.004673535823822021, 0.004716544151306152, 0.0046622719764709475, 0.004673535823822021, 0.004700160026550293, 0.004680704116821289, 0.004681727886199951, 0.00467251205444336, 0.0046561279296875, 0.004678656101226806, 0.004710400104522705, 0.004704256057739258, 0.004698112010955811, 0.004676608085632325, 0.004665408134460449, 0.0047400321960449215, 0.010270719528198241, 0.004685823917388916, 0.00467251205444336, 0.004701183795928955, 0.004676608085632325, 0.004710400104522705, 0.004678656101226806, 0.004681727886199951, 0.004667391777038574, 0.004702208042144776, 0.004678656101226806, 0.004673535823822021, 0.004676608085632325, 0.0046694397926330565, 0.0046694397926330565, 0.004689919948577881, 0.004683775901794434, 0.0046694397926330565, 0.00468281602859497, 0.004673471927642822, 0.004670464038848877, 0.004685823917388916, 0.0046715521812438964, 0.004660160064697266, 0.004662335872650147, 0.004670400142669678, 0.004724736213684082, 0.004695040225982666, 0.004712448120117187, 0.00467251205444336, 0.004679679870605469, 0.004693056106567383, 0.004698048114776612, 0.004684800148010254, 0.0046622719764709475, 0.00467251205444336, 0.00466431999206543, 0.0046694397926330565, 0.0046929922103881834, 0.004678656101226806, 0.004667391777038574, 0.00471347188949585, 0.004675583839416504, 0.004661248207092285, 0.004696063995361328, 0.0046622719764709475, 0.004682752132415771, 0.004675615787506104, 0.00467964792251587, 0.0046694397926330565, 0.0046888961791992185, 0.00464896011352539, 0.004680704116821289, 0.0046561279296875, 0.004663296222686767, 0.004652031898498535, 0.004866047859191895, 0.004696063995361328, 0.0046694397926330565, 0.0046592001914978025, 0.004873216152191162, 0.0047185921669006346, 0.004674560070037842, 0.010381312370300292, 0.0046929922103881834, 0.004715519905090332, 0.004690944194793701, 0.004695040225982666, 0.004683775901794434, 0.004685823917388916, 0.004704256057739258, 0.004679679870605469, 0.004682752132415771, 0.004710400104522705, 0.004685823917388916, 0.00505241584777832, 0.0046991357803344725, 0.00467251205444336, 0.0046694397926330565, 0.004670464038848877, 0.004680704116821289, 0.004693056106567383, 0.004693952083587647, 0.004708352088928222, 0.004681727886199951, 0.004679679870605469, 0.004667391777038574, 0.004699168205261231, 0.004682720184326172, 0.004687871932983399, 0.0046919679641723635, 0.00467251205444336, 0.004677696228027344, 0.004711359977722168, 0.004685855865478515, 0.004684768199920654, 0.004681727886199951, 0.004700160026550293, 0.004687871932983399, 0.004698112010955811, 0.0046929922103881834, 0.004710400104522705, 0.004701183795928955, 0.004695040225982666, 0.0046929922103881834, 0.004746240139007568, 0.004689919948577881, 0.004677631855010986, 0.004681727886199951, 0.004683775901794434, 0.004679679870605469, 0.004685823917388916, 0.00469708776473999, 0.004689919948577881, 0.004724736213684082, 0.004733952045440673, 0.004770815849304199, 0.00471347188949585, 0.004731904029846192, 0.004695072174072266, 0.00467964792251587, 0.00467251205444336, 0.004702208042144776, 0.004675583839416504, 0.004675583839416504, 0.00466431999206543, 0.01036083221435547, 0.0046929922103881834, 0.004674560070037842, 0.004666368007659912, 0.00467251205444336, 0.00470527982711792, 0.004666368007659912, 0.004685823917388916, 0.004676608085632325, 0.004674560070037842, 0.004673535823822021, 0.0046929922103881834, 0.004674560070037842, 0.00470630407333374, 0.004680704116821289, 0.004677631855010986, 0.004675583839416504, 0.004719615936279297, 0.004680704116821289, 0.004682752132415771, 0.00471449613571167, 0.004679679870605469, 0.004663296222686767, 0.004710400104522705, 0.004674560070037842, 0.00467251205444336, 0.004682752132415771, 0.004681727886199951, 0.004711423873901367, 0.004681727886199951, 0.004685887813568115, 0.004688831806182861, 0.004689919948577881, 0.00470527982711792, 0.00471449613571167, 0.004681727886199951, 0.004671487808227539, 0.0046715521812438964, 0.004675519943237305, 0.00471449613571167, 0.004683775901794434, 0.004712448120117187, 0.004680704116821289, 0.004716544151306152, 0.004685823917388916, 0.004683775901794434, 0.004697152137756348, 0.004722623825073242, 0.004668416023254395, 0.0046888961791992185, 0.004675583839416504, 0.004682752132415771, 0.004696063995361328, 0.004687871932983399, 0.004681727886199951, 0.004676608085632325, 0.004681727886199951, 0.0047257599830627444, 0.004681727886199951, 0.004685855865478515, 0.004689888000488281, 0.004684800148010254, 0.004683775901794434, 0.010320927619934082, 0.004685791969299316, 0.004675583839416504, 0.004682784080505371, 0.004712416172027588, 0.004682752132415771, 0.004687871932983399, 0.004663296222686767, 0.004726784229278564, 0.00470630407333374, 0.004694015979766846, 0.004684800148010254, 0.004683807849884033, 0.004673503875732422, 0.0046919679641723635, 0.004712448120117187, 0.004684800148010254, 0.004631552219390869, 0.004652031898498535, 0.004638720035552979, 0.004643839836120605, 0.004674560070037842, 0.004634624004364014, 0.00464793586730957, 0.004645887851715088, 0.004642816066741944, 0.0046356477737426755, 0.004689919948577881, 0.004663296222686767, 0.00471449613571167, 0.004652031898498535, 0.004646912097930909, 0.0046592001914978025, 0.004702208042144776, 0.004700160026550293, 0.005682176113128662, 0.00515993595123291, 0.0050432000160217285, 0.0049909758567810054, 0.004976640224456787, 0.00496127986907959, 0.005007359981536865, 0.005014527797698975, 0.004955135822296143, 0.005272575855255127, 0.005090303897857666, 0.005007359981536865, 0.005004288196563721, 0.00501145601272583, 0.0050032639503479, 0.004977663993835449, 0.005002272129058838, 0.004973536014556885, 0.004975615978240967, 0.005008384227752686, 0.00496230411529541, 0.004964352130889893, 0.004988927841186524, 0.004975615978240967, 0.0049428482055664065, 0.004923391819000244, 0.004996096134185791, 0.004959231853485108, 0.010363903999328614, 0.004710400104522705, 0.004719615936279297, 0.004661248207092285, 0.0047226881980895995, 0.004681727886199951, 0.004683775901794434, 0.004660223960876465, 0.004679679870605469, 0.004680704116821289, 0.004679679870605469, 0.00470527982711792, 0.004700160026550293, 0.004670464038848877, 0.00469708776473999, 0.004968448162078858, 0.0046787199974060055, 0.004639679908752441, 0.004667391777038574, 0.004631552219390869, 0.004719615936279297, 0.004671487808227539, 0.004673535823822021, 0.004666431903839112, 0.004714431762695312, 0.004723711967468262, 0.004690944194793701, 0.004682752132415771, 0.0046704959869384765, 0.005066751956939697, 0.0047247037887573244, 0.004682752132415771, 0.004682752132415771, 0.00466431999206543, 0.00466431999206543, 0.004696063995361328, 0.004681727886199951, 0.004696063995361328, 0.00467251205444336, 0.004686848163604736, 0.0046254081726074215, 0.004653056144714355, 0.004638720035552979, 0.0046254081726074215, 0.004636672019958496, 0.0046284799575805665, 0.0046295042037963864, 0.0046663999557495115, 0.00461513614654541, 0.004649983882904053, 0.004620287895202637, 0.004633600234985351, 0.004680704116821289, 0.004709375858306885, 0.0047226881980895995, 0.0047298879623413085, 0.0047144641876220705, 0.004686848163604736, 0.00466537618637085, 0.004689888000488281, 0.004651008129119873, 0.004624383926391602, 0.0046254081726074215, 0.010812416076660156, 0.004984831809997559, 0.004967455863952637, 0.00496943998336792, 0.0048855037689208985, 0.004897791862487793, 0.004894720077514648, 0.0049530878067016604, 0.0049203200340270994, 0.004903935909271241, 0.004932608127593994, 0.004895743846893311, 0.004931583881378174, 0.004960256099700928, 0.004947968006134033, 0.004939775943756103, 0.004963327884674072, 0.004967423915863037, 0.004966400146484375, 0.00497049617767334, 0.004958208084106445, 0.004966400146484375, 0.0050063362121582035, 0.004960256099700928, 0.0049459199905395506, 0.004988927841186524, 0.004976640224456787, 0.004967423915863037, 0.004965375900268554, 0.004999199867248535, 0.004988895893096924, 0.004984831809997559, 0.004966400146484375, 0.004985856056213379, 0.004956160068511963, 0.004968448162078858, 0.004940800189971924, 0.00496230411529541, 0.0049500160217285155, 0.004944896221160889, 0.0049428482055664065, 0.004958303928375244, 0.004982687950134277, 0.004951039791107178, 0.004957183837890625, 0.004993023872375488, 0.00495411205291748, 0.0049428482055664065, 0.004940832138061523, 0.005028831958770752, 0.004944896221160889, 0.004976640224456787, 0.004960256099700928, 0.0049428482055664065, 0.004975615978240967, 0.004960256099700928, 0.004936704158782959, 0.004930560111999512, 0.004975615978240967, 0.00495411205291748, 0.004946944236755371, 0.004966400146484375, 0.004974592208862305, 0.010241024017333985, 0.004670464038848877, 0.004703328132629394, 0.0046683201789855956, 0.004679679870605469, 0.00467251205444336, 0.00466534423828125, 0.004676608085632325, 0.004686848163604736, 0.004673535823822021, 0.004678656101226806, 0.004675583839416504, 0.004670464038848877, 0.004666368007659912, 0.004673535823822021, 0.00467251205444336, 0.00466534423828125, 0.004677631855010986, 0.0047084479331970215, 0.004663199901580811, 0.004684832096099854, 0.00467248010635376, 0.0046356477737426755, 0.004633600234985351, 0.0046284799575805665, 0.0046284799575805665, 0.004690944194793701, 0.004631552219390869, 0.004667391777038574, 0.00466431999206543, 0.004627456188201904, 0.004621312141418457, 0.004642816066741944, 0.004622335910797119, 0.004646912097930909, 0.00461516809463501, 0.004616191864013672, 0.004610047817230224, 0.004623360157012939, 0.0046991357803344725, 0.004689919948577881, 0.0046694397926330565, 0.0046694397926330565, 0.004690944194793701, 0.004679679870605469, 0.0047288317680358885, 0.004686848163604736, 0.004680704116821289, 0.004634624004364014, 0.00463974380493164, 0.004637695789337158, 0.004644864082336426, 0.004626431941986084, 0.004630527973175049, 0.00465715217590332, 0.004656159877777099, 0.004662240028381347, 0.004712448120117187, 0.0055203838348388675, 0.004751423835754395, 0.004674496173858643, 0.00465715217590332, 0.004637695789337158, 0.01031987190246582, 0.004689919948577881, 0.004703231811523437, 0.004700160026550293, 0.004671487808227539, 0.004673535823822021, 0.004690944194793701, 0.004673535823822021, 0.004696063995361328, 0.004686848163604736, 0.004674592018127441, 0.004696032047271728, 0.0046888961791992185, 0.0046888961791992185, 0.00466534423828125, 0.004680704116821289, 0.004671487808227539, 0.004670464038848877, 0.004680704116821289, 0.004690944194793701, 0.004684800148010254, 0.00467251205444336, 0.004671487808227539, 0.004660223960876465, 0.004679679870605469, 0.004744192123413086, 0.004680736064910889, 0.004701151847839356, 0.004704256057739258, 0.004689919948577881, 0.004698112010955811, 0.0046929922103881834, 0.0049192957878112795, 0.004748288154602051, 0.00469708776473999, 0.004689919948577881, 0.004695040225982666, 0.0046929922103881834, 0.0046919679641723635, 0.004675583839416504, 0.004745215892791748, 0.004859903812408447, 0.004769792079925537, 0.004729856014251709, 0.005243904113769531, 0.00515174388885498, 0.00520908784866333, 0.005025792121887207, 0.004997119903564453, 0.005040128231048584, 0.004974592208862305, 0.004997119903564453, 0.0049428482055664065, 0.004944896221160889, 0.00499513578414917, 0.004932544231414795, 0.0049489917755126955, 0.004955135822296143, 0.004975615978240967, 0.0049530878067016604, 0.004957183837890625, 0.004959231853485108, 0.004980735778808594, 0.010954751968383789, 0.00502784013748169, 0.004978687763214112, 0.005044223785400391, 0.004977663993835449, 0.0050032639503479, 0.004963327884674072, 0.004956160068511963, 0.004940800189971924, 0.004947968006134033, 0.004969471931457519, 0.004996096134185791, 0.004957183837890625, 0.004978687763214112, 0.004895743846893311, 0.004901887893676758, 0.004877312183380127, 0.004895743846893311, 0.004905983924865722, 0.004891647815704346, 0.004931615829467773, 0.004930592060089111, 0.005009344100952148, 0.005026815891265869, 0.0050022401809692385, 0.004996096134185791, 0.005074944019317627, 0.005021696090698242, 0.004974592208862305, 0.004897791862487793, 0.004923391819000244, 0.004918272018432617, 0.0049489917755126955, 0.00496230411529541, 0.004985856056213379, 0.005030911922454834, 0.004921343803405762, 0.0048895998001098635, 0.004898816108703613, 0.004981760025024414, 0.004973567962646484, 0.004960256099700928, 0.004987904071807861, 0.00497049617767334, 0.0049827837944030765, 0.0049469761848449706, 0.0049417920112609865, 0.004987904071807861, 0.0049459199905395506, 0.0049909758567810054, 0.004940800189971924, 0.004957183837890625, 0.004966400146484375, 0.0049500160217285155, 0.004935679912567138, 0.004992000102996826, 0.004973567962646484, 0.0049459199905395506, 0.004967423915863037, 0.004981760025024414, 0.004956192016601563, 0.004973536014556885, 0.004957183837890625]",tokens/s,202.92757705602773,,,main,False,False,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,google/gemma-2b,google/gemma-2b,cuda,0,42,,,True,,,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.223-212.873.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.0,,,,1.21.4,,,,0.12.0,,,gemma,MB,1548.333056,6797.39392,0.0,6150.946816,5419.87328,s,10,0.34557660293579096,0.034557660293579096,0.00019750828367446217,0.03442801475524902,0.034850107955932616,0.034850669288635255,0.03485111835479736,"[0.0344062385559082, 0.03483919906616211, 0.034415199279785154, 0.034409793853759765, 0.0343903694152832, 0.034380928039550784, 0.034592830657958984, 0.03485123062133789, 0.03484998321533203, 0.03444083023071289]",tokens/s,7407.908921645528,kWh,4.068712725372023e-07,2.2294371985946046e-07,1.9450072833814666e-06,2.5748222757781293e-06,tokens/kWh,99424337.907996,MB,1617.854464,6797.39392,0.0,6150.946816,5419.87584,s,10,13.92159716796875,1.3921597167968751,0.02023007918221962,1.398265441894531,1.411018359375,1.4207826782226562,1.4285941333007812,"[1.3698680419921876, 1.3984805908203124, 1.367292236328125, 1.3672281494140626, 1.39805029296875, 1.4088485107421875, 1.4305469970703124, 1.4019166259765625, 1.40419921875, 1.37516650390625]",tokens/s,45.25342835299988,kWh,1.659726406482387e-05,9.095122566918153e-06,5.263189748921836e-05,7.832428412096038e-05,tokens/kWh,804348.2389536524,,s,630,13.9198750553131,0.02209503977033827,0.0005755475498994588,0.021720064163208007,0.02271600570678711,0.022946099376678467,0.023882208786010748,"[0.021808128356933593, 0.02166169548034668, 0.02165657615661621, 0.02166886329650879, 0.021622783660888673, 0.021607423782348634, 0.021630975723266603, 0.02167091178894043, 0.02169856071472168, 0.021638143539428712, 0.021593088150024413, 0.021598207473754884, 0.021618688583374023, 0.0216944637298584, 0.021728256225585937, 0.02166783905029297, 0.02168934440612793, 0.02167500877380371, 0.021782527923583983, 0.02167193603515625, 0.021646335601806642, 0.02167807960510254, 0.021604352951049805, 0.02166783905029297, 0.021702655792236326, 0.02161664009094238, 0.0216484489440918, 0.021629888534545897, 0.02225766372680664, 0.021832704544067383, 0.02167807960510254, 0.02164121627807617, 0.021562368392944335, 0.021687295913696288, 0.021703680038452147, 0.02168422317504883, 0.0216627197265625, 0.02166886329650879, 0.021699583053588867, 0.02166988754272461, 0.02171801567077637, 0.02168627166748047, 0.021644287109375, 0.02167193603515625, 0.02166579246520996, 0.021731327056884766, 0.021803007125854493, 0.021715967178344727, 0.021659648895263672, 0.0216944637298584, 0.021696512222290038, 0.02163711929321289, 0.0216494083404541, 0.021651456832885742, 0.02164735984802246, 0.02165043258666992, 0.021607423782348634, 0.02165350341796875, 0.021625856399536132, 0.021588991165161133, 0.02168217658996582, 0.024484960556030274, 0.022811552047729493, 0.02183782386779785, 0.02168832015991211, 0.021630975723266603, 0.02166579246520996, 0.02166886329650879, 0.0216494083404541, 0.02166783905029297, 0.021554176330566405, 0.02145996856689453, 0.021555200576782226, 0.022889471054077147, 0.023215103149414062, 0.022715391159057616, 0.022543359756469726, 0.022550527572631835, 0.022647808074951172, 0.022575103759765625, 0.022525951385498046, 0.02254745674133301, 0.022602752685546876, 0.022530080795288086, 0.022583263397216797, 0.02254649543762207, 0.022559680938720704, 0.02248089599609375, 0.02255564880371094, 0.02246451187133789, 0.02267238426208496, 0.023768064498901367, 0.02281881523132324, 0.02259971237182617, 0.022700000762939453, 0.02262835121154785, 0.022594560623168947, 0.02266726493835449, 0.022580223083496095, 0.022611967086791994, 0.022550527572631835, 0.02260479927062988, 0.021702655792236326, 0.02164838409423828, 0.022377471923828125, 0.022599679946899414, 0.022600704193115235, 0.022181888580322266, 0.022687744140625, 0.02186137580871582, 0.0216494083404541, 0.021630975723266603, 0.02168524742126465, 0.021605375289916993, 0.02169856071472168, 0.021545984268188476, 0.021436416625976562, 0.02165760040283203, 0.02167193603515625, 0.021621759414672852, 0.022125568389892578, 0.021982208251953125, 0.02169241523742676, 0.021584896087646483, 0.021659648895263672, 0.02169856071472168, 0.021696512222290038, 0.02168115234375, 0.02167500877380371, 0.021739519119262696, 0.02163302421569824, 0.021593088150024413, 0.021831680297851562, 0.02164838409423828, 0.0216494083404541, 0.021790719985961913, 0.02173030471801758, 0.021720064163208007, 0.021704704284667968, 0.021644287109375, 0.021644287109375, 0.02165043258666992, 0.02164838409423828, 0.021651456832885742, 0.021695487976074217, 0.02164531135559082, 0.021666879653930663, 0.021662656784057616, 0.021639167785644533, 0.02169856071472168, 0.021729280471801758, 0.02169241523742676, 0.02168934440612793, 0.021704704284667968, 0.021642240524291992, 0.02168627166748047, 0.021727231979370116, 0.021612543106079102, 0.0216944637298584, 0.021619712829589844, 0.02168012809753418, 0.0216627197265625, 0.021639167785644533, 0.022392831802368163, 0.021679103851318358, 0.02168627166748047, 0.022041599273681642, 0.02211737632751465, 0.02167705535888672, 0.02168217658996582, 0.0216627197265625, 0.021644287109375, 0.02165760040283203, 0.02164531135559082, 0.021720064163208007, 0.021659648895263672, 0.021598207473754884, 0.02165452766418457, 0.02166476821899414, 0.021750783920288085, 0.021630975723266603, 0.021716991424560548, 0.02169343948364258, 0.02169856071472168, 0.021716991424560548, 0.021766143798828123, 0.02164019203186035, 0.021629951477050782, 0.021579776763916016, 0.02171801567077637, 0.02164838409423828, 0.02162892723083496, 0.02168832015991211, 0.021560319900512694, 0.02166681671142578, 0.02166988754272461, 0.021731327056884766, 0.021639167785644533, 0.021753856658935547, 0.022533119201660155, 0.02168012809753418, 0.021638143539428712, 0.02163711929321289, 0.02163206481933594, 0.021658559799194337, 0.021738496780395508, 0.021703680038452147, 0.021725183486938478, 0.02165452766418457, 0.021635072708129883, 0.02162483215332031, 0.021721088409423828, 0.02186854362487793, 0.02231808090209961, 0.021707775115966797, 0.021643264770507813, 0.02166886329650879, 0.02169856071472168, 0.021720064163208007, 0.021721088409423828, 0.02164121627807617, 0.021634048461914062, 0.02169753646850586, 0.021696512222290038, 0.02165043258666992, 0.021659648895263672, 0.02165043258666992, 0.021643264770507813, 0.021627904891967774, 0.02170163154602051, 0.021606399536132814, 0.02165452766418457, 0.021679103851318358, 0.02163711929321289, 0.0217262077331543, 0.02169139289855957, 0.02168627166748047, 0.02167398452758789, 0.02168320083618164, 0.02167296028137207, 0.02167296028137207, 0.02170675277709961, 0.02168320083618164, 0.02166476821899414, 0.02162483215332031, 0.021643264770507813, 0.021704704284667968, 0.02165862464904785, 0.02166374397277832, 0.02168422317504883, 0.021696512222290038, 0.021711872100830077, 0.021729248046875, 0.021646335601806642, 0.02160233688354492, 0.02172105598449707, 0.02167705535888672, 0.02165862464904785, 0.02163302421569824, 0.022205440521240235, 0.021816320419311523, 0.0217262077331543, 0.02167193603515625, 0.021733375549316408, 0.021638143539428712, 0.021716991424560548, 0.021773311614990236, 0.02169139289855957, 0.021799936294555664, 0.021687295913696288, 0.022388736724853517, 0.022238208770751954, 0.02224332809448242, 0.022487039566040038, 0.02266316795349121, 0.022616064071655274, 0.022598655700683593, 0.022639616012573242, 0.02262835121154785, 0.022633472442626954, 0.022593536376953126, 0.02368921661376953, 0.023427072525024413, 0.02289664077758789, 0.02272870445251465, 0.022722560882568358, 0.022599679946899414, 0.022573055267333983, 0.022617088317871094, 0.02262118339538574, 0.02310655975341797, 0.022605823516845702, 0.022583295822143554, 0.021609472274780273, 0.02168115234375, 0.02165350341796875, 0.022600704193115235, 0.02169753646850586, 0.021646335601806642, 0.021619775772094726, 0.021650367736816407, 0.021732351303100587, 0.02168832015991211, 0.02172313690185547, 0.021778432846069336, 0.022700031280517577, 0.022928384780883788, 0.02265907287597656, 0.0226713924407959, 0.022588447570800783, 0.022626239776611327, 0.022581247329711913, 0.021679103851318358, 0.02164735984802246, 0.02168524742126465, 0.021708799362182618, 0.021727231979370116, 0.02168217658996582, 0.0216944637298584, 0.021715967178344727, 0.021795839309692384, 0.02167193603515625, 0.021643264770507813, 0.022533119201660155, 0.022527999877929687, 0.022591487884521484, 0.022545408248901368, 0.022552576065063477, 0.022553600311279298, 0.021847103118896486, 0.02175276756286621, 0.02165657615661621, 0.02163711929321289, 0.021679103851318358, 0.021642240524291992, 0.021711872100830077, 0.02164531135559082, 0.021678112030029298, 0.02175484848022461, 0.021751808166503905, 0.021793792724609375, 0.022090751647949217, 0.02162073516845703, 0.02165555191040039, 0.021651456832885742, 0.02244607925415039, 0.02248192024230957, 0.022700031280517577, 0.02264067268371582, 0.02274198341369629, 0.02275328063964844, 0.02312396812438965, 0.021799936294555664, 0.021939199447631837, 0.022055936813354493, 0.02250547218322754, 0.022673408508300782, 0.024655872344970704, 0.02532352066040039, 0.02289664077758789, 0.022268928527832032, 0.022639616012573242, 0.022605823516845702, 0.022610944747924806, 0.022586368560791017, 0.022573055267333983, 0.022326271057128907, 0.023516159057617187, 0.02411008071899414, 0.02266828727722168, 0.022642688751220705, 0.02264473533630371, 0.02262015914916992, 0.022799360275268556, 0.022788095474243163, 0.022757375717163086, 0.022563840866088865, 0.022692863464355468, 0.02271232032775879, 0.02259660720825195, 0.02264575958251953, 0.022574079513549804, 0.02255462455749512, 0.02247065544128418, 0.022738943099975584, 0.022533119201660155, 0.022460416793823244, 0.02254643249511719, 0.022585344314575196, 0.024655872344970704, 0.02372198486328125, 0.02307788848876953, 0.02273587226867676, 0.022751232147216797, 0.021955583572387697, 0.023217151641845703, 0.023002111434936523, 0.02269593620300293, 0.02270412826538086, 0.022590463638305663, 0.02253209686279297, 0.022593536376953126, 0.02270412826538086, 0.02262835121154785, 0.02262937545776367, 0.022561792373657227, 0.022656000137329102, 0.022572032928466795, 0.02260479927062988, 0.022591487884521484, 0.022797311782836914, 0.02266111946105957, 0.022597631454467772, 0.022605823516845702, 0.022568960189819336, 0.02191974449157715, 0.02347110366821289, 0.02266111946105957, 0.022631423950195313, 0.022610944747924806, 0.02252288055419922, 0.02261299133300781, 0.022580223083496095, 0.022609920501708985, 0.02249113655090332, 0.022405120849609376, 0.022600704193115235, 0.02307583999633789, 0.022994943618774414, 0.02291097640991211, 0.022801408767700194, 0.02273689651489258, 0.022598720550537108, 0.022546367645263674, 0.022614015579223632, 0.02262015914916992, 0.02225868797302246, 0.02230067253112793, 0.022777856826782225, 0.022632448196411133, 0.02353971290588379, 0.02169856071472168, 0.021699583053588867, 0.021634048461914062, 0.021626880645751953, 0.02161664009094238, 0.021578752517700195, 0.021635072708129883, 0.021673023223876955, 0.02159507179260254, 0.02163302421569824, 0.02164735984802246, 0.021583871841430666, 0.022602752685546876, 0.022533119201660155, 0.022574079513549804, 0.022567935943603516, 0.022533119201660155, 0.02252288055419922, 0.022544384002685547, 0.022982656478881838, 0.02272153663635254, 0.022583295822143554, 0.022689792633056642, 0.022598655700683593, 0.022536191940307617, 0.022731775283813475, 0.022597631454467772, 0.02264371109008789, 0.022602752685546876, 0.02251468849182129, 0.02253926467895508, 0.02253926467895508, 0.023219200134277345, 0.022775808334350587, 0.022683647155761717, 0.02264678382873535, 0.02272870445251465, 0.022578176498413087, 0.022594560623168947, 0.022556671142578123, 0.02265088081359863, 0.022589439392089843, 0.022568960189819336, 0.022557695388793944, 0.02221670341491699, 0.021935104370117187, 0.021613567352294923, 0.021998592376708984, 0.02166067123413086, 0.02169343948364258, 0.02165452766418457, 0.021711872100830077, 0.02164121627807617, 0.02168627166748047, 0.021710847854614256, 0.02169036865234375, 0.021710847854614256, 0.021743616104125976, 0.021634048461914062, 0.02169856071472168, 0.023399423599243165, 0.023589887619018556, 0.022822912216186524, 0.022635520935058592, 0.022642688751220705, 0.022616064071655274, 0.022553600311279298, 0.022566911697387695, 0.022642688751220705, 0.022681600570678712, 0.022550527572631835, 0.022545408248901368, 0.022573055267333983, 0.022579200744628908, 0.022624256134033204, 0.022536191940307617, 0.022573055267333983, 0.02345062446594238, 0.02310860824584961, 0.023609344482421874, 0.02285264015197754, 0.022671327590942383, 0.022631423950195313, 0.022598655700683593, 0.02294272041320801, 0.022549503326416014, 0.022494239807128905, 0.02255766487121582, 0.022817792892456053, 0.023001087188720702, 0.023928831100463867, 0.02285977554321289, 0.02271334457397461, 0.02169343948364258, 0.02169139289855957, 0.021607423782348634, 0.021679103851318358, 0.02168524742126465, 0.021629951477050782, 0.02162892723083496, 0.021625856399536132, 0.02172211265563965, 0.022568960189819336, 0.022544384002685547, 0.02161664009094238, 0.021615615844726564, 0.02166374397277832, 0.02163302421569824, 0.022625280380249024, 0.021687295913696288, 0.021611520767211914, 0.021621759414672852, 0.02253209686279297, 0.022560768127441407, 0.022533119201660155, 0.022541311264038084, 0.0216760311126709, 0.021618688583374023, 0.02165760040283203, 0.02163609504699707, 0.02165043258666992, 0.02165456008911133, 0.021706720352172852, 0.02168320083618164, 0.02165043258666992, 0.02167807960510254, 0.02164019203186035, 0.0217262077331543, 0.021627904891967774, 0.02165555191040039, 0.02164121627807617, 0.02162892723083496, 0.02165247917175293, 0.02167193603515625, 0.022467584609985353, 0.022023168563842774, 0.021761024475097656, 0.02162380790710449, 0.021615615844726564, 0.02164735984802246, 0.021609472274780273, 0.021643264770507813, 0.02164121627807617, 0.02166886329650879, 0.02167398452758789, 0.02167193603515625, 0.021702655792236326, 0.02165043258666992, 0.021627904891967774, 0.0216944637298584, 0.02184806442260742, 0.02168217658996582, 0.021607423782348634, 0.021578752517700195, 0.021699583053588867, 0.021695487976074217, 0.02164838409423828, 0.02167500877380371, 0.021596160888671875, 0.02166476821899414, 0.021658655166625976, 0.021648351669311523, 0.02168320083618164, 0.021642240524291992, 0.021611520767211914, 0.02168934440612793, 0.021621759414672852, 0.02187059211730957, 0.021627904891967774, 0.021613567352294923, 0.02168832015991211, 0.02167091178894043, 0.02186137580871582, 0.02183782386779785, 0.02181427192687988, 0.02169856071472168, 0.02167807960510254, 0.021724159240722657, 0.02333286476135254, 0.024927232742309572, 0.022948863983154297, 0.022831104278564454, 0.02244607925415039, 0.02187980842590332, 0.021740543365478517, 0.02163199996948242, 0.02164838409423828, 0.02164838409423828, 0.021644287109375]",tokens/s,45.25902693067161,,,,,,MB,934.121472,6768.033792,0.0,6138.363904,6060.931072,s,1,6.98855615234375,6.98855615234375,0.0,6.98855615234375,6.98855615234375,6.98855615234375,6.98855615234375,[6.98855615234375],,kWh,1.9968520284723784e-06,1.0784003533201232e-06,3.6133362239995637e-06,6.688588605792065e-06, bfloat16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,s,s,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/s/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6694737c-1b7f88db07f5900351a51119;c5600467-201d-45eb-9b2e-f9b2bd21e23b) Repository Not Found for url: https://huggingface.co/s/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: s is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-40b,tiiuae/falcon-40b,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ self.load_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 250, in load_model_with_no_weights self.pretrained_model = self.automodel_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 559, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3704, in from_pretrained config = cls._autoset_attn_implementation( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1481, in _autoset_attn_implementation cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1572, in _check_and_enable_flash_attn_2 raise ValueError( ValueError: FalconForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpxu37mxk2/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,google/gemma-7b,google/gemma-7b,cuda,0,42,,,True,,,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.223-212.873.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.0,,,,1.21.4,,,,0.12.0,,,gemma,MB,1513.836544,19300.614144,0.0,18654.16704,17596.665856,s,10,0.9678243331909179,0.0967824333190918,0.00022752553448988485,0.09668753814697266,0.0969394401550293,0.09717565650939941,0.0973646295928955,"[0.096662109375, 0.09741187286376952, 0.09674845123291016, 0.09685391998291015, 0.09669939422607422, 0.09664739227294922, 0.09662448120117187, 0.09688694763183593, 0.09661408233642578, 0.0966756820678711]",tokens/s,2645.1081174614374,kWh,1.1433838122261771e-06,6.265148640965145e-07,6.3110600702115226e-06,8.080958746534214e-06,tokens/kWh,31679409.341100033,MB,1576.194048,19302.711296,0.0,18656.264192,17596.668416,s,10,31.100587158203126,3.1100587158203123,0.005821752515242594,3.1100438232421874,3.1170797119140623,3.118783142089844,3.120145886230469,"[3.109525146484375, 3.106936279296875, 3.1105625, 3.116701171875, 3.102742919921875, 3.11375048828125, 3.120486572265625, 3.10103076171875, 3.1057978515625, 3.113053466796875]",tokens/s,20.25685228369814,kWh,3.662836542673222e-05,2.0074121935414773e-05,0.00014339067080218926,0.00020009315816433624,tokens/kWh,314853.34420209506,,s,630,31.098767349243197,0.04936312277657645,0.00027610899523596976,0.049368576049804686,0.049692774963378906,0.04978790283203125,0.05009594440460205,"[0.049397758483886715, 0.04917452621459961, 0.04925439834594727, 0.049430526733398435, 0.04904243087768555, 0.04905881500244141, 0.049124351501464845, 0.049012737274169924, 0.049006591796875, 0.049296382904052735, 0.049345535278320314, 0.049539070129394534, 0.0491038703918457, 0.04914585494995117, 0.049122303009033204, 0.04903424072265625, 0.04909260940551758, 0.04944588851928711, 0.04964147186279297, 0.049519615173339845, 0.049509376525878904, 0.04953497695922852, 0.049465343475341796, 0.049503231048583986, 0.049667072296142575, 0.049560577392578124, 0.049097728729248044, 0.04926873779296875, 0.04913151931762695, 0.04919398498535156, 0.04917145538330078, 0.0492564468383789, 0.04906086349487305, 0.04899430465698242, 0.04913459014892578, 0.04912332916259766, 0.049314815521240236, 0.04935475158691406, 0.049345535278320314, 0.04916428756713867, 0.04915609741210938, 0.049083393096923826, 0.049154048919677736, 0.04909056091308594, 0.049185791015625, 0.04916121673583984, 0.04950835037231445, 0.04936601638793945, 0.04983193588256836, 0.049734657287597656, 0.04978790283203125, 0.05015039825439453, 0.04985651016235351, 0.04961075210571289, 0.049604606628417966, 0.04986880111694336, 0.049734657287597656, 0.04967628860473633, 0.049555454254150394, 0.04944076919555664, 0.049783809661865235, 0.04914176177978516, 0.0492308464050293, 0.04938547134399414, 0.049306625366210936, 0.04909568023681641, 0.04901785659790039, 0.04905779266357422, 0.049137664794921876, 0.04903014373779297, 0.049012737274169924, 0.049068031311035154, 0.04947353744506836, 0.04949913787841797, 0.0495022087097168, 0.049279998779296875, 0.04942131042480469, 0.04946432113647461, 0.049391616821289064, 0.04940390396118164, 0.04961382293701172, 0.049568767547607424, 0.04950732803344727, 0.049565696716308595, 0.049288192749023435, 0.04936294555664063, 0.04907622528076172, 0.04956159973144531, 0.04958617782592773, 0.049582080841064455, 0.04937113571166992, 0.0490250244140625, 0.049188865661621096, 0.04907929611206055, 0.04917657470703125, 0.049127422332763675, 0.04923494338989258, 0.04916326522827148, 0.049137664794921876, 0.04940083312988281, 0.04954009628295898, 0.04955136108398438, 0.049539070129394534, 0.04935987091064453, 0.04958720016479492, 0.04912947082519531, 0.04918374252319336, 0.04913151931762695, 0.04907929611206055, 0.0492410888671875, 0.04940902328491211, 0.04953702545166016, 0.04916428756713867, 0.049168384552001954, 0.04922777557373047, 0.049084415435791014, 0.04947558212280274, 0.049320960998535154, 0.04966400146484375, 0.04936703872680664, 0.049203201293945314, 0.04914790344238281, 0.049154048919677736, 0.049309696197509766, 0.04934041595458984, 0.0496814079284668, 0.04931071853637695, 0.04939059066772461, 0.0495206413269043, 0.049539070129394534, 0.04943360137939453, 0.04942438507080078, 0.04915097427368164, 0.049957889556884766, 0.049718273162841796, 0.04944588851928711, 0.04912025451660156, 0.04904857635498047, 0.04904243087768555, 0.049035263061523435, 0.04916428756713867, 0.048979969024658204, 0.04909056091308594, 0.0490885124206543, 0.049018878936767575, 0.049140735626220705, 0.04955551910400391, 0.04955744171142578, 0.04909568023681641, 0.048966655731201174, 0.049546241760253906, 0.04901683044433594, 0.04917145538330078, 0.04905779266357422, 0.04904652786254883, 0.04950527954101563, 0.04948889541625977, 0.04905062484741211, 0.04919500732421875, 0.04919500732421875, 0.049223678588867184, 0.0496517105102539, 0.04953395080566406, 0.04954111862182617, 0.04945100784301758, 0.04905881500244141, 0.04950527954101563, 0.04977766418457031, 0.04966092681884766, 0.04954111862182617, 0.04965478515625, 0.049203201293945314, 0.04910182571411133, 0.04962508773803711, 0.049460224151611325, 0.049116161346435545, 0.04961280059814453, 0.04957593536376953, 0.04975513458251953, 0.049509376525878904, 0.04961280059814453, 0.049527809143066405, 0.04934143829345703, 0.0491673583984375, 0.04954009628295898, 0.04960563278198242, 0.04970393753051758, 0.04961280059814453, 0.049631233215332034, 0.0496814079284668, 0.04941209411621094, 0.04942438507080078, 0.04943462371826172, 0.04954111862182617, 0.049492992401123044, 0.04962713623046875, 0.04947353744506836, 0.049522689819335934, 0.049737728118896485, 0.04953702545166016, 0.049503231048583986, 0.049793022155761715, 0.0490885124206543, 0.04952985763549805, 0.04955033493041992, 0.049858558654785154, 0.04956159973144531, 0.04941209411621094, 0.04947148895263672, 0.0494919662475586, 0.049601535797119144, 0.049544193267822265, 0.04952678298950195, 0.04905062484741211, 0.049324031829833984, 0.04905779266357422, 0.04910899353027344, 0.0495728645324707, 0.04918783950805664, 0.04901580810546875, 0.04912639999389649, 0.04904550552368164, 0.049037311553955076, 0.049160190582275394, 0.049031169891357425, 0.04986777496337891, 0.04929843139648438, 0.04965580749511719, 0.04917145538330078, 0.04908544158935547, 0.04911513519287109, 0.04914483261108398, 0.04914995193481445, 0.050460670471191404, 0.04991692733764649, 0.04974387359619141, 0.04964659118652344, 0.04969267272949219, 0.049568767547607424, 0.049114112854003904, 0.04916940689086914, 0.049064960479736325, 0.04924313735961914, 0.0495810546875, 0.049620990753173826, 0.050065406799316405, 0.050141185760498044, 0.04996198272705078, 0.049895423889160156, 0.049821697235107425, 0.0493199348449707, 0.04947148895263672, 0.04919193649291992, 0.049102848052978515, 0.04903219223022461, 0.04896051025390625, 0.048968704223632815, 0.04898713684082031, 0.049121280670166016, 0.048933887481689455, 0.0489881591796875, 0.04907110214233398, 0.04901785659790039, 0.04915097427368164, 0.04949094390869141, 0.0496732177734375, 0.05024153518676758, 0.04919500732421875, 0.04903936004638672, 0.04903936004638672, 0.04908748626708984, 0.04907417678833008, 0.04916326522827148, 0.04973158264160156, 0.04947558212280274, 0.04945305633544922, 0.04947148895263672, 0.04913980865478516, 0.04911196899414062, 0.049035263061523435, 0.04898918533325195, 0.04914483261108398, 0.04902092742919922, 0.04904959869384766, 0.04910489654541016, 0.04919500732421875, 0.049522689819335934, 0.04913356781005859, 0.049240062713623044, 0.04957388687133789, 0.049160190582275394, 0.04910079956054687, 0.049716224670410154, 0.04944793701171875, 0.04904345703125, 0.049111038208007815, 0.04904857635498047, 0.049068031311035154, 0.049148929595947265, 0.04907417678833008, 0.049380352020263675, 0.04934143829345703, 0.04925439834594727, 0.049075199127197267, 0.04925132751464844, 0.04917657470703125, 0.0496732177734375, 0.04964966583251953, 0.0496814079284668, 0.049563648223876954, 0.04917248153686524, 0.04918067169189453, 0.04940697479248047, 0.04954111862182617, 0.04937932968139649, 0.04949401473999023, 0.049463294982910154, 0.04943974304199219, 0.049484798431396484, 0.0495206413269043, 0.049549312591552735, 0.049500160217285157, 0.04900556945800781, 0.04929843139648438, 0.04954316711425781, 0.049601535797119144, 0.049501182556152344, 0.04949708938598633, 0.04943360137939453, 0.04964761734008789, 0.04993535995483398, 0.04977664184570312, 0.0497336311340332, 0.05002342224121094, 0.04971724700927734, 0.04949606323242187, 0.049667072296142575, 0.04944588851928711, 0.049067008972167966, 0.04901171112060547, 0.04931071853637695, 0.04904550552368164, 0.049091583251953126, 0.04908236694335937, 0.049514495849609375, 0.04969062423706055, 0.04955340957641602, 0.04957388687133789, 0.04907212829589844, 0.049138687133789063, 0.04937625503540039, 0.049160190582275394, 0.04985343933105469, 0.04928614425659179, 0.04907929611206055, 0.04915097427368164, 0.0491069450378418, 0.049132545471191405, 0.04965785598754883, 0.049186817169189455, 0.04912332916259766, 0.04954316711425781, 0.049562625885009766, 0.04961177444458008, 0.049691646575927735, 0.04995174407958984, 0.049631233215332034, 0.04957491302490234, 0.04945305633544922, 0.04959641647338867, 0.04921452713012695, 0.049430465698242186, 0.049143806457519534, 0.0492410888671875, 0.04938547134399414, 0.049121280670166016, 0.04911718368530273, 0.04922163009643555, 0.0495022087097168, 0.049631233215332034, 0.049503231048583986, 0.04945305633544922, 0.04938137435913086, 0.04957491302490234, 0.04928102493286133, 0.04928614425659179, 0.04943769454956055, 0.049601535797119144, 0.049552383422851565, 0.049448959350585936, 0.04965785598754883, 0.04913151931762695, 0.04942950439453125, 0.04945510482788086, 0.04904550552368164, 0.04908031845092774, 0.04899532699584961, 0.049293312072753906, 0.0495206413269043, 0.04957491302490234, 0.04948787307739258, 0.04955033493041992, 0.04965990447998047, 0.04955648040771484, 0.049552383422851565, 0.04962508773803711, 0.04951244735717773, 0.04962508773803711, 0.04964761734008789, 0.049509376525878904, 0.04957593536376953, 0.04964352035522461, 0.049635326385498044, 0.04960768127441406, 0.04965478515625, 0.049634304046630856, 0.049658878326416016, 0.0491253776550293, 0.04902604675292969, 0.04933222579956055, 0.04973056030273437, 0.04955340957641602, 0.04912844848632812, 0.04945305633544922, 0.04969369506835938, 0.04967116928100586, 0.04978790283203125, 0.04937625503540039, 0.04919807815551758, 0.04960768127441406, 0.05053440093994141, 0.04968243026733398, 0.04976128005981445, 0.04964761734008789, 0.04984832000732422, 0.04976128005981445, 0.04964966583251953, 0.04964044952392578, 0.049702911376953124, 0.04978384017944336, 0.04963836669921875, 0.0504535026550293, 0.04998144149780273, 0.04957388687133789, 0.04902707290649414, 0.049007614135742186, 0.04924313735961914, 0.0494837760925293, 0.04933529663085937, 0.048895999908447264, 0.04901990509033203, 0.04905267333984375, 0.04902195358276367, 0.048996353149414064, 0.049086463928222655, 0.04921343994140625, 0.04946124649047851, 0.04954111862182617, 0.049479679107666014, 0.04900864028930664, 0.04902604675292969, 0.049075199127197267, 0.04911820983886719, 0.04901171112060547, 0.049111038208007815, 0.04904243087768555, 0.04901990509033203, 0.04905267333984375, 0.049084415435791014, 0.04894924926757813, 0.04907110214233398, 0.049122303009033204, 0.04900454330444336, 0.04902912139892578, 0.04912025451660156, 0.049517566680908204, 0.049259521484375, 0.04913151931762695, 0.04910079956054687, 0.04904755020141602, 0.0490967025756836, 0.04904550552368164, 0.04913663864135742, 0.049122303009033204, 0.04905062484741211, 0.04960255813598633, 0.04920115280151367, 0.049089534759521485, 0.04911308670043945, 0.04920012664794922, 0.049756160736083986, 0.04969472122192383, 0.04956979370117188, 0.04976332855224609, 0.049326080322265625, 0.04910079956054687, 0.049181697845458984, 0.04909056091308594, 0.04917248153686524, 0.0490885124206543, 0.04915507125854492, 0.04918374252319336, 0.0490967025756836, 0.04923392105102539, 0.049219585418701174, 0.04902195358276367, 0.048947200775146485, 0.04889497756958008, 0.049205249786376956, 0.04906291198730469, 0.04895641708374023, 0.04900966262817383, 0.04904755020141602, 0.04897382354736328, 0.04893900680541992, 0.04901068878173828, 0.04904652786254883, 0.04904755020141602, 0.04900454330444336, 0.04904755020141602, 0.049053695678710936, 0.04898713684082031, 0.04895129776000977, 0.04909056091308594, 0.049073150634765625, 0.04899327850341797, 0.04951244735717773, 0.04988415908813477, 0.04980121612548828, 0.049462272644042966, 0.049522689819335934, 0.04948787307739258, 0.04953190231323242, 0.04908544158935547, 0.04909465789794922, 0.049006591796875, 0.049393665313720705, 0.049140735626220705, 0.04915507125854492, 0.049040382385253906, 0.04913356781005859, 0.04917248153686524, 0.04906086349487305, 0.04911820983886719, 0.049084415435791014, 0.049306625366210936, 0.04907929611206055, 0.049320960998535154, 0.04975209426879883, 0.04961481475830078, 0.04960870361328125, 0.04966604614257813, 0.049486846923828126, 0.04915609741210938, 0.04955033493041992, 0.04959436798095703, 0.049734657287597656, 0.049721343994140625, 0.050108417510986325, 0.049888256072998044, 0.04969472122192383, 0.04957183837890625, 0.04928204727172852, 0.049617919921875, 0.0495882568359375, 0.04951241683959961, 0.04948582458496094, 0.04967628860473633, 0.04952576065063476, 0.0491253776550293, 0.049073150634765625, 0.048914432525634766, 0.04899737548828125, 0.04906598281860351, 0.0490618896484375, 0.04907110214233398, 0.049426433563232425, 0.049274879455566405, 0.0490618896484375, 0.049070079803466796, 0.049430526733398435, 0.049476608276367184, 0.04913663864135742, 0.0495022087097168, 0.049547264099121094, 0.04947763061523437, 0.049465343475341796, 0.04949606323242187, 0.049535999298095705, 0.04941516876220703, 0.049430526733398435, 0.04919398498535156, 0.04924518585205078, 0.049070079803466796, 0.04950630569458008, 0.04977459335327149, 0.04954009628295898, 0.04942950439453125, 0.049135616302490234, 0.04937113571166992, 0.04962815856933594, 0.04959743881225586, 0.04945612716674805, 0.049535999298095705, 0.049514495849609375, 0.049620990753173826, 0.049514495849609375, 0.049514495849609375, 0.04981043243408203, 0.04956671905517578, 0.04956671905517578, 0.04951551818847656, 0.049544193267822265, 0.049582080841064455, 0.049721343994140625, 0.04961587142944336, 0.04984320068359375, 0.04926976013183594, 0.04909465789794922, 0.04937011337280273, 0.04928409576416016, 0.04921651077270508, 0.04926668930053711, 0.04905984115600586, 0.0492564468383789, 0.04951551818847656, 0.0496097297668457, 0.049917953491210934, 0.049757183074951174, 0.049576961517333984]",tokens/s,20.258037655480646,,,,,,MB,935.317504,19279.642624,0.0,18649.972736,18648.241152,s,1,7.20255078125,7.20255078125,0.0,7.20255078125,7.20255078125,7.20255078125,7.20255078125,[7.20255078125],,kWh,2.8533447888852e-06,1.5476262946301547e-06,9.332507465986684e-06,1.3733478549502038e-05, bfloat16-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-125m,,cuda,0,42,,,,,True,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,d35829e539df8480b726c647eeabf91e41eae047,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,,MB,1145.126912,1021.837312,0.0,375.390208,302.523392,s,26,0.1730886416435242,0.006657255447827854,0.0002481283321355283,0.006683007955551147,0.00685208010673523,0.007028840184211731,0.00743286395072937,"[0.007555935859680176, 0.007063648223876953, 0.006448448181152344, 0.006697760105133056, 0.006711967945098877, 0.006708735942840576, 0.006407040119171143, 0.006714816093444824, 0.006406239986419678, 0.0066418881416320805, 0.0064271678924560545, 0.00667190408706665, 0.006628287792205811, 0.006407936096191406, 0.006410336017608643, 0.006410496234893799, 0.006703616142272949, 0.006694111824035644, 0.00661840009689331, 0.006731872081756592, 0.006708896160125733, 0.006421120166778565, 0.00644323205947876, 0.006924416065216064, 0.0067506241798400875, 0.006779744148254395]",tokens/s,38454.285254072434,kWh,7.873261593006276e-08,4.3141790385561764e-08,1.8187255290535997e-07,3.037469592209845e-07,tokens/kWh,842806791.0755702,MB,1145.126912,1021.837312,0.0,375.390208,312.411136,s,26,9.946028289794924,0.38253954960749703,0.00739916788964748,0.3835676879882812,0.3865096130371094,0.38839906311035155,0.40465960693359376,"[0.4098717956542969, 0.38902304077148436, 0.38239901733398435, 0.38649209594726563, 0.3843893127441406, 0.38381002807617187, 0.37556607055664065, 0.38332534790039063, 0.37828363037109375, 0.3843154602050781, 0.3792252197265625, 0.38153680419921876, 0.3831302185058594, 0.3774757995605469, 0.3781383361816406, 0.378805419921875, 0.3845868225097656, 0.3812719421386719, 0.3842580261230469, 0.38477362060546877, 0.38410842895507813, 0.3668628540039062, 0.3688902282714844, 0.3865271301269531, 0.38419686889648436, 0.3847647705078125]",tokens/s,164.68885390972218,kWh,4.482345601364418e-06,2.455868333930434e-06,7.953401049326249e-06,1.4891614984621101e-05,tokens/kWh,4230568.683454514,,s,1637,10.105503751754739,0.00617318494303895,0.0008769229226547649,0.006077439785003662,0.00616759672164917,0.006470035362243652,0.01276919818878173,"[0.006543360233306885, 0.0064839677810668945, 0.006542335987091064, 0.00643071985244751, 0.006493184089660644, 0.006386688232421875, 0.006296576023101807, 0.006281216144561768, 0.006240255832672119, 0.006330368041992188, 0.006258687973022461, 0.006166528224945069, 0.006118400096893311, 0.006293504238128662, 0.00628326416015625, 0.0065812478065490725, 0.00662937593460083, 0.006623231887817383, 0.006609920024871826, 0.006500351905822754, 0.006501376152038574, 0.006609920024871826, 0.006612991809844971, 0.00657919979095459, 0.006635519981384277, 0.0065710082054138185, 0.006464511871337891, 0.006609920024871826, 0.006627327919006347, 0.006546432018280029, 0.006426623821258545, 0.0063846721649169925, 0.006410208225250244, 0.006551551818847656, 0.006402048110961914, 0.006518847942352295, 0.0064951682090759275, 0.006526976108551025, 0.006636544227600098, 0.006663167953491211, 0.006688767910003662, 0.0066304001808166506, 0.006555647850036621, 0.006501376152038574, 0.006685696125030518, 0.006565887928009034, 0.006533120155334473, 0.006561791896820069, 0.006387712001800537, 0.006502399921417237, 0.006526976108551025, 0.006478847980499268, 0.0067041277885437015, 0.0064542717933654785, 0.006445055961608887, 0.006491168022155762, 0.006471648216247559, 0.006515711784362793, 0.006586368083953857, 0.006515711784362793, 0.0065372161865234375, 0.0066447358131408694, 0.014210047721862793, 0.006441984176635742, 0.0062341117858886715, 0.0064204797744750975, 0.006189055919647217, 0.006264832019805908, 0.006231040000915527, 0.006081535816192627, 0.006025216102600098, 0.005934080123901367, 0.006043647766113281, 0.006177792072296143, 0.006076416015625, 0.00582041597366333, 0.005827583789825439, 0.005851136207580566, 0.005910528182983398, 0.0062197761535644535, 0.006468607902526856, 0.006418432235717773, 0.006469632148742676, 0.006590464115142822, 0.0066375679969787596, 0.00672051191329956, 0.006307839870452881, 0.00620851182937622, 0.006188032150268555, 0.006165503978729248, 0.006168575763702393, 0.006131711959838867, 0.006137856006622314, 0.006096896171569824, 0.00616755199432373, 0.006121471881866455, 0.0061562881469726565, 0.006103040218353272, 0.00612665605545044, 0.006094783782958984, 0.006138879776000977, 0.00611737585067749, 0.0061337599754333495, 0.006135807991027832, 0.006091775894165039, 0.006148096084594727, 0.006078464031219482, 0.00613478422164917, 0.006111231803894043, 0.00613478422164917, 0.006094848155975342, 0.006124544143676758, 0.006071296215057373, 0.006127615928649902, 0.006094848155975342, 0.006127615928649902, 0.006079487800598145, 0.006130688190460205, 0.006178815841674804, 0.006095871925354004, 0.006110208034515381, 0.006096896171569824, 0.006074368000030517, 0.006082560062408447, 0.006106112003326416, 0.012788736343383789, 0.005829631805419922, 0.005848063945770264, 0.00582144021987915, 0.005856256008148194, 0.005815296173095703, 0.005815328121185303, 0.00582857608795166, 0.006107135772705078, 0.006031360149383545, 0.0060282878875732426, 0.006011903762817383, 0.0060026879310607914, 0.006016064167022705, 0.005997504234313965, 0.005994495868682862, 0.006007808208465576, 0.006089727878570556, 0.006103072166442871, 0.006096864223480224, 0.0060999679565429685, 0.006080512046813965, 0.006113279819488526, 0.006090752124786377, 0.006095871925354004, 0.006091775894165039, 0.00613478422164917, 0.006114304065704346, 0.006150144100189209, 0.006079487800598145, 0.006114304065704346, 0.006082560062408447, 0.006090752124786377, 0.006118400096893311, 0.006142975807189942, 0.00611846399307251, 0.006137792110443115, 0.006158336162567139, 0.006148096084594727, 0.006188096046447754, 0.006086624145507812, 0.006137824058532715, 0.006105088233947754, 0.006116352081298828, 0.006085631847381592, 0.006108160018920898, 0.006124544143676758, 0.006137856006622314, 0.006102015972137451, 0.006090752124786377, 0.006118400096893311, 0.006105088233947754, 0.006102015972137451, 0.0061265921592712404, 0.006107135772705078, 0.0060999679565429685, 0.0061337599754333495, 0.006085631847381592, 0.006120448112487793, 0.006072319984436035, 0.006111231803894043, 0.006077439785003662, 0.006100992202758789, 0.013244416236877441, 0.006139904022216797, 0.006074368000030517, 0.006115327835083008, 0.006090752124786377, 0.0061562881469726565, 0.006119423866271972, 0.006115327835083008, 0.006085631847381592, 0.006162432193756104, 0.00611737585067749, 0.006121471881866455, 0.00611846399307251, 0.0061101441383361816, 0.006091775894165039, 0.006177792072296143, 0.006149119853973388, 0.006070335865020752, 0.006164415836334229, 0.006132736206054688, 0.006195199966430664, 0.006171648025512695, 0.00616755199432373, 0.00613478422164917, 0.0061337599754333495, 0.006128640174865723, 0.006144000053405762, 0.006153215885162353, 0.006144000053405762, 0.006165503978729248, 0.006119423866271972, 0.006135807991027832, 0.006108160018920898, 0.006146048069000244, 0.006136832237243653, 0.006138879776000977, 0.006100992202758789, 0.006146080017089843, 0.0061255359649658205, 0.006173696041107178, 0.006178815841674804, 0.006118400096893311, 0.006106143951416016, 0.006149087905883789, 0.006147071838378906, 0.006113279819488526, 0.0061521921157836916, 0.0060928001403808595, 0.006123519897460937, 0.0060897917747497554, 0.006092735767364502, 0.006096896171569824, 0.006132736206054688, 0.006097919940948486, 0.00613478422164917, 0.006106112003326416, 0.006118400096893311, 0.006124544143676758, 0.006162432193756104, 0.00612556791305542, 0.006123519897460937, 0.006128672122955322, 0.006086624145507812, 0.01324339199066162, 0.0061265921592712404, 0.006111231803894043, 0.006082560062408447, 0.006087679862976075, 0.006082560062408447, 0.006102015972137451, 0.006109183788299561, 0.006138879776000977, 0.006098944187164307, 0.006130688190460205, 0.006104063987731933, 0.006115327835083008, 0.006105088233947754, 0.006137856006622314, 0.006095871925354004, 0.006127615928649902, 0.006079487800598145, 0.0060999679565429685, 0.006090752124786377, 0.006098944187164307, 0.00611737585067749, 0.006078464031219482, 0.006690847873687744, 0.006228960037231445, 0.006060031890869141, 0.006046720027923584, 0.006104063987731933, 0.006074368000030517, 0.006128640174865723, 0.006054944038391114, 0.006082528114318848, 0.006055935859680176, 0.006088704109191895, 0.006071296215057373, 0.006077439785003662, 0.0060999679565429685, 0.006054912090301514, 0.006074368000030517, 0.006051839828491211, 0.006055935859680176, 0.00603545618057251, 0.006069248199462891, 0.0060405759811401364, 0.0060928001403808595, 0.006091775894165039, 0.006115327835083008, 0.006032383918762207, 0.006094848155975342, 0.0060293121337890625, 0.00606822395324707, 0.006043647766113281, 0.006093823909759521, 0.006045695781707764, 0.0061224961280822755, 0.006041632175445556, 0.006059999942779541, 0.006080512046813965, 0.006095871925354004, 0.006022143840789795, 0.006087679862976075, 0.006022143840789795, 0.006072383880615234, 0.013231040000915528, 0.006089727878570556, 0.006046751976013184, 0.006080480098724366, 0.0060631041526794435, 0.006120480060577392, 0.006107103824615478, 0.00608358383178711, 0.006062079906463623, 0.00612556791305542, 0.00606822395324707, 0.006164480209350586, 0.006095871925354004, 0.006050848007202149, 0.006076384067535401, 0.006061056137084961, 0.0061675839424133305, 0.006051807880401611, 0.006103040218353272, 0.006104063987731933, 0.006112256050109863, 0.006080512046813965, 0.006078464031219482, 0.006073344230651856, 0.006090752124786377, 0.006049791812896729, 0.0060999679565429685, 0.006070335865020752, 0.006127552032470703, 0.006065152168273926, 0.00608358383178711, 0.006105088233947754, 0.006120448112487793, 0.0060928001403808595, 0.0060928001403808595, 0.006088704109191895, 0.006067200183868408, 0.006045695781707764, 0.006116352081298828, 0.006077439785003662, 0.006094848155975342, 0.00607539176940918, 0.0060661759376525876, 0.006081535816192627, 0.006065152168273926, 0.006110208034515381, 0.006103040218353272, 0.00606822395324707, 0.006074368000030517, 0.006087679862976075, 0.006032383918762207, 0.006100992202758789, 0.006096896171569824, 0.00611843204498291, 0.006040544033050537, 0.006094848155975342, 0.0060293121337890625, 0.006080512046813965, 0.00616755199432373, 0.00608358383178711, 0.006072351932525635, 0.006068192005157471, 0.006044672012329101, 0.012703743934631348, 0.005827583789825439, 0.005839871883392334, 0.005801983833312988, 0.005803008079528809, 0.00582041597366333, 0.0058091521263122555, 0.005816319942474365, 0.005805056095123291, 0.005829631805419922, 0.005791744232177734, 0.005796864032745362, 0.005791744232177734, 0.005914624214172363, 0.005793791770935058, 0.005825535774230957, 0.005818367958068848, 0.005795839786529541, 0.005810175895690918, 0.005803008079528809, 0.005829631805419922, 0.005810175895690918, 0.005814271926879883, 0.005822463989257813, 0.005811200141906738, 0.005803008079528809, 0.005785600185394287, 0.005814271926879883, 0.005804031848907471, 0.005844992160797119, 0.00592793607711792, 0.005993472099304199, 0.00582041597366333, 0.00620851182937622, 0.006699007987976074, 0.006155263900756836, 0.006151167869567871, 0.006080512046813965, 0.006184959888458252, 0.006048768043518066, 0.006065152168273926, 0.00602623987197876, 0.006090752124786377, 0.006052864074707031, 0.006088704109191895, 0.0060364799499511715, 0.006094848155975342, 0.006020095825195313, 0.006096896171569824, 0.0060293121337890625, 0.0060590081214904785, 0.006030335903167725, 0.006115327835083008, 0.0060405759811401364, 0.006047743797302246, 0.006136832237243653, 0.006067200183868408, 0.006048768043518066, 0.006060031890869141, 0.006047743797302246, 0.006080543994903565, 0.006019040107727051, 0.006044672012329101, 0.013203455924987792, 0.006097919940948486, 0.006043647766113281, 0.00606822395324707, 0.006038591861724853, 0.006099904060363769, 0.006072319984436035, 0.006145023822784424, 0.006089727878570556, 0.006055935859680176, 0.006046720027923584, 0.005912576198577881, 0.0059658241271972655, 0.00591974401473999, 0.005916672229766846, 0.005922815799713135, 0.005929984092712402, 0.0059064321517944334, 0.005886975765228271, 0.005920767784118652, 0.005948416233062744, 0.006031360149383545, 0.006119423866271972, 0.006138879776000977, 0.006094848155975342, 0.0060364799499511715, 0.006094848155975342, 0.007301119804382325, 0.006285376071929931, 0.006147039890289307, 0.006143968105316162, 0.0061296639442443845, 0.006113344192504882, 0.006086592197418213, 0.006069248199462891, 0.006033408164978027, 0.006041600227355957, 0.006061056137084961, 0.006074368000030517, 0.006080512046813965, 0.006150144100189209, 0.0060999679565429685, 0.006030335903167725, 0.00608460807800293, 0.006048768043518066, 0.006094848155975342, 0.0060702719688415525, 0.006105088233947754, 0.0060702719688415525, 0.006109183788299561, 0.006057983875274659, 0.006077439785003662, 0.0060702719688415525, 0.006089727878570556, 0.00605292797088623, 0.006078400135040283, 0.006061056137084961, 0.006148096084594727, 0.00608460807800293, 0.006061056137084961, 0.0060282878875732426, 0.00606822395324707, 0.006049791812896729, 0.012676095962524414, 0.005815296173095703, 0.005815296173095703, 0.005789696216583252, 0.005853184223175049, 0.0058122239112854005, 0.0058091521263122555, 0.005808127880096436, 0.005795904159545898, 0.00582144021987915, 0.005789631843566894, 0.005798912048339844, 0.005786687850952149, 0.005817344188690185, 0.005804992198944091, 0.005805056095123291, 0.005789696216583252, 0.005776383876800537, 0.005804031848907471, 0.006078464031219482, 0.0061265921592712404, 0.006054912090301514, 0.006061056137084961, 0.006057983875274659, 0.006053887844085694, 0.006047743797302246, 0.006074368000030517, 0.006053887844085694, 0.006073344230651856, 0.006043647766113281, 0.0060375680923461916, 0.006089663982391357, 0.006047743797302246, 0.0060405759811401364, 0.0061296639442443845, 0.006097919940948486, 0.006060031890869141, 0.006104063987731933, 0.006049791812896729, 0.0060590081214904785, 0.006024191856384278, 0.006061056137084961, 0.006042623996734619, 0.006081535816192627, 0.006071296215057373, 0.0064849920272827145, 0.006145023822784424, 0.006056960105895996, 0.006174719810485839, 0.006031392097473144, 0.006093791961669922, 0.006041600227355957, 0.006078464031219482, 0.006080575942993164, 0.006068160057067871, 0.006038591861724853, 0.006074304103851318, 0.006118400096893311, 0.006090752124786377, 0.00601087999343872, 0.006157311916351318, 0.006021120071411133, 0.006071296215057373, 0.013154303550720215, 0.006054912090301514, 0.006031360149383545, 0.006079487800598145, 0.006792191982269287, 0.006147071838378906, 0.006053887844085694, 0.006114304065704346, 0.006074368000030517, 0.006094848155975342, 0.0060590081214904785, 0.006072383880615234, 0.006068160057067871, 0.006043647766113281, 0.006243328094482422, 0.006071296215057373, 0.0060631041526794435, 0.00603545618057251, 0.006094848155975342, 0.006081535816192627, 0.00612556791305542, 0.006073344230651856, 0.00607539176940918, 0.006056960105895996, 0.006079487800598145, 0.006046720027923584, 0.006098944187164307, 0.006093823909759521, 0.006112287998199463, 0.006093791961669922, 0.006105088233947754, 0.0061634559631347655, 0.006114304065704346, 0.006113279819488526, 0.006098944187164307, 0.006069248199462891, 0.006052896022796631, 0.006065120220184326, 0.006176767826080322, 0.006088704109191895, 0.00611737585067749, 0.006090752124786377, 0.0060702719688415525, 0.006098976135253906, 0.0060210881233215335, 0.00607539176940918, 0.006062079906463623, 0.006081535816192627, 0.006030335903167725, 0.006094848155975342, 0.0060293121337890625, 0.006100992202758789, 0.006067200183868408, 0.00608358383178711, 0.006020095825195313, 0.006098944187164307, 0.006032383918762207, 0.006057983875274659, 0.0060631041526794435, 0.006107135772705078, 0.006056992053985596, 0.006037471771240234, 0.006154240131378174, 0.012718079566955566, 0.005825535774230957, 0.005831679821014404, 0.005849088191986084, 0.005825535774230957, 0.005788671970367432, 0.0057866559028625485, 0.005824480056762695, 0.0057825279235839844, 0.005855231761932373, 0.005793791770935058, 0.005807104110717773, 0.005797887802124023, 0.0058122239112854005, 0.005780479907989502, 0.005814271926879883, 0.006235136032104492, 0.006039552211761475, 0.006051839828491211, 0.006017024040222168, 0.0060375361442565915, 0.006048799991607666, 0.00618284797668457, 0.006076416015625, 0.00603545618057251, 0.0060928001403808595, 0.0060375680923461916, 0.006115263938903809, 0.006047743797302246, 0.006091775894165039, 0.006024191856384278, 0.006080512046813965, 0.006034431934356689, 0.006082560062408447, 0.006039552211761475, 0.006078464031219482, 0.006085631847381592, 0.006093823909759521, 0.0060590081214904785, 0.006069248199462891, 0.006042623996734619, 0.006090752124786377, 0.006078464031219482, 0.006064127922058105, 0.006077439785003662, 0.006181920051574707, 0.006172639846801758, 0.006057983875274659, 0.00608460807800293, 0.0060364799499511715, 0.006100992202758789, 0.006064127922058105, 0.0060928001403808595, 0.006041632175445556, 0.0060845761299133305, 0.0060364799499511715, 0.006308864116668702, 0.006102015972137451, 0.006111231803894043, 0.0060405759811401364, 0.006078464031219482, 0.006039552211761475, 0.006051839828491211, 0.01323519992828369, 0.006103040218353272, 0.006050816059112549, 0.006104063987731933, 0.006069248199462891, 0.006172671794891358, 0.006091775894165039, 0.006085631847381592, 0.00606822395324707, 0.006090752124786377, 0.006113279819488526, 0.0060702719688415525, 0.006107200145721436, 0.0061132159233093265, 0.006048768043518066, 0.006091839790344238, 0.006078400135040283, 0.00613478422164917, 0.006105088233947754, 0.006047743797302246, 0.006090752124786377, 0.006074368000030517, 0.006071296215057373, 0.005817344188690185, 0.005819392204284668, 0.005783616065979004, 0.005801919937133789, 0.005819392204284668, 0.005790719985961914, 0.006072319984436035, 0.006076416015625, 0.006069248199462891, 0.006069248199462891, 0.006069248199462891, 0.0060631041526794435, 0.006078464031219482, 0.006054912090301514, 0.006039552211761475, 0.006064127922058105, 0.0060661759376525876, 0.0060928001403808595, 0.006050816059112549, 0.0060661759376525876, 0.006128640174865723, 0.006065152168273926, 0.00606822395324707, 0.006090752124786377, 0.006022143840789795, 0.0060661759376525876, 0.006015999794006348, 0.006054912090301514, 0.006043647766113281, 0.00608358383178711, 0.0061562881469726565, 0.006078464031219482, 0.006039552211761475, 0.006074399948120117, 0.006030303955078125, 0.00608460807800293, 0.006045695781707764, 0.006033408164978027, 0.0060405759811401364, 0.006039552211761475, 0.013119487762451172, 0.006107135772705078, 0.006054912090301514, 0.006065152168273926, 0.006031360149383545, 0.006074368000030517, 0.006042623996734619, 0.006097919940948486, 0.006090752124786377, 0.006082560062408447, 0.00603545618057251, 0.006299647808074951, 0.006087679862976075, 0.006136832237243653, 0.006074368000030517, 0.006054912090301514, 0.0060702719688415525, 0.00603545618057251, 0.006060031890869141, 0.006037504196166992, 0.006079487800598145, 0.0060375680923461916, 0.006063039779663086, 0.006042623996734619, 0.006074368000030517, 0.006067200183868408, 0.006064127922058105, 0.006031360149383545, 0.00608460807800293, 0.006039552211761475, 0.006077439785003662, 0.006031360149383545, 0.006046720027923584, 0.00602623987197876, 0.006056960105895996, 0.006038527965545654, 0.006149119853973388, 0.006073344230651856, 0.00618393611907959, 0.006106112003326416, 0.006105088233947754, 0.006052864074707031, 0.006052896022796631, 0.006022111892700195, 0.006064127922058105, 0.00606931209564209, 0.006098879814147949, 0.006080512046813965, 0.006091775894165039, 0.006031424045562744, 0.006006720066070556, 0.0060702719688415525, 0.0062024321556091305, 0.006081471920013428, 0.006121471881866455, 0.006041600227355957, 0.006054912090301514, 0.0060702719688415525, 0.006043647766113281, 0.006138879776000977, 0.006140927791595459, 0.006051839828491211, 0.006033408164978027, 0.01265561580657959, 0.0058122239112854005, 0.005810175895690918, 0.005774335861206055, 0.005772287845611572, 0.005836800098419189, 0.005783552169799804, 0.005807104110717773, 0.005776383876800537, 0.005827583789825439, 0.00580511999130249, 0.0057701759338378905, 0.0058122239112854005, 0.005774335861206055, 0.005827583789825439, 0.005783552169799804, 0.005811200141906738, 0.005751808166503906, 0.005805056095123291, 0.005788671970367432, 0.005797887802124023, 0.006077439785003662, 0.006094848155975342, 0.006082560062408447, 0.006104063987731933, 0.006054912090301514, 0.0061265921592712404, 0.006049791812896729, 0.006087679862976075, 0.006042623996734619, 0.006378496170043945, 0.006094848155975342, 0.006113279819488526, 0.006039584159851074, 0.006100959777832031, 0.006025216102600098, 0.006077439785003662, 0.006065152168273926, 0.006079487800598145, 0.006081535816192627, 0.006105088233947754, 0.00611737585067749, 0.006057983875274659, 0.006087679862976075, 0.006072319984436035, 0.006103040218353272, 0.006053887844085694, 0.006085631847381592, 0.006038527965545654, 0.00607539176940918, 0.005996543884277344, 0.0060702719688415525, 0.006031360149383545, 0.006115327835083008, 0.006008831977844238, 0.006159359931945801, 0.006021120071411133, 0.00608358383178711, 0.00603545618057251, 0.00606822395324707, 0.006047743797302246, 0.006091775894165039, 0.0060282878875732426, 0.012697600364685058, 0.005836800098419189, 0.005846015930175781, 0.005794816017150879, 0.005826560020446778, 0.005836800098419189, 0.0057825279235839844, 0.005832704067230224, 0.005793791770935058, 0.005836800098419189, 0.005796864032745362, 0.005796927928924561, 0.00579475212097168, 0.005817344188690185, 0.005943295955657959, 0.0062228479385375976, 0.006078464031219482, 0.006017024040222168, 0.0060405759811401364, 0.006030335903167725, 0.00602623987197876, 0.006057983875274659, 0.00603545618057251, 0.006078464031219482, 0.006047743797302246, 0.006078464031219482, 0.005991424083709717, 0.006073344230651856, 0.006054912090301514, 0.006114304065704346, 0.006017024040222168, 0.0060702719688415525, 0.006005760192871094, 0.006047743797302246, 0.006049791812896729, 0.006077439785003662, 0.0060282878875732426, 0.006076416015625, 0.006017024040222168, 0.006079487800598145, 0.006023168087005615, 0.006081535816192627, 0.006017024040222168, 0.006060031890869141, 0.006013951778411865, 0.006060031890869141, 0.006025216102600098, 0.006076416015625, 0.006023168087005615, 0.006062143802642822, 0.005989312171936035, 0.006064127922058105, 0.006067200183868408, 0.006080543994903565, 0.006030303955078125, 0.006061056137084961, 0.006033408164978027, 0.006045695781707764, 0.006041600227355957, 0.0060282878875732426, 0.006015999794006348, 0.006025216102600098, 0.006008831977844238, 0.01265561580657959, 0.005801983833312988, 0.005784639835357666, 0.005783487796783447, 0.005814271926879883, 0.005797952175140381, 0.0057927041053771975, 0.005807104110717773, 0.005783552169799804, 0.005816319942474365, 0.005793791770935058, 0.005810175895690918, 0.005840896129608154, 0.006039552211761475, 0.005947391986846923, 0.005924863815307617, 0.005898240089416504, 0.005874720096588135, 0.005919712066650391, 0.00591974401473999, 0.006913023948669434, 0.0062197761535644535, 0.0060282878875732426, 0.006119423866271972, 0.00602726411819458, 0.006096896171569824, 0.006044672012329101, 0.006127615928649902, 0.0060661759376525876, 0.006085663795471191, 0.006049759864807129, 0.0060405759811401364, 0.006065152168273926, 0.006067200183868408, 0.00607539176940918, 0.006050816059112549, 0.006017024040222168, 0.006060031890869141, 0.006034431934356689, 0.00606928014755249, 0.006172639846801758, 0.006048768043518066, 0.006072319984436035, 0.0060405759811401364, 0.0060631041526794435, 0.006022143840789795, 0.006088704109191895, 0.006034431934356689, 0.006051839828491211, 0.006033408164978027, 0.006033408164978027, 0.00601907205581665, 0.006044672012329101, 0.006043647766113281, 0.0060405759811401364, 0.006041600227355957, 0.006044672012329101, 0.0060364799499511715, 0.0060661759376525876, 0.006022143840789795, 0.006065152168273926, 0.0060364799499511715, 0.006072319984436035, 0.013180928230285644, 0.00608358383178711, 0.006030335903167725, 0.006052864074707031, 0.006343679904937744, 0.006112256050109863, 0.006033408164978027, 0.0060897917747497554, 0.006024127960205078, 0.006062079906463623, 0.006025216102600098, 0.0060282878875732426, 0.006017024040222168, 0.006049791812896729, 0.0060282878875732426, 0.006055935859680176, 0.006041600227355957, 0.0060590081214904785, 0.0060364799499511715, 0.006039552211761475, 0.006147071838378906, 0.006065152168273926, 0.006085631847381592, 0.006072319984436035, 0.0060702719688415525, 0.006061056137084961, 0.006062079906463623, 0.0064737281799316405, 0.0062156801223754886, 0.006107135772705078, 0.006519807815551758, 0.006919167995452881, 0.00612556791305542, 0.006120448112487793, 0.006054912090301514, 0.006120448112487793, 0.006107135772705078, 0.006110208034515381, 0.00608460807800293, 0.006158336162567139, 0.006033408164978027, 0.006078464031219482, 0.006039552211761475, 0.0060590081214904785, 0.006100992202758789, 0.006089727878570556, 0.006081535816192627, 0.0060364799499511715, 0.006046720027923584, 0.006043647766113281, 0.006008831977844238, 0.006047743797302246, 0.00607539176940918, 0.006087679862976075, 0.006061056137084961, 0.006039552211761475, 0.006047743797302246, 0.006044672012329101, 0.006071296215057373, 0.006016032218933105, 0.006052832126617432, 0.006012928009033203, 0.0060702719688415525, 0.01316864013671875, 0.0060661759376525876, 0.006042623996734619, 0.006047743797302246, 0.006043647766113281, 0.006055935859680176, 0.0060631041526794435, 0.006050816059112549, 0.006056960105895996, 0.006072319984436035, 0.0060661759376525876, 0.006060031890869141, 0.006046720027923584, 0.0060631041526794435, 0.0060405759811401364, 0.006051839828491211, 0.006038527965545654, 0.006030335903167725, 0.006009856224060059, 0.006042623996734619, 0.006037504196166992, 0.006072319984436035, 0.006033408164978027, 0.006098944187164307, 0.006037504196166992, 0.006073344230651856, 0.0060293121337890625, 0.006100992202758789, 0.006044672012329101, 0.006085631847381592, 0.006045695781707764, 0.00608460807800293, 0.006051839828491211, 0.006077439785003662, 0.006051839828491211, 0.00603545618057251, 0.00602623987197876, 0.0060405759811401364, 0.006046720027923584, 0.006012928009033203, 0.006103040218353272, 0.00601087999343872, 0.006000639915466309, 0.006033408164978027, 0.00601907205581665, 0.00602623987197876, 0.006025216102600098, 0.006032383918762207, 0.005987360000610352, 0.006031360149383545, 0.0059944639205932616, 0.006037504196166992, 0.006062079906463623, 0.006041600227355957, 0.006081535816192627, 0.006042623996734619, 0.00603545618057251, 0.0060067839622497555, 0.006043647766113281, 0.0060405759811401364, 0.006039552211761475, 0.0060364799499511715, 0.006053952217102051, 0.013111231803894043, 0.006137856006622314, 0.006031360149383545, 0.006039552211761475, 0.006034431934356689, 0.006031360149383545, 0.006030335903167725, 0.006015999794006348, 0.006034431934356689, 0.006065152168273926, 0.00601907205581665, 0.006069248199462891, 0.0060282878875732426, 0.006072319984436035, 0.006037504196166992, 0.006169600009918213, 0.006023168087005615, 0.006054912090301514, 0.005781568050384521, 0.005993408203125, 0.006142975807189942, 0.006190080165863037, 0.006102015972137451, 0.006131711959838867, 0.0060928001403808595, 0.006120448112487793, 0.006088704109191895, 0.006145023822784424, 0.006094848155975342, 0.006176767826080322, 0.006131711959838867, 0.006128640174865723, 0.006085631847381592, 0.006097919940948486, 0.0060928001403808595, 0.006112256050109863, 0.006144000053405762, 0.006106112003326416, 0.00613478422164917, 0.006091775894165039, 0.0062904319763183595, 0.006228991985321045, 0.006150144100189209, 0.006057983875274659, 0.0061224961280822755, 0.006100992202758789, 0.006123519897460937, 0.006093823909759521, 0.006074368000030517, 0.006090752124786377, 0.0060590081214904785, 0.006154240131378174, 0.006107168197631836, 0.006144032001495361, 0.006068192005157471, 0.0060927681922912594, 0.00607539176940918, 0.006145023822784424, 0.0061082239151000976, 0.0061244797706604005, 0.0060928001403808595, 0.006110208034515381, 0.00608460807800293, 0.013316096305847168, 0.006100992202758789, 0.006105088233947754, 0.006048768043518066, 0.006106112003326416, 0.006079487800598145, 0.00613478422164917, 0.00605292797088623, 0.00613369607925415, 0.006108160018920898, 0.006113344192504882, 0.006335423946380615, 0.0061010560989379885, 0.0060835199356079105, 0.006090752124786377, 0.006105088233947754, 0.006067200183868408, 0.006085631847381592, 0.006071296215057373, 0.006113279819488526, 0.006087679862976075, 0.006137856006622314, 0.006040639877319336, 0.006119359970092773, 0.006239232063293457, 0.00612556791305542, 0.006140927791595459, 0.006112256050109863, 0.006085631847381592, 0.006124544143676758, 0.006087679862976075, 0.006086656093597412, 0.00611737585067749, 0.006093823909759521, 0.006127615928649902, 0.006069248199462891, 0.006080512046813965, 0.006074368000030517, 0.006127615928649902, 0.006085631847381592, 0.006098944187164307, 0.006079487800598145, 0.006088704109191895, 0.006088704109191895, 0.006107135772705078, 0.006081535816192627, 0.006116352081298828, 0.006103040218353272, 0.006114304065704346, 0.006086656093597412, 0.006055935859680176, 0.006091775894165039, 0.006079487800598145, 0.006096896171569824, 0.006052864074707031, 0.00611737585067749, 0.006061056137084961, 0.006118400096893311, 0.006077439785003662, 0.006096896171569824, 0.006048768043518066, 0.006112256050109863, 0.006071296215057373, 0.01325772762298584, 0.006104063987731933, 0.006103040218353272, 0.006065152168273926, 0.006103040218353272, 0.0060631041526794435, 0.006080512046813965, 0.006060031890869141, 0.0061224961280822755, 0.006079487800598145, 0.006113279819488526, 0.006067200183868408, 0.006123519897460937, 0.006077439785003662, 0.006118400096893311, 0.006089727878570556, 0.0060928001403808595, 0.006078464031219482, 0.006062079906463623, 0.00607539176940918, 0.006083615779876709, 0.006100959777832031, 0.006072319984436035, 0.006107135772705078, 0.006087679862976075, 0.006114304065704346, 0.006076416015625, 0.006111231803894043, 0.006089727878570556, 0.006120448112487793, 0.006123519897460937, 0.006116352081298828, 0.006077439785003662, 0.006095871925354004, 0.006049791812896729, 0.006107135772705078, 0.006087679862976075, 0.006088704109191895, 0.006076416015625, 0.006082560062408447, 0.006081535816192627, 0.006082560062408447, 0.006045695781707764, 0.0061634559631347655, 0.006095871925354004, 0.006077439785003662, 0.006130688190460205, 0.006069248199462891, 0.0060928001403808595, 0.006060031890869141, 0.00608460807800293, 0.006064127922058105, 0.006123519897460937, 0.006097919940948486, 0.006132736206054688, 0.006071296215057373, 0.006095871925354004, 0.006061056137084961, 0.006102015972137451, 0.006091775894165039, 0.006090752124786377, 0.0060661759376525876, 0.006064127922058105, 0.012652544021606446, 0.005834752082824707, 0.005786623954772949, 0.005798912048339844, 0.005798975944519043, 0.0057998719215393065, 0.005859392166137695, 0.005782464027404785, 0.00582144021987915, 0.005798912048339844, 0.005844992160797119, 0.005785600185394287, 0.005850111961364746, 0.005904384136199951, 0.0058009600639343266, 0.005794816017150879, 0.005815296173095703, 0.005772287845611572, 0.005781504154205322, 0.005826560020446778, 0.005793791770935058, 0.005833727836608887, 0.005774335861206055, 0.005811200141906738, 0.0057794561386108395, 0.005801983833312988, 0.005804031848907471, 0.005891071796417236, 0.005839871883392334, 0.0058091521263122555, 0.005816319942474365, 0.005803008079528809, 0.005863423824310303, 0.005814271926879883, 0.005832704067230224, 0.005829631805419922, 0.005803008079528809, 0.00581324815750122, 0.0058122239112854005, 0.005857279777526855, 0.005815296173095703, 0.005856256008148194, 0.005819392204284668, 0.005832704067230224, 0.005792768001556397, 0.005818367958068848, 0.005830656051635743, 0.005847040176391601, 0.005796927928924561, 0.005823423862457275, 0.005858304023742676, 0.005785600185394287, 0.005836800098419189, 0.005801023960113526, 0.005804992198944091, 0.005811200141906738, 0.005817344188690185, 0.005819392204284668, 0.005824512004852295, 0.005799935817718506, 0.0057712640762329105, 0.005815296173095703, 0.005808127880096436, 0.012734463691711426, 0.005828639984130859, 0.005795807838439942, 0.005890048027038574, 0.005792768001556397, 0.005826560020446778, 0.005792768001556397, 0.005824512004852295, 0.005795839786529541, 0.005815296173095703, 0.005798912048339844, 0.005801983833312988, 0.005806079864501953, 0.0058132801055908205, 0.005851103782653809, 0.005795872211456299, 0.005831647872924805, 0.005761023998260498, 0.00582144021987915, 0.005818399906158447, 0.005805024147033691, 0.005893152236938476, 0.005790688037872314, 0.005890048027038574, 0.005780479907989502, 0.005819392204284668, 0.0057784318923950195, 0.005826591968536377, 0.0057999038696289066, 0.005839871883392334, 0.005774335861206055, 0.005856256008148194, 0.0057794561386108395, 0.005817344188690185, 0.005788671970367432, 0.005785600185394287, 0.005810175895690918, 0.005788671970367432, 0.005808127880096436, 0.0058009600639343266, 0.0058122239112854005, 0.005791744232177734, 0.005785600185394287, 0.005789696216583252, 0.005787648200988769, 0.0058245759010314944, 0.00579475212097168, 0.005823488235473633, 0.005788671970367432, 0.005829631805419922, 0.005772287845611572, 0.005796864032745362, 0.005791744232177734, 0.005804031848907471, 0.005794816017150879, 0.005780479907989502, 0.005822463989257813, 0.005790719985961914, 0.0058122239112854005, 0.005783552169799804, 0.00606822395324707, 0.006998015880584717, 0.0068915200233459475, 0.01359769630432129, 0.00616761589050293, 0.006172607898712159, 0.00613478422164917, 0.006124544143676758, 0.006089727878570556, 0.00613478422164917, 0.006058015823364258, 0.0061214399337768555, 0.006090752124786377, 0.006142975807189942, 0.006069248199462891, 0.006212607860565185, 0.0060928001403808595, 0.006102015972137451, 0.006079487800598145, 0.006085631847381592, 0.006039552211761475, 0.006093823909759521, 0.006102015972137451, 0.006078464031219482, 0.006124544143676758, 0.006065152168273926, 0.006119423866271972, 0.006058047771453858, 0.006090688228607178, 0.006088704109191895, 0.0061931519508361815, 0.006148096084594727, 0.006141952037811279, 0.006079487800598145, 0.006139904022216797, 0.006116352081298828, 0.006094848155975342, 0.006112256050109863, 0.006071296215057373, 0.0060631041526794435, 0.006065152168273926, 0.006112256050109863, 0.006094848155975342, 0.006148096084594727, 0.006116352081298828, 0.00612556791305542, 0.0060999999046325686, 0.006146016120910644, 0.006421504020690918, 0.0062873601913452145, 0.006189055919647217, 0.006200319766998291, 0.006172671794891358, 0.006218751907348633, 0.006254591941833496, 0.006098944187164307, 0.006206463813781738, 0.006081535816192627, 0.006151167869567871, 0.006067200183868408, 0.006138879776000977, 0.006104063987731933, 0.006155327796936035, 0.006107071876525879, 0.006173696041107178, 0.006070335865020752, 0.013321151733398437, 0.006103040218353272, 0.0061224961280822755, 0.006087679862976075, 0.006094848155975342, 0.006080543994903565, 0.0060692157745361325, 0.0060661759376525876, 0.006102015972137451, 0.006121471881866455, 0.006089727878570556, 0.006103040218353272, 0.006078464031219482, 0.0061562881469726565, 0.00606822395324707, 0.006100992202758789, 0.00606822395324707, 0.006115327835083008, 0.006053887844085694, 0.006105088233947754, 0.006054912090301514, 0.006081535816192627, 0.006065152168273926, 0.0061112961769104, 0.006059967994689942, 0.006091775894165039, 0.006064127922058105, 0.006089727878570556, 0.00612556791305542, 0.006097919940948486, 0.006061056137084961, 0.0060702719688415525, 0.00607539176940918, 0.006104063987731933, 0.006076416015625, 0.006076416015625, 0.0061562881469726565, 0.006087679862976075, 0.006089727878570556, 0.006054912090301514, 0.006100992202758789, 0.006073344230651856, 0.0060928001403808595, 0.006065152168273926, 0.006110208034515381, 0.006055935859680176, 0.006153215885162353, 0.0060631041526794435, 0.006112256050109863, 0.006088704109191895, 0.006051839828491211, 0.006091775894165039, 0.0060999679565429685, 0.006095871925354004, 0.006093823909759521, 0.00611737585067749, 0.006077439785003662, 0.006121471881866455, 0.006098944187164307, 0.006098944187164307, 0.006065152168273926, 0.006127615928649902, 0.006136832237243653, 0.013393919944763183, 0.006111231803894043, 0.006118400096893311, 0.006079487800598145, 0.006132736206054688, 0.006081535816192627, 0.0061224961280822755, 0.006077439785003662, 0.006108191967010498, 0.006085599899291992, 0.006178815841674804, 0.0060999679565429685, 0.006105088233947754, 0.006107135772705078, 0.006093823909759521, 0.006102015972137451, 0.006089727878570556, 0.006073344230651856, 0.00607539176940918, 0.006103040218353272, 0.006088768005371094, 0.0061511039733886715, 0.006074368000030517, 0.006110208034515381, 0.006069248199462891, 0.006104063987731933, 0.006093823909759521, 0.006131711959838867, 0.0060928001403808595, 0.006135871887207031, 0.006064064025878906, 0.006137856006622314, 0.006127615928649902, 0.006103040218353272, 0.006080512046813965, 0.006087744235992431, 0.006099904060363769, 0.0060928001403808595, 0.006155263900756836, 0.006090752124786377, 0.006154240131378174, 0.006086656093597412, 0.006090752124786377, 0.006081535816192627, 0.0060999679565429685, 0.00606822395324707, 0.00612556791305542, 0.006148096084594727, 0.006140927791595459, 0.0060631041526794435, 0.006073344230651856, 0.0060999679565429685, 0.006157311916351318, 0.006097983837127685, 0.006046656131744385, 0.006104063987731933, 0.00606822395324707, 0.006088704109191895, 0.006078464031219482, 0.006097919940948486, 0.006041600227355957, 0.006108160018920898, 0.00603545618057251]",tokens/s,161.9909348621779,,,main,False,False,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,databricks/dbrx-base,databricks/dbrx-base,cuda,0,42,,,True,,,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",dbrx,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-350m,,cuda,0,42,,,,,True,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,d35829e539df8480b726c647eeabf91e41eae047,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,,MB,1155.096576,1455.947776,0.0,809.500672,723.533824,s,15,0.173844352722168,0.011589623514811197,0.0003799565019044181,0.01169200038909912,0.011827808189392091,0.012035174560546875,0.012418913116455078,"[0.012514847755432129, 0.011829600334167481, 0.011780799865722657, 0.011762816429138184, 0.011121055603027345, 0.011691712379455566, 0.011717568397521973, 0.011177023887634278, 0.011095871925354004, 0.011142815589904785, 0.011825119972229004, 0.01169200038909912, 0.011690336227416992, 0.011700032234191894, 0.011102751731872558]",tokens/s,22088.72442429554,kWh,1.3687280103232357e-07,7.499941463656173e-08,3.880742128667657e-07,5.99946428535651e-07,tokens/kWh,426704765.3318726,MB,1155.096576,1455.947776,0.0,809.500672,749.216256,s,15,10.518155212402345,0.7012103474934896,0.010939446017162301,0.7029287719726562,0.7118413452148438,0.7139651794433594,0.7139718444824219,"[0.7139735107421875, 0.7086609497070312, 0.7139616088867188, 0.7029287719726562, 0.7076468505859375, 0.7025215454101562, 0.70141259765625, 0.6947970581054688, 0.6885543212890625, 0.6679688720703125, 0.7027149047851563, 0.7057315063476562, 0.7060153198242187, 0.7033875732421875, 0.6978798217773438]",tokens/s,89.84465250006157,kWh,8.281585413480256e-06,4.537992137930547e-06,1.557408670500015e-05,2.839366425641095e-05,tokens/kWh,2218804.85135959,,s,944,10.667562987327564,0.011300384520474126,0.0014620249435332547,0.011165696144104004,0.011293696403503419,0.01140469717979431,0.022238607692718505,"[0.011946016311645508, 0.011907039642333984, 0.012149760246276856, 0.010962944030761718, 0.01061888027191162, 0.010612735748291016, 0.010579968452453613, 0.010592255592346191, 0.010800127983093261, 0.011176959991455078, 0.011200511932373047, 0.011199487686157226, 0.011238431930541993, 0.011479007720947265, 0.01143398380279541, 0.011431936264038087, 0.011265024185180664, 0.011222016334533692, 0.01135103988647461, 0.01123737621307373, 0.011149312019348144, 0.011206656455993653, 0.01123635196685791, 0.01124454402923584, 0.011256832122802735, 0.011200511932373047, 0.011184127807617187, 0.011299936294555665, 0.011191200256347657, 0.011272192001342773, 0.011169792175292969, 0.011192319869995117, 0.011144191741943359, 0.011204607963562012, 0.01117903995513916, 0.011198431968688964, 0.011172863960266113, 0.01124556827545166, 0.011199487686157226, 0.011174912452697755, 0.011217920303344727, 0.011229184150695801, 0.011172863960266113, 0.01120358371734619, 0.011223039627075194, 0.011226112365722657, 0.011174912452697755, 0.011188223838806152, 0.011195391654968261, 0.011210751533508301, 0.011783167839050293, 0.013306879997253418, 0.013592576026916504, 0.012147711753845216, 0.011752448081970214, 0.011341823577880859, 0.011339776039123535, 0.011367424011230469, 0.011053055763244628, 0.010977279663085938, 0.011018239974975585, 0.011346943855285644, 0.02332876777648926, 0.01091481590270996, 0.010882047653198243, 0.011111455917358399, 0.011453408241271973, 0.011422719955444336, 0.011257856369018555, 0.011213824272155762, 0.011287551879882812, 0.01124665641784668, 0.011330495834350586, 0.011282431602478027, 0.011202560424804688, 0.011185152053833008, 0.011208703994750976, 0.011286527633666991, 0.01124454402923584, 0.0112424955368042, 0.011131903648376466, 0.011344896316528321, 0.011210751533508301, 0.0112424955368042, 0.011263999938964844, 0.011302911758422851, 0.011213824272155762, 0.011325440406799316, 0.011231231689453124, 0.011256832122802735, 0.01124556827545166, 0.011246591567993165, 0.011188223838806152, 0.011262975692749023, 0.011303936004638672, 0.011294719696044921, 0.01122208023071289, 0.011206591606140136, 0.0112424955368042, 0.011308032035827637, 0.011212800025939941, 0.011331583976745606, 0.011241472244262696, 0.011345919609069824, 0.011208703994750976, 0.011266048431396485, 0.011198464393615723, 0.011206656455993653, 0.011239423751831054, 0.011283455848693847, 0.01124454402923584, 0.01136844825744629, 0.011285504341125489, 0.011257856369018555, 0.01122815990447998, 0.01116982364654541, 0.011263968467712403, 0.011196415901184082, 0.011222016334533692, 0.011268095970153809, 0.011254783630371093, 0.011423744201660157, 0.011216896057128906, 0.011208703994750976, 0.011232255935668945, 0.02350592041015625, 0.0112609281539917, 0.012005375862121583, 0.011383808135986329, 0.011304960250854493, 0.011283455848693847, 0.011226112365722657, 0.011202560424804688, 0.011263999938964844, 0.011246591567993165, 0.011610112190246581, 0.011381759643554687, 0.011267071723937988, 0.01233510398864746, 0.011413503646850585, 0.011245632171630859, 0.011271103858947754, 0.011197440147399903, 0.011265024185180664, 0.011243519783020019, 0.011241472244262696, 0.011213855743408203, 0.011274208068847656, 0.011226112365722657, 0.011177984237670899, 0.01124556827545166, 0.01125376033782959, 0.011276288032531738, 0.011593728065490723, 0.012101632118225097, 0.011258879661560058, 0.011426848411560058, 0.011340767860412597, 0.011250687599182128, 0.011226112365722657, 0.01131827163696289, 0.011140095710754394, 0.011276288032531738, 0.011244576454162597, 0.011317215919494628, 0.011682815551757812, 0.01140121555328369, 0.01131827163696289, 0.011275263786315918, 0.011298815727233886, 0.011229184150695801, 0.01125273609161377, 0.011293696403503419, 0.011226112365722657, 0.011249664306640626, 0.011280384063720703, 0.011209728240966797, 0.011223039627075194, 0.011274239540100098, 0.011243552207946777, 0.01130288028717041, 0.011293696403503419, 0.011228192329406738, 0.011246560096740722, 0.011202560424804688, 0.01121177577972412, 0.01123737621307373, 0.011276288032531738, 0.02346188735961914, 0.011205632209777832, 0.011248640060424805, 0.011196415901184082, 0.011185152053833008, 0.011194368362426758, 0.011103232383728028, 0.011118592262268067, 0.010899456024169921, 0.010898431777954102, 0.010889216423034668, 0.011025407791137695, 0.0111595516204834, 0.011188223838806152, 0.011126784324645997, 0.011218943595886231, 0.011179007530212403, 0.011150336265563965, 0.011134976387023926, 0.011125760078430176, 0.011155455589294434, 0.011191295623779298, 0.011165696144104004, 0.01113804817199707, 0.011150336265563965, 0.01121996784210205, 0.011215871810913085, 0.011169792175292969, 0.01112985610961914, 0.011187199592590333, 0.011195391654968261, 0.011234304428100587, 0.011191295623779298, 0.011135999679565429, 0.011169792175292969, 0.011190272331237794, 0.011175935745239257, 0.011165696144104004, 0.010827775955200195, 0.010895359992980956, 0.010907648086547851, 0.011126784324645997, 0.011133952140808106, 0.011169792175292969, 0.011165696144104004, 0.011185152053833008, 0.011100159645080567, 0.011106304168701172, 0.011142144203186035, 0.011210751533508301, 0.011107328414916993, 0.011194368362426758, 0.011146240234375, 0.011167743682861327, 0.011464703559875488, 0.011312128067016602, 0.011296768188476563, 0.011158528327941895, 0.01112166404724121, 0.011174912452697755, 0.01123532772064209, 0.011336704254150391, 0.011124735832214355, 0.022219776153564453, 0.011018239974975585, 0.011139072418212891, 0.01112985610961914, 0.011140095710754394, 0.011172863960266113, 0.0111278076171875, 0.011171872138977051, 0.011145183563232422, 0.011186176300048829, 0.011166720390319825, 0.011091967582702637, 0.011150336265563965, 0.011148287773132324, 0.011134976387023926, 0.011150336265563965, 0.011160575866699218, 0.011149312019348144, 0.0111278076171875, 0.01112063980102539, 0.01131007957458496, 0.01112985610961914, 0.011163647651672364, 0.01115340805053711, 0.01120358371734619, 0.011196415901184082, 0.011167743682861327, 0.011165696144104004, 0.011246591567993165, 0.011224063873291015, 0.011179007530212403, 0.011167743682861327, 0.011182080268859864, 0.011175935745239257, 0.011095040321350098, 0.011188223838806152, 0.01114521598815918, 0.011164671897888183, 0.011201536178588867, 0.01122815990447998, 0.01115340805053711, 0.011130880355834961, 0.011168767929077148, 0.011240447998046875, 0.01115340805053711, 0.01117081642150879, 0.011123711585998536, 0.011975680351257324, 0.014036992073059081, 0.011767807960510255, 0.011272192001342773, 0.011220992088317871, 0.011161600112915039, 0.011171839714050292, 0.011180031776428222, 0.011186176300048829, 0.011100159645080567, 0.01124454402923584, 0.011142144203186035, 0.011199487686157226, 0.011167776107788087, 0.011124704360961914, 0.011181056022644043, 0.023325727462768554, 0.01110524845123291, 0.011058176040649414, 0.010943488121032715, 0.010982399940490722, 0.010989567756652833, 0.011012096405029297, 0.011171839714050292, 0.011198464393615723, 0.011144224166870118, 0.011127776145935058, 0.011077695846557618, 0.011132863998413086, 0.011165696144104004, 0.011160575866699218, 0.011229184150695801, 0.01121177577972412, 0.011184127807617187, 0.011172863960266113, 0.01111244773864746, 0.011144191741943359, 0.011133952140808106, 0.011180031776428222, 0.011130880355834961, 0.011205632209777832, 0.011103232383728028, 0.011167743682861327, 0.01122713565826416, 0.011155455589294434, 0.01112063980102539, 0.011210751533508301, 0.011217920303344727, 0.011107328414916993, 0.011163647651672364, 0.01111244773864746, 0.011090944290161133, 0.011166720390319825, 0.011087871551513672, 0.011205632209777832, 0.01115340805053711, 0.011155455589294434, 0.01112063980102539, 0.011190272331237794, 0.01112985610961914, 0.011212800025939941, 0.01121177577972412, 0.011147263526916504, 0.011180031776428222, 0.011126784324645997, 0.011220992088317871, 0.011133983612060546, 0.01121174430847168, 0.011220992088317871, 0.011192319869995117, 0.011101183891296386, 0.011101183891296386, 0.011166751861572266, 0.011146240234375, 0.011119584083557128, 0.011091967582702637, 0.011132927894592285, 0.011150336265563965, 0.011187199592590333, 0.023353343963623048, 0.011106304168701172, 0.011140095710754394, 0.01115340805053711, 0.01093120002746582, 0.01091481590270996, 0.01095372772216797, 0.011004927635192872, 0.01112985610961914, 0.011101183891296386, 0.011104255676269532, 0.010960895538330078, 0.01094758415222168, 0.010938367843627929, 0.011023360252380371, 0.011216896057128906, 0.01113702392578125, 0.011067392349243164, 0.010962944030761718, 0.010912768363952637, 0.010910719871520995, 0.010944512367248535, 0.010978303909301757, 0.011214847564697266, 0.01115135955810547, 0.011186176300048829, 0.011133952140808106, 0.011190272331237794, 0.011031552314758301, 0.010978303909301757, 0.010934271812438966, 0.0111278076171875, 0.011119615554809571, 0.011147263526916504, 0.011101183891296386, 0.011167743682861327, 0.011180031776428222, 0.011135999679565429, 0.011117568016052246, 0.011113471984863281, 0.011185152053833008, 0.01125376033782959, 0.011240447998046875, 0.011155455589294434, 0.011187199592590333, 0.011194368362426758, 0.011183103561401368, 0.011171839714050292, 0.011149312019348144, 0.011119615554809571, 0.011183103561401368, 0.011135999679565429, 0.011386879920959473, 0.011439104080200196, 0.01154355239868164, 0.011313152313232423, 0.0112609281539917, 0.011216896057128906, 0.011262975692749023, 0.011165696144104004, 0.011180031776428222, 0.011125823974609376, 0.011188159942626954, 0.02225049591064453, 0.01052569580078125, 0.010568703651428223, 0.010543104171752929, 0.010622976303100586, 0.010621952056884766, 0.010568703651428223, 0.010541055679321289, 0.010622976303100586, 0.010554368019104005, 0.010582015991210938, 0.010565631866455078, 0.010530816078186036, 0.010882047653198243, 0.011097087860107421, 0.01112985610961914, 0.01112985610961914, 0.011177984237670899, 0.011173888206481934, 0.011111424446105958, 0.011111488342285156, 0.011157440185546876, 0.011141119956970215, 0.01115135955810547, 0.011143168449401856, 0.01112883186340332, 0.011140095710754394, 0.01136844825744629, 0.01112166404724121, 0.011107328414916993, 0.011119615554809571, 0.011123711585998536, 0.011150336265563965, 0.01120358371734619, 0.01112063980102539, 0.011146271705627441, 0.011225055694580079, 0.011163647651672364, 0.011206656455993653, 0.011156479835510253, 0.011208703994750976, 0.011188223838806152, 0.011194368362426758, 0.011114496231079102, 0.01114521598815918, 0.011217920303344727, 0.011176959991455078, 0.01121177577972412, 0.011148287773132324, 0.010975232124328613, 0.010976256370544434, 0.010999808311462403, 0.011003904342651367, 0.011169792175292969, 0.011200511932373047, 0.011224063873291015, 0.011240447998046875, 0.011239423751831054, 0.010964991569519043, 0.010970111846923827, 0.011025407791137695, 0.011190272331237794, 0.011210783958435059, 0.022187999725341797, 0.010556415557861328, 0.010572799682617188, 0.010587136268615722, 0.010652671813964844, 0.01060969638824463, 0.010614751815795898, 0.010569727897644043, 0.010598400115966796, 0.010579999923706054, 0.010570719718933105, 0.010619903564453125, 0.010581024169921875, 0.010570752143859862, 0.010605536460876464, 0.010647551536560058, 0.010609663963317872, 0.010669055938720704, 0.010605567932128907, 0.010513407707214355, 0.010582015991210938, 0.010595328330993652, 0.010590208053588868, 0.010577919960021973, 0.010600447654724121, 0.01055846405029297, 0.011377663612365722, 0.011400192260742188, 0.011234304428100587, 0.011131903648376466, 0.011103232383728028, 0.011122688293457032, 0.011140095710754394, 0.011132927894592285, 0.01112985610961914, 0.01115135955810547, 0.011166720390319825, 0.011167743682861327, 0.011173888206481934, 0.011142144203186035, 0.011146240234375, 0.011175935745239257, 0.01126195240020752, 0.011149312019348144, 0.011132927894592285, 0.011160575866699218, 0.01123840045928955, 0.011169792175292969, 0.011114496231079102, 0.011144191741943359, 0.011139072418212891, 0.010988544464111329, 0.010944512367248535, 0.010891263961791992, 0.010995712280273438, 0.011147263526916504, 0.011293696403503419, 0.011174912452697755, 0.011168767929077148, 0.011181056022644043, 0.01122815990447998, 0.011097087860107421, 0.011144191741943359, 0.022222848892211915, 0.010568703651428223, 0.010520640373229981, 0.010560447692871094, 0.010555392265319824, 0.010669055938720704, 0.010604543685913086, 0.010606592178344726, 0.010740736007690429, 0.010578944206237792, 0.010570752143859862, 0.010628095626831055, 0.010579968452453613, 0.010572799682617188, 0.010601471900939942, 0.01063424015045166, 0.010566656112670898, 0.010643551826477051, 0.010620832443237305, 0.010530816078186036, 0.010621952056884766, 0.010595328330993652, 0.010600447654724121, 0.010604543685913086, 0.010577919960021973, 0.0105830717086792, 0.010615776062011719, 0.010611712455749512, 0.010565631866455078, 0.010570752143859862, 0.010562560081481934, 0.01053593635559082, 0.010546175956726075, 0.01061888027191162, 0.010591232299804687, 0.010600447654724121, 0.010589183807373047, 0.010599424362182617, 0.010596351623535156, 0.010602527618408204, 0.010623968124389649, 0.010583040237426757, 0.010593279838562012, 0.010605567932128907, 0.010557439804077149, 0.01064140796661377, 0.010597375869750977, 0.010564607620239258, 0.010590208053588868, 0.010691583633422852, 0.010604543685913086, 0.010619903564453125, 0.01062604808807373, 0.010612735748291016, 0.010584063529968261, 0.010546175956726075, 0.010605567932128907, 0.010622976303100586, 0.010594304084777833, 0.010560511589050293, 0.010608639717102051, 0.010576895713806152, 0.010612768173217773, 0.023476192474365234, 0.011000831604003907, 0.010892288208007812, 0.011049983978271484, 0.011250687599182128, 0.011179007530212403, 0.01115443229675293, 0.011082847595214844, 0.01117686367034912, 0.011156479835510253, 0.011122688293457032, 0.011104255676269532, 0.011077631950378418, 0.011192319869995117, 0.011141119956970215, 0.011192319869995117, 0.01115340805053711, 0.011197471618652343, 0.011192288398742676, 0.0111278076171875, 0.011124735832214355, 0.011157504081726074, 0.011213824272155762, 0.01112883186340332, 0.0111595516204834, 0.011113471984863281, 0.011140128135681153, 0.011154399871826173, 0.01114521598815918, 0.011107328414916993, 0.011158528327941895, 0.01115340805053711, 0.011140095710754394, 0.011147263526916504, 0.011156479835510253, 0.011115519523620606, 0.011163647651672364, 0.011156479835510253, 0.01115135955810547, 0.01116262435913086, 0.011130880355834961, 0.011122688293457032, 0.011018239974975585, 0.010986495971679687, 0.011015232086181641, 0.011019200325012206, 0.0112609281539917, 0.011164671897888183, 0.011179007530212403, 0.011146240234375, 0.01121177577972412, 0.01115135955810547, 0.011171839714050292, 0.011141119956970215, 0.011281408309936524, 0.01120358371734619, 0.011191295623779298, 0.011168767929077148, 0.011190272331237794, 0.011324480056762695, 0.011292608261108398, 0.011140095710754394, 0.011161600112915039, 0.023377920150756838, 0.011315199851989746, 0.011220992088317871, 0.011183103561401368, 0.011199487686157226, 0.011177984237670899, 0.011420672416687011, 0.011294783592224121, 0.011265983581542969, 0.011155455589294434, 0.011230208396911622, 0.011168767929077148, 0.011123711585998536, 0.01124556827545166, 0.011197440147399903, 0.011199487686157226, 0.01115443229675293, 0.011144191741943359, 0.01113804817199707, 0.011110400199890137, 0.011140095710754394, 0.011111424446105958, 0.011141119956970215, 0.011123711585998536, 0.0111595516204834, 0.011122688293457032, 0.011305983543395997, 0.01131929588317871, 0.011192319869995117, 0.011145248413085937, 0.011159520149230957, 0.011119615554809571, 0.011196415901184082, 0.01122815990447998, 0.011187199592590333, 0.011146240234375, 0.0111595516204834, 0.011174976348876953, 0.011206591606140136, 0.011270144462585448, 0.01117081642150879, 0.011179007530212403, 0.011171839714050292, 0.01115443229675293, 0.011140095710754394, 0.011191295623779298, 0.011165696144104004, 0.011374591827392578, 0.01121996784210205, 0.011220992088317871, 0.011185152053833008, 0.011156479835510253, 0.011411456108093262, 0.011210751533508301, 0.01112166404724121, 0.011126784324645997, 0.011168767929077148, 0.011199487686157226, 0.011240447998046875, 0.011131903648376466, 0.011180031776428222, 0.01113702392578125, 0.011262975692749023, 0.023784448623657226, 0.011406335830688476, 0.011194368362426758, 0.010918911933898925, 0.010921983718872071, 0.011123711585998536, 0.01111244773864746, 0.011032575607299805, 0.011065343856811523, 0.011059200286865235, 0.011974656105041503, 0.01143295955657959, 0.011126784324645997, 0.011075648307800293, 0.01116870403289795, 0.011183103561401368, 0.011117568016052246, 0.011180031776428222, 0.011175935745239257, 0.011152383804321288, 0.011309056282043458, 0.011322367668151855, 0.011193344116210938, 0.011293696403503419, 0.011202560424804688, 0.011185152053833008, 0.011208703994750976, 0.011263999938964844, 0.01115135955810547, 0.011171839714050292, 0.011182080268859864, 0.011140095710754394, 0.011177984237670899, 0.011115519523620606, 0.011192352294921874, 0.011250656127929687, 0.011204607963562012, 0.011172863960266113, 0.011143168449401856, 0.011181056022644043, 0.011270144462585448, 0.011250687599182128, 0.01125376033782959, 0.011148287773132324, 0.011201536178588867, 0.011172863960266113, 0.01122815990447998, 0.011149312019348144, 0.011142144203186035, 0.011180031776428222, 0.011180031776428222, 0.011220992088317871, 0.011209728240966797, 0.011199487686157226, 0.011199487686157226, 0.011239423751831054, 0.011174912452697755, 0.01120358371734619, 0.011176992416381835, 0.011191264152526855, 0.011183103561401368, 0.01112063980102539, 0.011161600112915039, 0.02329599952697754, 0.011206656455993653, 0.011105279922485351, 0.011160575866699218, 0.011181056022644043, 0.011124735832214355, 0.010936320304870606, 0.01092300796508789, 0.011149312019348144, 0.01091481590270996, 0.010949631690979005, 0.01113702392578125, 0.011117568016052246, 0.011149312019348144, 0.01125376033782959, 0.011171839714050292, 0.011158528327941895, 0.011171839714050292, 0.01114521598815918, 0.011100159645080567, 0.011187264442443848, 0.011203519821166992, 0.011156479835510253, 0.011087871551513672, 0.011157504081726074, 0.011142144203186035, 0.011171839714050292, 0.011243519783020019, 0.011124735832214355, 0.011180031776428222, 0.011176959991455078, 0.011116543769836425, 0.011148287773132324, 0.011171839714050292, 0.01115443229675293, 0.01115340805053711, 0.011131903648376466, 0.011108415603637695, 0.011206591606140136, 0.011194368362426758, 0.011298815727233886, 0.01116262435913086, 0.011177984237670899, 0.011129920005798339, 0.01116051197052002, 0.011156479835510253, 0.011405311584472656, 0.011185152053833008, 0.011182080268859864, 0.011201536178588867, 0.011184127807617187, 0.011123711585998536, 0.011195391654968261, 0.01122713565826416, 0.011183199882507324, 0.011243424415588378, 0.011175935745239257, 0.01132748794555664, 0.011174912452697755, 0.011171839714050292, 0.011160575866699218, 0.011133952140808106, 0.011172863960266113, 0.022174720764160157, 0.010556415557861328, 0.01095372772216797, 0.011195391654968261, 0.011160575866699218, 0.011092032432556152, 0.01113491153717041, 0.01117084789276123, 0.011182047843933106, 0.011116543769836425, 0.011103232383728028, 0.011118592262268067, 0.011139072418212891, 0.011116543769836425, 0.011149312019348144, 0.011268095970153809, 0.011103232383728028, 0.011250687599182128, 0.011189248085021973, 0.011140159606933594, 0.011080639839172363, 0.011131903648376466, 0.011133952140808106, 0.011148287773132324, 0.01115443229675293, 0.011152383804321288, 0.011223039627075194, 0.011332608222961426, 0.011132927894592285, 0.010942463874816894, 0.010919936180114746, 0.010942463874816894, 0.010935296058654785, 0.01093734359741211, 0.011166720390319825, 0.010996800422668457, 0.010977215766906738, 0.010974207878112792, 0.010911744117736816, 0.010969087600708008, 0.01095577621459961, 0.011114496231079102, 0.011009023666381837, 0.01102233600616455, 0.010984448432922364, 0.0109619197845459, 0.0109486083984375, 0.010946559906005859, 0.010987520217895508, 0.011146240234375, 0.01111244773864746, 0.011174912452697755, 0.011181056022644043, 0.011263999938964844, 0.011266048431396485, 0.011032575607299805, 0.010979328155517578, 0.011013119697570802, 0.010957823753356934, 0.010988544464111329, 0.011087871551513672, 0.011174943923950196, 0.011246560096740722]",tokens/s,88.49256396436706,,,main,False,False,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,-,-,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 106, in _inner_fn validate_repo_id(arg_value) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 160, in validate_repo_id raise HFValidationError( huggingface_hub.errors.HFValidationError: Repo id must use alphanumeric chars or '-', '_', '.', '--' and '..' are forbidden, '-' and '.' cannot start or end the name, max length is 96: '-'. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 466, in cached_file raise EnvironmentError( OSError: Incorrect path_or_model_id: '-'. Please provide either the path to a local folder or the repo_id of a model on the Hub. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,stabilityai/stablelm-2-12b,stabilityai/stablelm-2-12b,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ self.load_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 250, in load_model_with_no_weights self.pretrained_model = self.automodel_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3710, in from_pretrained model = cls(config, *model_args, **model_kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/stablelm/modeling_stablelm.py"", line 1188, in __init__ self.model = StableLmModel(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/stablelm/modeling_stablelm.py"", line 966, in __init__ [StableLmDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/stablelm/modeling_stablelm.py"", line 966, in [StableLmDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/stablelm/modeling_stablelm.py"", line 748, in __init__ self.mlp = StableLmMLP(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/stablelm/modeling_stablelm.py"", line 206, in __init__ self.down_proj = nn.Linear(self.intermediate_size, self.hidden_size, bias=False) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/linear.py"", line 98, in __init__ self.weight = Parameter(torch.empty((out_features, in_features), **factory_kwargs)) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_device.py"", line 78, in __torch_function__ return func(*args, **kwargs) torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 136.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,v,v,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/v/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66932582-1e5003da4d0765525ccfcb16;6ac2c33f-5e42-4697-bd12-a9125e25a5ba) Repository Not Found for url: https://huggingface.co/v/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: v is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,openai-community/gpt2,openai-community/gpt2,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1204.568064,1032.323072,0.0,385.875968,318.907904,s,22,0.16983635187149052,0.00771983417597684,0.0007607731323240949,0.007601696014404297,0.007774636840820312,0.008059390234947204,0.010461356220245358,"[0.011095999717712402, 0.008073887825012206, 0.007334271907806397, 0.007314432144165039, 0.0071922879219055175, 0.007555744171142578, 0.00756934404373169, 0.007273280143737793, 0.007361184120178223, 0.007558623790740967, 0.007629280090332031, 0.0076835198402404785, 0.007595935821533203, 0.007606783866882324, 0.007643680095672608, 0.007644288063049317, 0.007616256237030029, 0.007358751773834229, 0.00759660816192627, 0.007783936023712158, 0.007657311916351319, 0.007690944194793701]",tokens/s,33161.33406034032,kWh,8.940365078050382e-08,4.898381483806803e-08,2.1730629715216028e-07,3.5569376277073213e-07,tokens/kWh,719720239.134496,MB,1204.895744,1032.323072,0.0,385.875968,328.799232,s,22,10.209191284179687,0.46405414928089483,0.009746597556020337,0.465557861328125,0.4700826110839844,0.47039546966552737,0.48987033386230466,"[0.4950452880859375, 0.4652591857910156, 0.45673638916015624, 0.4462599792480469, 0.4533187255859375, 0.4634225769042969, 0.4551590576171875, 0.44665042114257814, 0.45565463256835936, 0.461884765625, 0.46849453735351565, 0.4646158142089844, 0.4639776611328125, 0.46777743530273436, 0.4682359619140625, 0.4681744384765625, 0.46585653686523437, 0.46617596435546876, 0.47025906372070314, 0.46821295166015625, 0.46761724853515624, 0.47040264892578126]",tokens/s,135.76001873407603,kWh,5.27711459329481e-06,2.8914946297763693e-06,9.482387452934294e-06,1.765099667600548e-05,tokens/kWh,3569203.5501678684,,s,1385,10.357688314914704,0.007478475317627945,0.0010003552119716708,0.007383039951324463,0.007503872013092041,0.007754931163787841,0.015381130142211916,"[0.01041919994354248, 0.00820736026763916, 0.008085503578186035, 0.008085503578186035, 0.00808243179321289, 0.007862400054931641, 0.007790463924407959, 0.007799808025360107, 0.007370751857757568, 0.007364607810974121, 0.0074700798988342285, 0.007488512039184571, 0.00758681583404541, 0.007523392200469971, 0.00755398416519165, 0.008069151878356934, 0.007755743980407715, 0.008038399696350097, 0.008651776313781738, 0.007857151985168457, 0.007881728172302246, 0.007774208068847656, 0.0077907519340515135, 0.00786518383026123, 0.0077844481468200685, 0.00764518404006958, 0.007683104038238526, 0.0074925761222839355, 0.007401472091674805, 0.007522304058074952, 0.00743833589553833, 0.007535615921020508, 0.007802879810333252, 0.007635968208312988, 0.007602176189422607, 0.007464960098266602, 0.007529471874237061, 0.007635968208312988, 0.007515135765075683, 0.007827455997467042, 0.007940095901489258, 0.007751679897308349, 0.007764992237091065, 0.007757887840270996, 0.007925695896148681, 0.00774348783493042, 0.007848959922790527, 0.007831552028656007, 0.007799808025360107, 0.007796735763549805, 0.007796735763549805, 0.007903232097625732, 0.008030207633972167, 0.007968768119812012, 0.007972864151000977, 0.00799948787689209, 0.007831552028656007, 0.0077844481468200685, 0.007747583866119385, 0.007703551769256592, 0.007557119846343994, 0.007692287921905518, 0.01633791923522949, 0.007371776103973389, 0.0074106879234313965, 0.007315455913543701, 0.007531519889831543, 0.007741439819335938, 0.007697408199310303, 0.007796735763549805, 0.007781375885009765, 0.007507967948913574, 0.007390207767486572, 0.007321599960327148, 0.007374847888946533, 0.007361536026000977, 0.007402495861053467, 0.007409696102142334, 0.007421919822692871, 0.007324672222137451, 0.007337984085083008, 0.007279615879058838, 0.007336959838867187, 0.0073400321006774905, 0.007327744007110596, 0.007513088226318359, 0.007321599960327148, 0.007280640125274658, 0.007350272178649903, 0.007245823860168457, 0.0074629120826721195, 0.007368703842163086, 0.007280640125274658, 0.007344128131866455, 0.007315455913543701, 0.007327744007110596, 0.007324672222137451, 0.007336959838867187, 0.0073062400817871095, 0.0073134078979492185, 0.007318528175354004, 0.007331840038299561, 0.007329792022705078, 0.007342080116271973, 0.007307263851165771, 0.007323647975921631, 0.007307263851165771, 0.007327744007110596, 0.007422976016998291, 0.0073359360694885255, 0.007371776103973389, 0.007315455913543701, 0.007305215835571289, 0.00724889612197876, 0.007331840038299561, 0.0073431038856506346, 0.007352320194244385, 0.007300096035003662, 0.00733900785446167, 0.0073400321006774905, 0.00734822416305542, 0.007334911823272705, 0.007333888053894043, 0.0073359360694885255, 0.007302144050598145, 0.01490329647064209, 0.007054336071014404, 0.0069918718338012695, 0.0070266880989074704, 0.007069695949554444, 0.0070891518592834475, 0.0070830078125, 0.007050240039825439, 0.00707583999633789, 0.007085055828094483, 0.007012351989746094, 0.007008255958557129, 0.0070563840866088865, 0.007080959796905518, 0.007079936027526855, 0.007049215793609619, 0.007053311824798584, 0.007068672180175781, 0.007078911781311035, 0.006989823818206787, 0.007078911781311035, 0.007073791980743408, 0.007096320152282715, 0.007185408115386963, 0.00709939193725586, 0.007647232055664062, 0.007386112213134765, 0.007338016033172607, 0.007336927890777588, 0.007352320194244385, 0.007318528175354004, 0.007336959838867187, 0.007344128131866455, 0.007329792022705078, 0.00733900785446167, 0.007308288097381592, 0.007290880203247071, 0.007327744007110596, 0.007324672222137451, 0.007367712020874023, 0.007366687774658203, 0.007374783992767334, 0.007314432144165039, 0.007383039951324463, 0.0074291200637817386, 0.007409664154052734, 0.007362559795379638, 0.007367680072784424, 0.00738918399810791, 0.007334911823272705, 0.007387135982513428, 0.007332863807678222, 0.007368703842163086, 0.007333888053894043, 0.007345151901245117, 0.00742195177078247, 0.007374879837036133, 0.007357408046722412, 0.0073564801216125485, 0.007358399868011475, 0.0073062400817871095, 0.007300096035003662, 0.0073431038856506346, 0.015045632362365722, 0.0071198720932006835, 0.00709222412109375, 0.007038976192474365, 0.0070563840866088865, 0.007102464199066162, 0.0070860800743103025, 0.007057407855987549, 0.007061503887176514, 0.007102464199066162, 0.007132160186767578, 0.007087103843688965, 0.007122943878173828, 0.007090176105499267, 0.007156735897064209, 0.007118879795074463, 0.007068640232086182, 0.007062528133392334, 0.007093247890472412, 0.007077919960021973, 0.007062496185302735, 0.007069695949554444, 0.007060480117797851, 0.007066624164581299, 0.00709222412109375, 0.007007232189178467, 0.007046144008636474, 0.007030784130096435, 0.0070594558715820314, 0.007087103843688965, 0.007064576148986816, 0.007049215793609619, 0.007078911781311035, 0.007255040168762207, 0.007126016139984131, 0.007058432102203369, 0.007074816226959229, 0.007073791980743408, 0.0070256638526916505, 0.007060480117797851, 0.007045184135437012, 0.007034815788269043, 0.007046144008636474, 0.007066624164581299, 0.007128064155578613, 0.007109632015228271, 0.007085055828094483, 0.007063551902770996, 0.007071743965148926, 0.007038976192474365, 0.007065599918365479, 0.007045119762420654, 0.007059519767761231, 0.007068607807159424, 0.007053311824798584, 0.007021567821502686, 0.007077888011932373, 0.007031807899475098, 0.007014400005340577, 0.007102464199066162, 0.007048192024230957, 0.007049215793609619, 0.007079936027526855, 0.014714879989624024, 0.007004159927368164, 0.007049215793609619, 0.007030784130096435, 0.007069695949554444, 0.007061503887176514, 0.007067647933959961, 0.00707583999633789, 0.007045119762420654, 0.007070720195770264, 0.007053311824798584, 0.007107583999633789, 0.007108607769012451, 0.007104544162750244, 0.0070778560638427735, 0.007109632015228271, 0.007081984043121338, 0.007049215793609619, 0.007110655784606934, 0.007098368167877197, 0.007087103843688965, 0.007029759883880615, 0.0070522880554199216, 0.0070225920677185055, 0.007068672180175781, 0.007029759883880615, 0.007041024208068848, 0.007018496036529541, 0.007064576148986816, 0.007073791980743408, 0.007018496036529541, 0.007407616138458252, 0.007468031883239746, 0.007352320194244385, 0.007312384128570557, 0.007314432144165039, 0.007318528175354004, 0.007310336112976074, 0.007332863807678222, 0.007349247932434082, 0.007310336112976074, 0.007321599960327148, 0.007218175888061523, 0.007292928218841553, 0.007268352031707764, 0.007350272178649903, 0.007215104103088379, 0.007234560012817383, 0.007362559795379638, 0.007323647975921631, 0.007286784172058106, 0.007238656044006348, 0.007235583782196045, 0.007472127914428711, 0.007321599960327148, 0.007373824119567871, 0.007361536026000977, 0.007367680072784424, 0.007315455913543701, 0.007185408115386963, 0.007207935810089112, 0.007258111953735351, 0.007246880054473877, 0.015384544372558593, 0.0073359360694885255, 0.007353343963623047, 0.007360511779785156, 0.007353343963623047, 0.007361536026000977, 0.0073768959045410155, 0.0073471999168395995, 0.007326720237731933, 0.007336959838867187, 0.007361536026000977, 0.007388160228729248, 0.0074700798988342285, 0.007368703842163086, 0.0073359360694885255, 0.0073359360694885255, 0.007341055870056152, 0.007329792022705078, 0.007336959838867187, 0.007328767776489258, 0.007326720237731933, 0.007316480159759522, 0.007428095817565918, 0.007337984085083008, 0.007367680072784424, 0.007323647975921631, 0.007284736156463623, 0.007309311866760254, 0.00729804801940918, 0.007417856216430664, 0.007373824119567871, 0.007319551944732666, 0.007274496078491211, 0.007320576190948487, 0.007327744007110596, 0.00733900785446167, 0.007342080116271973, 0.007288832187652588, 0.007352320194244385, 0.007384064197540283, 0.007334911823272705, 0.007327744007110596, 0.007341055870056152, 0.007345183849334717, 0.007560160160064697, 0.0073768959045410155, 0.007345151901245117, 0.007375872135162354, 0.007367680072784424, 0.007310336112976074, 0.0073400321006774905, 0.00728985595703125, 0.007333888053894043, 0.007323647975921631, 0.0073359360694885255, 0.007323647975921631, 0.007321599960327148, 0.0074414081573486324, 0.007350272178649903, 0.00733900785446167, 0.007354368209838867, 0.007344128131866455, 0.007352320194244385, 0.015380479812622071, 0.007318528175354004, 0.007279615879058838, 0.007360511779785156, 0.007261184215545655, 0.007245823860168457, 0.007204864025115967, 0.007261184215545655, 0.007258111953735351, 0.0072111358642578125, 0.007275392055511474, 0.007315455913543701, 0.007301119804382325, 0.007341184139251709, 0.007237504005432129, 0.007241727828979493, 0.007192575931549072, 0.007265279769897461, 0.007307263851165771, 0.007305215835571289, 0.0073062400817871095, 0.007306272029876709, 0.0072980160713195805, 0.007377920150756836, 0.007321663856506348, 0.007336895942687988, 0.007334911823272705, 0.007324672222137451, 0.007284736156463623, 0.007290880203247071, 0.007341055870056152, 0.007315455913543701, 0.0073359360694885255, 0.007301119804382325, 0.00727347183227539, 0.007302144050598145, 0.007362559795379638, 0.007322624206542969, 0.007418879985809326, 0.007321599960327148, 0.007328767776489258, 0.0070574722290039065, 0.007052224159240722, 0.00709119987487793, 0.0070225920677185055, 0.007054336071014404, 0.007061503887176514, 0.007060480117797851, 0.007070720195770264, 0.007035903930664063, 0.0069970240592956544, 0.007097311973571777, 0.007181312084197998, 0.007011328220367432, 0.0070522880554199216, 0.007074816226959229, 0.007073791980743408, 0.0070891518592834475, 0.007069695949554444, 0.00708403205871582, 0.007062528133392334, 0.00703385591506958, 0.007067647933959961, 0.014791680335998534, 0.007354368209838867, 0.007095295906066895, 0.007050240039825439, 0.007076863765716553, 0.00709222412109375, 0.007090176105499267, 0.007048192024230957, 0.007065599918365479, 0.007094272136688232, 0.0070891518592834475, 0.007067647933959961, 0.007105535984039306, 0.007112703800201416, 0.007088160037994385, 0.007068640232086182, 0.00709939193725586, 0.007058432102203369, 0.00709222412109375, 0.007073791980743408, 0.007030784130096435, 0.007049215793609619, 0.007229440212249756, 0.007047167778015137, 0.007077888011932373, 0.007035903930664063, 0.007031807899475098, 0.007080959796905518, 0.007066624164581299, 0.007038976192474365, 0.007049215793609619, 0.0070553598403930665, 0.00707583999633789, 0.007047167778015137, 0.007067647933959961, 0.007068672180175781, 0.007070720195770264, 0.00703385591506958, 0.007060480117797851, 0.007064576148986816, 0.007144447803497315, 0.007136256217956543, 0.0069847040176391605, 0.007074816226959229, 0.007093247890472412, 0.007093247890472412, 0.007061503887176514, 0.007039999961853028, 0.007098368167877197, 0.007131135940551757, 0.0070563840866088865, 0.007072768211364746, 0.0071352958679199215, 0.007176127910614014, 0.007051263809204102, 0.007237631797790528, 0.007090176105499267, 0.007069695949554444, 0.0070594558715820314, 0.007041024208068848, 0.007063551902770996, 0.007072768211364746, 0.007070720195770264, 0.014858240127563477, 0.007054336071014404, 0.007064576148986816, 0.0070860800743103025, 0.007095295906066895, 0.007044095993041993, 0.00709119987487793, 0.007087103843688965, 0.007038976192474365, 0.007068672180175781, 0.007064576148986816, 0.007076863765716553, 0.007085055828094483, 0.007045119762420654, 0.007069695949554444, 0.007081984043121338, 0.007097343921661377, 0.007104512214660645, 0.007057407855987549, 0.007085055828094483, 0.007070720195770264, 0.007063551902770996, 0.007021567821502686, 0.007047167778015137, 0.007071807861328125, 0.007054272174835205, 0.007097343921661377, 0.007021567821502686, 0.0070522880554199216, 0.007120895862579346, 0.0070553598403930665, 0.007080959796905518, 0.007177216053009033, 0.00724070405960083, 0.00758681583404541, 0.007448575973510742, 0.007428095817565918, 0.007396351814270019, 0.00733081579208374, 0.007371776103973389, 0.007404543876647949, 0.007351295948028564, 0.0073400321006774905, 0.007336959838867187, 0.007353343963623047, 0.0073359360694885255, 0.0074065918922424315, 0.007365632057189942, 0.007329792022705078, 0.007293983936309815, 0.007301087856292725, 0.007342080116271973, 0.008068096160888672, 0.007449600219726562, 0.0074741759300231934, 0.007407616138458252, 0.0074106879234313965, 0.007334911823272705, 0.007352320194244385, 0.007361536026000977, 0.007346176147460938, 0.007352320194244385, 0.007296031951904297, 0.015389663696289062, 0.007357439994812012, 0.007371776103973389, 0.007364607810974121, 0.007349247932434082, 0.00734822416305542, 0.0073431038856506346, 0.007329855918884277, 0.007304128170013428, 0.0073175039291381834, 0.007336959838867187, 0.0073431038856506346, 0.007300096035003662, 0.007304192066192627, 0.007356416225433349, 0.007391232013702393, 0.007362559795379638, 0.007314432144165039, 0.007341055870056152, 0.007308288097381592, 0.0073697280883789065, 0.007057407855987549, 0.007079936027526855, 0.0070266880989074704, 0.007064576148986816, 0.007069695949554444, 0.00714035177230835, 0.007076863765716553, 0.00704307222366333, 0.00708403205871582, 0.007071743965148926, 0.007070720195770264, 0.007062528133392334, 0.007367680072784424, 0.007411712169647216, 0.0074065918922424315, 0.007396351814270019, 0.007363584041595459, 0.0073994240760803225, 0.00740556812286377, 0.00740556812286377, 0.0074065918922424315, 0.007387135982513428, 0.007434239864349365, 0.007355391979217529, 0.007365632057189942, 0.007404543876647949, 0.0074106879234313965, 0.007417856216430664, 0.007406623840332032, 0.00744652795791626, 0.007376863956451416, 0.007442431926727295, 0.007585792064666748, 0.007540768146514893, 0.007431136131286621, 0.007428095817565918, 0.007366655826568603, 0.007354368209838867, 0.00739737606048584, 0.007473152160644531, 0.007417856216430664, 0.00744652795791626, 0.01552998447418213, 0.00740556812286377, 0.00738918399810791, 0.007488512039184571, 0.007451648235321045, 0.007375872135162354, 0.007414783954620361, 0.007392255783081054, 0.0074403839111328125, 0.007346176147460938, 0.0074035201072692874, 0.0074065918922424315, 0.007407616138458252, 0.007394303798675537, 0.0074711041450500485, 0.007407616138458252, 0.007390207767486572, 0.00738099193572998, 0.007378943920135498, 0.007396351814270019, 0.007400512218475342, 0.0073471360206604, 0.007354368209838867, 0.00734822416305542, 0.007400447845458984, 0.007506944179534912, 0.007393343925476074, 0.007407551765441895, 0.0074332160949707035, 0.007365632057189942, 0.007393280029296875, 0.007422976016998291, 0.007417856216430664, 0.007413760185241699, 0.007398399829864502, 0.00744652795791626, 0.008076288223266602, 0.007924736022949219, 0.0074741759300231934, 0.007420928001403809, 0.007412735939025879, 0.0073768959045410155, 0.007449600219726562, 0.007415808200836181, 0.007364607810974121, 0.007385087966918945, 0.007408639907836914, 0.007345151901245117, 0.007522304058074952, 0.00744652795791626, 0.007435264110565186, 0.007473152160644531, 0.00739737606048584, 0.007414783954620361, 0.007404543876647949, 0.0074414081573486324, 0.007398399829864502, 0.007386112213134765, 0.007395328044891358, 0.007394303798675537, 0.007452672004699707, 0.007427072048187256, 0.007391232013702393, 0.01561292839050293, 0.0074301438331604, 0.007452672004699707, 0.007398399829864502, 0.007396351814270019, 0.0074291200637817386, 0.0073381118774414065, 0.007432064056396484, 0.0074414081573486324, 0.0074926080703735356, 0.007330976009368897, 0.007334752082824707, 0.00733081579208374, 0.007378943920135498, 0.007337984085083008, 0.007308288097381592, 0.007235583782196045, 0.007310336112976074, 0.00743833589553833, 0.007365632057189942, 0.0073768959045410155, 0.007336959838867187, 0.0072837119102478025, 0.007391232013702393, 0.0073431038856506346, 0.007326720237731933, 0.007323647975921631, 0.00729804801940918, 0.007359488010406494, 0.007318528175354004, 0.007326720237731933, 0.007302144050598145, 0.007634943962097168, 0.007422976016998291, 0.007267327785491944, 0.007258111953735351, 0.007333888053894043, 0.007302207946777344, 0.007405504226684571, 0.007341119766235351, 0.00736352014541626, 0.007336959838867187, 0.007318528175354004, 0.007396351814270019, 0.007363584041595459, 0.007320576190948487, 0.007312384128570557, 0.007367743968963623, 0.007434175968170166, 0.007654399871826172, 0.007465983867645264, 0.007383039951324463, 0.0073134078979492185, 0.007375872135162354, 0.007378943920135498, 0.007287807941436767, 0.007364607810974121, 0.007337984085083008, 0.007341055870056152, 0.0075008001327514645, 0.0073175039291381834, 0.007315455913543701, 0.00733900785446167, 0.015458304405212403, 0.007296000003814697, 0.007337984085083008, 0.00729804801940918, 0.007323647975921631, 0.007380127906799316, 0.00736348819732666, 0.007358528137207031, 0.007389056205749512, 0.007334911823272705, 0.007320576190948487, 0.007346176147460938, 0.007337984085083008, 0.007327744007110596, 0.007325695991516113, 0.007292928218841553, 0.007398399829864502, 0.007338143825531006, 0.007310175895690918, 0.007361536026000977, 0.007417984008789062, 0.007294847965240479, 0.007318624019622803, 0.007339935779571533, 0.007356416225433349, 0.007330880165100097, 0.007352255821228028, 0.007327744007110596, 0.007312384128570557, 0.007353375911712647, 0.007387328147888184, 0.007373600006103515, 0.007363584041595459, 0.007331840038299561, 0.007437312126159668, 0.0073175039291381834, 0.007345151901245117, 0.007342080116271973, 0.007284736156463623, 0.007300096035003662, 0.007321599960327148, 0.007332863807678222, 0.007334911823272705, 0.007299071788787842, 0.007346176147460938, 0.007374847888946533, 0.007418879985809326, 0.007355391979217529, 0.007385087966918945, 0.007436287879943848, 0.0074106879234313965, 0.007309311866760254, 0.0072991042137146, 0.007371744155883789, 0.007448575973510742, 0.007453695774078369, 0.007404543876647949, 0.007375872135162354, 0.0073768959045410155, 0.007392255783081054, 0.007379968166351319, 0.007508992195129394, 0.007475200176239013, 0.015498240470886231, 0.007387135982513428, 0.007431168079376221, 0.00738099193572998, 0.007372799873352051, 0.007469056129455566, 0.007420928001403809, 0.00740556812286377, 0.007417856216430664, 0.007393280029296875, 0.007412735939025879, 0.007420928001403809, 0.0074291200637817386, 0.007391232013702393, 0.007435264110565186, 0.007420928001403809, 0.007391232013702393, 0.007380032062530518, 0.007403456211090088, 0.00740556812286377, 0.007404543876647949, 0.007384064197540283, 0.007350304126739502, 0.00739734411239624, 0.007441472053527832, 0.007508927822113037, 0.007403615951538086, 0.007401375770568847, 0.007452672004699707, 0.007448575973510742, 0.00742195177078247, 0.007404543876647949, 0.007458879947662354, 0.007466944217681884, 0.007437312126159668, 0.007404543876647949, 0.007425024032592774, 0.007388160228729248, 0.007378943920135498, 0.007395328044891358, 0.007401472091674805, 0.007472127914428711, 0.0074035201072692874, 0.007322624206542969, 0.007465983867645264, 0.007479296207427978, 0.007442431926727295, 0.0075008001327514645, 0.0074711041450500485, 0.00742195177078247, 0.007450623989105225, 0.007331840038299561, 0.00739737606048584, 0.007432191848754883, 0.0074076480865478515, 0.007423967838287354, 0.007395328044891358, 0.007356416225433349, 0.007387135982513428, 0.007460896015167236, 0.007407584190368652, 0.0074332160949707035, 0.0074997758865356446, 0.015566847801208495, 0.007383039951324463, 0.007413760185241699, 0.007398399829864502, 0.0074332160949707035, 0.0074926080703735356, 0.007423999786376953, 0.007383039951324463, 0.007747583866119385, 0.007550975799560547, 0.00748748779296875, 0.007366655826568603, 0.007409664154052734, 0.00742195177078247, 0.007400447845458984, 0.007478271961212158, 0.007391232013702393, 0.0074035201072692874, 0.007426112174987793, 0.0075847039222717285, 0.007460864067077637, 0.007452672004699707, 0.007407616138458252, 0.007414783954620361, 0.00738918399810791, 0.007448575973510742, 0.007468031883239746, 0.007391232013702393, 0.0074332160949707035, 0.007372799873352051, 0.007370751857757568, 0.007377920150756836, 0.007384064197540283, 0.007387135982513428, 0.007419904232025146, 0.00742195177078247, 0.007367680072784424, 0.007427072048187256, 0.007386112213134765, 0.007359488010406494, 0.007437312126159668, 0.007413760185241699, 0.0073994240760803225, 0.0074301438331604, 0.007449600219726562, 0.007404543876647949, 0.0074301438331604, 0.0074106879234313965, 0.007416831970214844, 0.007452672004699707, 0.007401472091674805, 0.007367680072784424, 0.007414783954620361, 0.007432191848754883, 0.007505919933319092, 0.007428095817565918, 0.007392255783081054, 0.007385087966918945, 0.007402495861053467, 0.0074035201072692874, 0.007396351814270019, 0.007423999786376953, 0.007414783954620361, 0.015532032012939453, 0.007367680072784424, 0.007404543876647949, 0.007422976016998291, 0.00743936014175415, 0.00743936014175415, 0.0074106879234313965, 0.007436287879943848, 0.007402495861053467, 0.007413760185241699, 0.0074301438331604, 0.007417856216430664, 0.007459839820861816, 0.007390207767486572, 0.00748748779296875, 0.007413760185241699, 0.0074291200637817386, 0.007370751857757568, 0.007432191848754883, 0.007427072048187256, 0.007412735939025879, 0.007425024032592774, 0.007402495861053467, 0.007404543876647949, 0.007400447845458984, 0.007401472091674805, 0.007456768035888672, 0.007490560054779053, 0.0074403839111328125, 0.007452672004699707, 0.007434239864349365, 0.007427072048187256, 0.0074301438331604, 0.0074967041015625, 0.00743936014175415, 0.007426047801971435, 0.007392255783081054, 0.007379968166351319, 0.0074414081573486324, 0.007509024143218994, 0.007423967838287354, 0.007371776103973389, 0.00744652795791626, 0.007374847888946533, 0.007366655826568603, 0.007432191848754883, 0.007418879985809326, 0.007409664154052734, 0.007427072048187256, 0.007443456172943115, 0.007431168079376221, 0.007355391979217529, 0.007412735939025879, 0.0074741759300231934, 0.007463935852050781, 0.007543807983398438, 0.007426047801971435, 0.007374847888946533, 0.007394303798675537, 0.00744652795791626, 0.007404543876647949, 0.007416831970214844, 0.007404543876647949, 0.01551360034942627, 0.0073768959045410155, 0.007444479942321777, 0.007412735939025879, 0.007396351814270019, 0.007422976016998291, 0.007427072048187256, 0.007476223945617676, 0.0072325119972229, 0.007314432144165039, 0.007525375843048096, 0.007476223945617676, 0.007453695774078369, 0.007435264110565186, 0.007451648235321045, 0.007251967906951904, 0.007233535766601563, 0.00719974422454834, 0.007192575931549072, 0.007222271919250488, 0.007353375911712647, 0.007359456062316894, 0.0074403839111328125, 0.0074065918922424315, 0.007491583824157715, 0.0074065918922424315, 0.007403584003448487, 0.007288767814636231, 0.00739737606048584, 0.007398399829864502, 0.0073994240760803225, 0.0074065918922424315, 0.007391232013702393, 0.0073768959045410155, 0.007359488010406494, 0.007422976016998291, 0.007407616138458252, 0.007426047801971435, 0.007414783954620361, 0.007396383762359619, 0.007373792171478271, 0.007396351814270019, 0.007383039951324463, 0.007396351814270019, 0.007364607810974121, 0.007379968166351319, 0.007357439994812012, 0.007459839820861816, 0.0074106879234313965, 0.00740556812286377, 0.0074117441177368164, 0.007384031772613525, 0.007416831970214844, 0.007395328044891358, 0.007394303798675537, 0.0074291200637817386, 0.007386112213134765, 0.0074291200637817386, 0.007422976016998291, 0.007375872135162354, 0.007352320194244385, 0.007377920150756836, 0.007392255783081054, 0.014935040473937988, 0.007081984043121338, 0.007071743965148926, 0.007077888011932373, 0.007065599918365479, 0.007081984043121338, 0.007016448020935059, 0.007072768211364746, 0.007077888011932373, 0.007222271919250488, 0.007771135807037354, 0.00790937614440918, 0.007539711952209473, 0.007427072048187256, 0.007445504188537597, 0.0074700798988342285, 0.007508992195129394, 0.007539711952209473, 0.00744755220413208, 0.007559167861938477, 0.007396351814270019, 0.007411712169647216, 0.007394303798675537, 0.007408639907836914, 0.007426047801971435, 0.007393280029296875, 0.007375872135162354, 0.007367680072784424, 0.007375967979431152, 0.007335840225219726, 0.007787615776062012, 0.007509920120239258, 0.007444479942321777, 0.007417856216430664, 0.007432191848754883, 0.007448575973510742, 0.0074403839111328125, 0.007426047801971435, 0.007452672004699707, 0.007428095817565918, 0.007444479942321777, 0.007390207767486572, 0.007373824119567871, 0.007394303798675537, 0.007412735939025879, 0.007372799873352051, 0.007400447845458984, 0.007368703842163086, 0.007371776103973389, 0.007414783954620361, 0.007373824119567871, 0.0074065918922424315, 0.007420928001403809, 0.0074106879234313965, 0.007415808200836181, 0.00738918399810791, 0.007394303798675537, 0.0073994240760803225, 0.007427072048187256, 0.007481344223022461, 0.007452672004699707, 0.007615488052368164, 0.007414783954620361, 0.015503359794616698, 0.0073994240760803225, 0.0074158720970153805, 0.007406527996063232, 0.007411712169647216, 0.007413760185241699, 0.007394303798675537, 0.007742464065551758, 0.0074741759300231934, 0.0076902399063110355, 0.008204287528991699, 0.007608320236206055, 0.007482367992401123, 0.0074403839111328125, 0.007464960098266602, 0.0074741759300231934, 0.007412735939025879, 0.007400479793548584, 0.007428063869476319, 0.007402527809143067, 0.00744547176361084, 0.007416831970214844, 0.007423999786376953, 0.007400447845458984, 0.007398399829864502, 0.007391232013702393, 0.0074301438331604, 0.007414783954620361, 0.007427072048187256, 0.00784281587600708, 0.007472127914428711, 0.007518208026885987, 0.007412735939025879, 0.007436287879943848, 0.007373824119567871, 0.007453695774078369, 0.00753766393661499, 0.007423999786376953, 0.007365632057189942, 0.007428095817565918, 0.007408639907836914, 0.0074065918922424315, 0.0073820161819458, 0.007437312126159668, 0.0074301438331604, 0.007475200176239013, 0.0074106879234313965, 0.007404543876647949, 0.007416831970214844, 0.007494656085968018, 0.007451648235321045, 0.007398399829864502, 0.007356416225433349, 0.007383039951324463, 0.007390207767486572, 0.007407616138458252, 0.0075008001327514645, 0.007443456172943115, 0.007513088226318359, 0.007413760185241699, 0.007374847888946533, 0.00740556812286377, 0.007419904232025146, 0.015706111907958984, 0.007416831970214844, 0.007414783954620361, 0.007367680072784424, 0.007383039951324463, 0.007465983867645264, 0.007415808200836181, 0.0073994240760803225, 0.007544832229614258, 0.007572480201721191, 0.007402495861053467, 0.007402495861053467, 0.007513088226318359, 0.00744755220413208, 0.007390207767486572, 0.007437312126159668, 0.007401472091674805, 0.007349247932434082, 0.007364607810974121, 0.007392255783081054, 0.007417856216430664, 0.0074670081138610836, 0.007475200176239013, 0.0074332160949707035, 0.007486464023590088, 0.007415808200836181, 0.007459839820861816, 0.007414783954620361, 0.007407616138458252, 0.007390207767486572, 0.0074332160949707035, 0.007383039951324463, 0.007368703842163086, 0.007402495861053467, 0.007412735939025879, 0.007453695774078369, 0.007512063980102539, 0.007412735939025879, 0.00738099193572998, 0.007419904232025146, 0.00740556812286377, 0.007427072048187256, 0.00738918399810791, 0.007404543876647949, 0.007396351814270019, 0.0073820161819458, 0.007374847888946533, 0.0074967041015625, 0.0074823999404907224, 0.007394271850585938, 0.007426047801971435, 0.007425024032592774, 0.0074414081573486324, 0.007459839820861816, 0.007450623989105225, 0.007398399829864502, 0.007434239864349365, 0.007374847888946533, 0.00743939208984375, 0.00738915205001831, 0.007408639907836914, 0.007426047801971435, 0.007469056129455566, 0.015568896293640137, 0.007375872135162354, 0.007453695774078369, 0.007475200176239013, 0.007422016143798828, 0.007447487831115722, 0.007414783954620361, 0.0074106879234313965, 0.007624703884124756, 0.007417856216430664, 0.007364607810974121, 0.007427072048187256, 0.007404543876647949, 0.007414783954620361, 0.007408639907836914, 0.007412735939025879, 0.007633920192718506, 0.007391232013702393, 0.007360511779785156, 0.007388160228729248, 0.007395328044891358, 0.007415808200836181, 0.007392255783081054, 0.007388160228729248, 0.0073471999168395995, 0.007425024032592774, 0.007422976016998291, 0.007409664154052734, 0.007401472091674805, 0.007413760185241699, 0.007402495861053467, 0.007417920112609863, 0.007379968166351319, 0.007383999824523926, 0.0075304961204528805, 0.007417856216430664, 0.007520319938659668, 0.0074505600929260256, 0.007416831970214844, 0.007413760185241699, 0.007451648235321045, 0.007383039951324463, 0.0074035201072692874, 0.00742195177078247, 0.007394303798675537, 0.007363584041595459, 0.007385087966918945, 0.007409664154052734, 0.007398399829864502, 0.007391232013702393, 0.007377920150756836, 0.007401472091674805, 0.007412735939025879, 0.007408639907836914, 0.007398399829864502, 0.007414783954620361, 0.007407616138458252, 0.007391232013702393, 0.007383039951324463, 0.007392255783081054, 0.007379968166351319, 0.007384064197540283, 0.007390207767486572, 0.015604736328125, 0.00743936014175415, 0.007373824119567871, 0.00744755220413208, 0.007378943920135498, 0.00742195177078247, 0.007460864067077637, 0.007413760185241699, 0.007361599922180176, 0.007460800170898438, 0.007415808200836181, 0.007408639907836914, 0.007423999786376953, 0.007407616138458252, 0.007416831970214844, 0.0074065918922424315, 0.00739737606048584, 0.007427072048187256, 0.007392255783081054, 0.0074035201072692874, 0.007387135982513428, 0.007505919933319092, 0.007385087966918945, 0.007395359992980957, 0.007390175819396973, 0.007387135982513428, 0.007409664154052734, 0.007387135982513428, 0.007417856216430664, 0.007730175971984863, 0.007521279811859131, 0.007845888137817383, 0.007542816162109375, 0.007914463996887207, 0.008094719886779785, 0.007558144092559814, 0.0074700798988342285, 0.007525375843048096, 0.0074670081138610836, 0.007477248191833496, 0.007378943920135498, 0.00743833589553833, 0.007582719802856445, 0.0074629120826721195, 0.007414783954620361, 0.007391232013702393, 0.007437312126159668, 0.0073431038856506346, 0.007412799835205078, 0.007379903793334961, 0.007402495861053467, 0.007387135982513428, 0.007435264110565186, 0.007411712169647216, 0.007359488010406494, 0.007461887836456299, 0.007560192108154297, 0.007468031883239746, 0.007400447845458984, 0.007452672004699707, 0.007457791805267334, 0.007368703842163086, 0.007425024032592774]",tokens/s,133.71709573512163,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Salesforce/codegen-6B-nl,Salesforce/codegen-6B-nl,cuda,0,42,,,True,,,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",codegen,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-410m,,cuda,0,42,,,,,True,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,217063f5c507ed7cc255df7e1f64c4333a0b4dfe,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,,MB,1278.38208,1713.897472,0.0,1067.450368,1018.226688,s,10,0.1711367034912109,0.017113670349121094,0.0006044720308494674,0.017062576293945312,0.017929014587402342,0.018202859115600586,0.01842193473815918,"[0.018476703643798827, 0.017089599609375, 0.016498239517211914, 0.017047168731689454, 0.017160512924194335, 0.016970815658569335, 0.017868160247802734, 0.01707798385620117, 0.016465215682983397, 0.016482303619384766]",tokens/s,14958.801635041858,kWh,1.9755586151162386e-07,1.0825147990690798e-07,5.015727721237629e-07,8.073801135422947e-07,tokens/kWh,317074938.6888254,MB,1278.38208,1713.897472,0.0,1067.450368,1059.39456,s,10,10.570229980468751,1.057022998046875,0.01434517760269983,1.0608662719726563,1.0653480590820312,1.071996600341797,1.0773154333496093,"[1.0786451416015626, 1.0614322509765626, 1.0485587158203125, 1.0598475341796876, 1.06387060546875, 1.0626763916015625, 1.061619140625, 1.06030029296875, 1.020062744140625, 1.0532171630859375]",tokens/s,59.601352209373765,kWh,1.2347086212728956e-05,6.765710445814322e-06,2.1629694932874916e-05,4.07424915914182e-05,tokens/kWh,1546297.183583883,,s,629,10.7047331676483,0.017018653684655508,0.0020937808719674184,0.01680588722229004,0.01699450836181641,0.017415782165527343,0.03315052490234375,"[0.017915903091430666, 0.0176363525390625, 0.017562623977661132, 0.01740902328491211, 0.017509376525878906, 0.01741619110107422, 0.01743974494934082, 0.01756159973144531, 0.01738444709777832, 0.017515520095825195, 0.01739776039123535, 0.017354751586914064, 0.017535999298095704, 0.01760665512084961, 0.0174202880859375, 0.017514495849609374, 0.01740390396118164, 0.017494016647338868, 0.017672191619873046, 0.017306623458862306, 0.017474559783935546, 0.01769267272949219, 0.01754521560668945, 0.01742233657836914, 0.016741376876831054, 0.016734207153320312, 0.016705535888671876, 0.016700416564941405, 0.017678335189819337, 0.017415168762207032, 0.0172042236328125, 0.016962560653686523, 0.016944128036499022, 0.016825344085693358, 0.016714752197265623, 0.016738304138183592, 0.01679769515991211, 0.01680281639099121, 0.016870399475097657, 0.016694271087646484, 0.0168222713470459, 0.01675468826293945, 0.016842752456665038, 0.016837631225585938, 0.016772096633911132, 0.01677414321899414, 0.016746496200561522, 0.016700416564941405, 0.01683865547180176, 0.016744447708129884, 0.016858112335205077, 0.01675468826293945, 0.016859136581420898, 0.01679667282104492, 0.016776191711425782, 0.01670867156982422, 0.016846847534179688, 0.01671878433227539, 0.016760831832885743, 0.01681715202331543, 0.016878591537475587, 0.01684787178039551, 0.034307071685791016, 0.01676288032531738, 0.016781312942504883, 0.016820287704467772, 0.01678335952758789, 0.01689695930480957, 0.016829439163208008, 0.01681920051574707, 0.016767999649047852, 0.016933887481689454, 0.01680793571472168, 0.017054719924926756, 0.016825344085693358, 0.016721920013427736, 0.01678339195251465, 0.016764896392822266, 0.016841728210449217, 0.016765983581542968, 0.016816095352172853, 0.01677107238769531, 0.017076223373413087, 0.016883712768554687, 0.01681612777709961, 0.01677414321899414, 0.016990207672119142, 0.016911359786987306, 0.01681305694580078, 0.016841728210449217, 0.016842752456665038, 0.016939008712768554, 0.01681100845336914, 0.01680793571472168, 0.01686425590515137, 0.016858112335205077, 0.016897024154663084, 0.01678643226623535, 0.01679871940612793, 0.0168222713470459, 0.016845823287963867, 0.01698099136352539, 0.01678745651245117, 0.016850944519042968, 0.01677414321899414, 0.01685606384277344, 0.01683660888671875, 0.016849920272827147, 0.016753664016723634, 0.016878591537475587, 0.016842752456665038, 0.016767999649047852, 0.016804927825927733, 0.0167126407623291, 0.0168222713470459, 0.017168384552001953, 0.016925695419311524, 0.01680179214477539, 0.01682636833190918, 0.016888832092285155, 0.016784383773803712, 0.016837631225585938, 0.016861183166503906, 0.01684480094909668, 0.016865280151367186, 0.03311513519287109, 0.016079872131347657, 0.016054271697998047, 0.01617407989501953, 0.016210943222045898, 0.01614028739929199, 0.01617100715637207, 0.016243711471557617, 0.016130048751831053, 0.016185375213623048, 0.016128992080688475, 0.01620479965209961, 0.016130048751831053, 0.01617203140258789, 0.016676864624023437, 0.01662156867980957, 0.016734272003173827, 0.016788415908813477, 0.0167956485748291, 0.01679871940612793, 0.016792575836181642, 0.01685196876525879, 0.01678335952758789, 0.01677824020385742, 0.016889856338500975, 0.016748544692993163, 0.016763904571533202, 0.016746496200561522, 0.01677107238769531, 0.016729087829589845, 0.016717824935913086, 0.016709632873535156, 0.016737279891967775, 0.01676697540283203, 0.0168089599609375, 0.016707584381103514, 0.016772096633911132, 0.01665433692932129, 0.016546815872192384, 0.016548864364624022, 0.01656831932067871, 0.01680076789855957, 0.01682636833190918, 0.01704550361633301, 0.01692163276672363, 0.01686832046508789, 0.016888832092285155, 0.016829439163208008, 0.016906240463256835, 0.016746496200561522, 0.01685196876525879, 0.016767999649047852, 0.01682022476196289, 0.016829439163208008, 0.01683148765563965, 0.016578559875488282, 0.016721920013427736, 0.01681920051574707, 0.0168089599609375, 0.01684889602661133, 0.01684889602661133, 0.016788480758666992, 0.01679052734375, 0.0344453125, 0.016760831832885743, 0.016747520446777343, 0.01675775909423828, 0.016744447708129884, 0.0168089599609375, 0.016782367706298828, 0.016795616149902343, 0.016745471954345705, 0.01677004814147949, 0.016849920272827147, 0.016886783599853517, 0.016883712768554687, 0.016792575836181642, 0.01679769515991211, 0.01678335952758789, 0.016873472213745116, 0.016793600082397463, 0.01678233528137207, 0.01682636833190918, 0.01681715202331543, 0.01680384063720703, 0.016781312942504883, 0.016855039596557618, 0.016827392578125, 0.01684889602661133, 0.01682022476196289, 0.016852991104125976, 0.01677414321899414, 0.016788543701171874, 0.016879552841186522, 0.01677004814147949, 0.01680281639099121, 0.01680793571472168, 0.01680076789855957, 0.016821247100830078, 0.01677926445007324, 0.016753664016723634, 0.01681920051574707, 0.0168724479675293, 0.01676288032531738, 0.016887807846069337, 0.016788480758666992, 0.016764928817749023, 0.016845823287963867, 0.01722163200378418, 0.01670662307739258, 0.016764863967895508, 0.016740352630615234, 0.016862207412719727, 0.016849920272827147, 0.016827392578125, 0.016772096633911132, 0.01676595115661621, 0.016769023895263673, 0.016880640029907225, 0.01683251190185547, 0.016883712768554687, 0.01680998420715332, 0.016850944519042968, 0.01681100845336914, 0.01679667282104492, 0.016827392578125, 0.034411518096923825, 0.016741376876831054, 0.01679052734375, 0.016880640029907225, 0.01681715202331543, 0.0167587833404541, 0.01679974365234375, 0.01681920051574707, 0.01677004814147949, 0.016827392578125, 0.016803871154785155, 0.017103839874267578, 0.016885759353637696, 0.01676192092895508, 0.01679046440124512, 0.016842752456665038, 0.01683558464050293, 0.0168222713470459, 0.016849920272827147, 0.017146879196166993, 0.01835212707519531, 0.01741926383972168, 0.016946176528930663, 0.016669696807861328, 0.016795679092407225, 0.016792543411254884, 0.016785408020019533, 0.01684480094909668, 0.01702400016784668, 0.01662566375732422, 0.016874496459960937, 0.016769023895263673, 0.017132543563842775, 0.017356800079345702, 0.016907264709472656, 0.01681612777709961, 0.01682329559326172, 0.01675775909423828, 0.016529407501220703, 0.016728063583374024, 0.016826431274414064, 0.01677510452270508, 0.01676700782775879, 0.016827360153198242, 0.016903167724609376, 0.01681612777709961, 0.016910335540771485, 0.01682329559326172, 0.016756736755371093, 0.01681305694580078, 0.01682329559326172, 0.016752639770507814, 0.016792575836181642, 0.016683008193969725, 0.016594944000244142, 0.016623615264892578, 0.016518144607543944, 0.01702809524536133, 0.01704140853881836, 0.017705984115600586, 0.017150976181030272, 0.01684787178039551, 0.016837631225585938, 0.03449651336669922, 0.016915456771850586, 0.017155071258544922, 0.016904191970825197, 0.016784383773803712, 0.01680281639099121, 0.01677516746520996, 0.017356800079345702, 0.016895008087158204, 0.016777183532714842, 0.016731136322021483, 0.016880735397338868, 0.016845727920532228, 0.016830495834350586, 0.01685091209411621, 0.01677414321899414, 0.01722470474243164, 0.016881696701049803, 0.01688470458984375, 0.0168222713470459, 0.01678643226623535, 0.01683967971801758, 0.01680281639099121, 0.016902143478393555, 0.016865280151367186, 0.016886783599853517, 0.016825344085693358, 0.016863231658935548, 0.016776191711425782, 0.01683046340942383, 0.01678233528137207, 0.016785408020019533, 0.016982015609741212, 0.016902143478393555, 0.01683558464050293, 0.016884735107421875, 0.016909311294555664, 0.016873472213745116, 0.016907264709472656, 0.016908287048339844, 0.01680899238586426, 0.01685807991027832, 0.01682022476196289, 0.016761856079101564, 0.016820255279541015, 0.016922592163085937, 0.016743423461914063, 0.01675468826293945, 0.01681920051574707, 0.016874496459960937, 0.016776191711425782, 0.01678233528137207, 0.016785408020019533, 0.016897024154663084, 0.016772096633911132, 0.01686944007873535, 0.016851903915405274, 0.016892927169799805, 0.016778303146362306, 0.016824256896972655, 0.016816192626953125, 0.01689900779724121, 0.01683660888671875, 0.036327423095703124, 0.01703219223022461, 0.01710393524169922, 0.016817087173461913, 0.016899072647094726, 0.016750591278076172, 0.016756767272949218, 0.016770015716552733, 0.0168222713470459, 0.016827392578125, 0.01684889602661133, 0.017040384292602538, 0.01682022476196289, 0.016733184814453125, 0.016862207412719727, 0.016788480758666992, 0.016781312942504883, 0.016882688522338866, 0.01682329559326172, 0.01659187126159668, 0.01661337661743164, 0.016764928817749023, 0.01683456039428711, 0.01681612777709961, 0.01678233528137207, 0.01680076789855957, 0.01684889602661133, 0.01679974365234375, 0.01685196876525879, 0.016784383773803712, 0.01676288032531738, 0.01678745651245117, 0.01684889602661133, 0.01682022476196289, 0.01683967971801758, 0.01679667282104492, 0.016899072647094726, 0.01677926445007324, 0.01672300720214844, 0.016747455596923828, 0.016737279891967775, 0.01683558464050293, 0.016785408020019533, 0.016846847534179688, 0.01685196876525879, 0.01682841682434082, 0.01677926445007324, 0.0168089599609375, 0.016728063583374024, 0.016881664276123046, 0.016862207412719727, 0.016842752456665038, 0.01679052734375, 0.0168724479675293, 0.016909311294555664, 0.016869375228881836, 0.01679974365234375, 0.01685203170776367, 0.01689900779724121, 0.016879615783691407, 0.016969728469848632, 0.01680281639099121, 0.016916479110717773, 0.03441049575805664, 0.01678233528137207, 0.016878591537475587, 0.016867328643798828, 0.016906272888183593, 0.016855072021484375, 0.016803775787353516, 0.01684889602661133, 0.016743423461914063, 0.01686425590515137, 0.016777215957641603, 0.01679155158996582, 0.01683148765563965, 0.01679871940612793, 0.016691200256347655, 0.01658060836791992, 0.016900096893310547, 0.016781312942504883, 0.01680588722229004, 0.01658880043029785, 0.016672767639160157, 0.016756736755371093, 0.016767999649047852, 0.01685196876525879, 0.01683456039428711, 0.01684480094909668, 0.01677414321899414, 0.016829439163208008, 0.0168089599609375, 0.0168407039642334, 0.016730112075805666, 0.01677824020385742, 0.01681510353088379, 0.016852991104125976, 0.01660825538635254, 0.016529407501220703, 0.016729087829589845, 0.016850944519042968, 0.016776191711425782, 0.016905216217041014, 0.016899072647094726, 0.016858112335205077, 0.016751615524291993, 0.01678335952758789, 0.01680486488342285, 0.01703321647644043, 0.01685606384277344, 0.0167956485748291, 0.01677824020385742, 0.016827455520629885, 0.016819135665893555, 0.01701171112060547, 0.016941055297851563, 0.01677516746520996, 0.016793600082397463, 0.01680384063720703, 0.01684377670288086, 0.01764659118652344, 0.01704652786254883, 0.01680691146850586, 0.016776191711425782, 0.016875520706176757, 0.016821247100830078, 0.033111038208007815, 0.016175104141235352, 0.01620582389831543, 0.0161474552154541, 0.016202751159667968, 0.016130048751831053, 0.01618841552734375, 0.01616896057128906, 0.01616998481750488, 0.016022527694702148, 0.016058368682861326, 0.016438272476196288, 0.016368640899658202, 0.016148544311523436, 0.016179136276245117, 0.016207872390747072, 0.016157695770263672, 0.016145408630371092, 0.016154624938964843, 0.016175104141235352, 0.016135168075561524, 0.016141311645507812, 0.01617715263366699, 0.016192512512207033, 0.016149503707885742, 0.01617203140258789, 0.01620582389831543, 0.01618841552734375, 0.01617817687988281, 0.01616486358642578, 0.016149503707885742, 0.016175167083740234, 0.016184255599975585, 0.016210943222045898, 0.016226303100585936, 0.016166912078857423, 0.01616281509399414, 0.01618022346496582, 0.01618022346496582, 0.016228351593017578, 0.016114688873291014, 0.016234495162963866, 0.01613926315307617, 0.01623040008544922, 0.01620582389831543, 0.01616896057128906, 0.016210943222045898, 0.01618841552734375, 0.016150527954101563, 0.01621708869934082, 0.01621708869934082, 0.016154624938964843, 0.016206911087036133, 0.016170944213867187, 0.01621504020690918, 0.016286720275878908, 0.01617100715637207, 0.016166912078857423, 0.01623859214782715, 0.016227327346801757, 0.016150527954101563, 0.01616998481750488, 0.01620992088317871, 0.03316428756713867, 0.01618022346496582, 0.01618841552734375, 0.0161976318359375, 0.01616383934020996, 0.0161474552154541, 0.01618022346496582, 0.016453632354736326, 0.016564224243164064, 0.016552959442138672, 0.016711679458618164, 0.016925695419311524, 0.01661235237121582, 0.016556032180786134, 0.016530431747436524, 0.016542720794677734, 0.016748544692993163, 0.01676288032531738, 0.01677516746520996, 0.01685196876525879, 0.016865280151367186, 0.01679052734375, 0.016910335540771485, 0.01702809524536133, 0.01679667282104492, 0.016855039596557618, 0.01680691146850586, 0.016842752456665038, 0.016894975662231446, 0.016906240463256835, 0.016724992752075195, 0.016850944519042968, 0.016756736755371093, 0.016846847534179688, 0.01680486488342285, 0.01681305694580078, 0.016785408020019533, 0.01678950309753418, 0.01676595115661621, 0.016663551330566406, 0.01663385581970215, 0.016707584381103514, 0.016869375228881836, 0.016927743911743166, 0.016885759353637696, 0.016894975662231446, 0.01682636833190918, 0.016753664016723634, 0.016761856079101564, 0.01681100845336914, 0.016777215957641603, 0.016864288330078126, 0.016827360153198242, 0.016713727951049806, 0.01682431983947754, 0.016841728210449217, 0.01683558464050293, 0.01680588722229004, 0.016679935455322266, 0.01680691146850586, 0.0167587833404541, 0.016886783599853517, 0.016726015090942382]",tokens/s,58.75905453682437,,,main,False,False,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,.,.,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 373, in cached_file raise EnvironmentError( OSError: . does not appear to have a file named config.json. Checkout 'https://huggingface.co/./tree/None' for available files. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,a,a,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/a/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66947498-58a4713947ab72d54b3de1b8;267061e9-cc25-442b-aa61-3d3134f8a64d) Repository Not Found for url: https://huggingface.co/a/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: a is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-7B,Qwen/Qwen-7B,cuda,0,42,,,True,,,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",qwen,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-66b,facebook/opt-66b,cuda,0,42,,,True,,,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",opt,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,stabilityai/stablelm-base-alpha-7b,stabilityai/stablelm-base-alpha-7b,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1293.328384,17197.170688,0.0,16550.723584,16322.776064,s,10,0.7599241867065429,0.07599241867065429,0.00024290902259590754,0.07589332962036133,0.07620593872070312,0.0764411766052246,0.07662936691284179,"[0.07667641448974609, 0.07589459228515626, 0.07587564849853516, 0.07587039947509766, 0.07583782196044922, 0.07592412567138672, 0.07589206695556641, 0.07593353271484375, 0.0758659210205078, 0.0761536636352539]",tokens/s,3368.757100750875,kWh,8.99471250999255e-07,4.927137167214823e-07,5.531755267151757e-06,6.923940234872494e-06,tokens/kWh,36973167.20191394,MB,1293.656064,17197.170688,0.0,16550.723584,16467.883008,s,10,22.345561767578126,2.2345561767578124,0.003168750314599697,2.2338046875,2.235604345703125,2.239733447265625,2.243036728515625,"[2.243862548828125, 2.232825927734375, 2.23384765625, 2.2322685546875, 2.23376171875, 2.233000244140625, 2.234686767578125, 2.233957275390625, 2.23345947265625, 2.2338916015625]",tokens/s,28.193518093337296,kWh,2.6314910228555454e-05,1.4420524698562076e-05,0.0001301421588274483,0.00017087759375456578,tokens/kWh,368684.9669154864,,s,629,22.99268313217163,0.036554345202180656,0.009064589335446004,0.035448833465576174,0.03555430908203125,0.035712205505371096,0.11167555816650392,"[0.03600588989257812, 0.03602227020263672, 0.03600384140014649, 0.03599359893798828, 0.03607244873046875, 0.03603558349609375, 0.035985408782958986, 0.03606528091430664, 0.03609804916381836, 0.03605404663085938, 0.036080608367919924, 0.03611443328857422, 0.0359741439819336, 0.03572531127929687, 0.0355153923034668, 0.03543756866455078, 0.035394561767578124, 0.03545600128173828, 0.03553177642822265, 0.03572326278686523, 0.03568544006347656, 0.035393470764160155, 0.03542015838623047, 0.035399681091308595, 0.03540172958374024, 0.035419136047363284, 0.035414142608642576, 0.035396480560302736, 0.035460094451904296, 0.03542732620239258, 0.035434497833251956, 0.03542732620239258, 0.03540172958374024, 0.03543251037597656, 0.03541292953491211, 0.03545804977416992, 0.03542630386352539, 0.03548057556152344, 0.035465217590332034, 0.035448833465576174, 0.035507198333740234, 0.03551846313476562, 0.03546316909790039, 0.0354969596862793, 0.03549491119384766, 0.035484672546386715, 0.03552460861206055, 0.03553484725952148, 0.03548672103881836, 0.03550822448730469, 0.035489791870117186, 0.03552665710449219, 0.03552358245849609, 0.03553177642822265, 0.03551129531860352, 0.03550822448730469, 0.035520511627197264, 0.035517440795898435, 0.03554816055297851, 0.035566593170166014, 0.035568641662597655, 0.035530784606933596, 0.11168867492675781, 0.0353546257019043, 0.03534950256347656, 0.035345535278320316, 0.03540166473388672, 0.03545695877075195, 0.035366912841796876, 0.03536383819580078, 0.03534131240844727, 0.03534131240844727, 0.035335166931152344, 0.035364864349365234, 0.03536588668823242, 0.03535257720947266, 0.035386367797851564, 0.03539763259887695, 0.035399681091308595, 0.03536076736450195, 0.03540172958374024, 0.035383296966552735, 0.035383296966552735, 0.03541401672363281, 0.03536896133422852, 0.035389438629150394, 0.03539046478271484, 0.03542736053466797, 0.03539759826660156, 0.03544166564941406, 0.035416065216064455, 0.03543756866455078, 0.03540377426147461, 0.03544268798828125, 0.035438591003417966, 0.03547340774536133, 0.035409984588623045, 0.03543545532226562, 0.03542015838623047, 0.03544371032714844, 0.035460094451904296, 0.03548262405395508, 0.03546214294433594, 0.03544985580444336, 0.03545804977416992, 0.035460094451904296, 0.03548262405395508, 0.035465217590332034, 0.03547750473022461, 0.03548780822753906, 0.03548976135253906, 0.03550204849243164, 0.03547238540649414, 0.03549491119384766, 0.03555635070800781, 0.0355153923034668, 0.035525665283203126, 0.0355184326171875, 0.03551232147216797, 0.0355153923034668, 0.035517440795898435, 0.03550822448730469, 0.035517440795898435, 0.03553996658325195, 0.0355676155090332, 0.11171225738525391, 0.035378177642822264, 0.0354068489074707, 0.03533315277099609, 0.035334110260009766, 0.035323902130126955, 0.03539865493774414, 0.03533824157714844, 0.035335166931152344, 0.03540070343017578, 0.035465217590332034, 0.035499038696289065, 0.03536790466308594, 0.035388416290283206, 0.035383296966552735, 0.035351551055908204, 0.03538739013671875, 0.03535564804077149, 0.03535871887207031, 0.035399742126464843, 0.03537606430053711, 0.03538534545898438, 0.035369983673095705, 0.035434497833251956, 0.03545702362060547, 0.035438591003417966, 0.035410945892333984, 0.03540889739990234, 0.03538534545898438, 0.03540582275390625, 0.0354068489074707, 0.03545087814331055, 0.03542630386352539, 0.035448833465576174, 0.035454975128173825, 0.035419136047363284, 0.03543961715698242, 0.03544166564941406, 0.03546828842163086, 0.035466270446777345, 0.035501022338867184, 0.03545907211303711, 0.03545395278930664, 0.03549593734741211, 0.03544678497314453, 0.0354856948852539, 0.0354785270690918, 0.035514366149902346, 0.03549184036254883, 0.03548876953125, 0.03547545623779297, 0.0354785270690918, 0.03555430221557617, 0.03554304122924805, 0.035573760986328126, 0.03583180618286133, 0.03582156753540039, 0.035591167449951173, 0.035533824920654294, 0.03553484725952148, 0.03551846313476562, 0.035552257537841796, 0.035527679443359376, 0.11165491485595704, 0.0353259506225586, 0.03532799911499023, 0.035323936462402346, 0.03534332656860351, 0.03532185745239258, 0.035335166931152344, 0.035369983673095705, 0.035356670379638674, 0.03535564804077149, 0.03536076736450195, 0.03532492828369141, 0.03534438323974609, 0.035399681091308595, 0.035388416290283206, 0.035451904296875, 0.035448833465576174, 0.03538534545898438, 0.035410945892333984, 0.03537919998168945, 0.035383296966552735, 0.0353966064453125, 0.03539558410644531, 0.035418113708496096, 0.035388416290283206, 0.03540172958374024, 0.03538534545898438, 0.0353966064453125, 0.03542015838623047, 0.03544371032714844, 0.03541401672363281, 0.03539251327514648, 0.035399681091308595, 0.035410945892333984, 0.035432449340820314, 0.03543142318725586, 0.03544063949584961, 0.035434497833251956, 0.0354252815246582, 0.03542732620239258, 0.035451904296875, 0.03545401763916016, 0.03546515274047852, 0.0354703369140625, 0.03555433654785156, 0.035482593536376957, 0.03546726226806641, 0.03548160171508789, 0.03547545623779297, 0.03548883056640625, 0.035484638214111325, 0.03547747039794922, 0.0355011215209961, 0.035507137298583985, 0.03547955322265625, 0.035484672546386715, 0.03550310516357422, 0.03552153778076172, 0.035530750274658206, 0.03551440048217774, 0.03552150344848633, 0.03550515365600586, 0.03551334381103516, 0.11157196807861328, 0.035304447174072266, 0.03529523086547852, 0.03532502365112305, 0.0353197135925293, 0.03533107376098633, 0.035345409393310545, 0.035426494598388675, 0.035368766784667965, 0.03535974502563476, 0.03537411117553711, 0.0353392333984375, 0.03534745788574219, 0.03535974502563476, 0.035381248474121094, 0.03539046478271484, 0.03536281585693359, 0.03536076736450195, 0.035366912841796876, 0.03536281585693359, 0.03535974502563476, 0.0354068489074707, 0.0354252815246582, 0.03542425537109375, 0.03539046478271484, 0.035379230499267576, 0.035378143310546874, 0.03540172958374024, 0.03542732620239258, 0.03540889739990234, 0.03540176010131836, 0.03541193771362305, 0.035527679443359376, 0.035775489807128906, 0.03570175933837891, 0.03551446533203125, 0.03556035232543946, 0.03545804977416992, 0.03547238540649414, 0.035448833465576174, 0.03546112060546875, 0.035471359252929685, 0.035681278228759765, 0.035757057189941405, 0.03548160171508789, 0.035492862701416016, 0.035465217590332034, 0.035490814208984374, 0.03547750473022461, 0.03547955322265625, 0.0354856948852539, 0.03551129531860352, 0.03549900817871094, 0.035530750274658206, 0.03552972793579102, 0.035530750274658206, 0.03550822448730469, 0.03552870559692383, 0.03551846313476562, 0.03551232147216797, 0.035509342193603514, 0.03556342315673828, 0.03555737686157227, 0.11177881622314453, 0.035323902130126955, 0.035329025268554685, 0.03532287979125977, 0.03532799911499023, 0.035310592651367184, 0.035351551055908204, 0.03531980895996094, 0.035315711975097655, 0.03532799911499023, 0.03532495880126953, 0.03535971069335937, 0.0353361930847168, 0.035345409393310545, 0.03534643173217773, 0.03533939361572266, 0.035348350524902344, 0.03536383819580078, 0.035372222900390625, 0.035373886108398436, 0.03535683059692383, 0.035355487823486326, 0.03538739013671875, 0.035383296966552735, 0.03541299057006836, 0.035410945892333984, 0.03542124938964844, 0.03545695877075195, 0.0354969596862793, 0.035547134399414065, 0.035409984588623045, 0.03544672012329102, 0.03543552017211914, 0.035432449340820314, 0.03543142318725586, 0.03544678497314453, 0.03544575881958008, 0.03544678497314453, 0.03548876953125, 0.035489791870117186, 0.03546214294433594, 0.035476478576660156, 0.03546419143676758, 0.03548160171508789, 0.03549388885498047, 0.035484672546386715, 0.035520511627197264, 0.035509246826171875, 0.03551641464233399, 0.035487743377685545, 0.035492862701416016, 0.035490814208984374, 0.035525630950927735, 0.03552358245849609, 0.03560755157470703, 0.035604480743408204, 0.035576831817626955, 0.035566593170166014, 0.03554816055297851, 0.035544063568115236, 0.035533824920654294, 0.03556454467773437, 0.035552257537841796, 0.11175730895996094, 0.03542937469482422, 0.03536793518066406, 0.035353599548339845, 0.035342334747314456, 0.03535769653320313, 0.035366912841796876, 0.0354252815246582, 0.035372032165527346, 0.03535776138305664, 0.035344318389892576, 0.03535871887207031, 0.03539251327514648, 0.035381248474121094, 0.03542425537109375, 0.035386497497558594, 0.03537907028198242, 0.035386367797851564, 0.03543040084838867, 0.03543654251098633, 0.03545907211303711, 0.035434497833251956, 0.03545910263061523, 0.035517406463623044, 0.03542732620239258, 0.03544268798828125, 0.03543756866455078, 0.03542937469482422, 0.03542835235595703, 0.03545292663574219, 0.0354150390625, 0.03562499237060547, 0.03566080093383789, 0.03545596694946289, 0.03544985580444336, 0.03545804977416992, 0.03545395278930664, 0.03546419143676758, 0.03546112060546875, 0.03546623992919922, 0.03545907211303711, 0.03548160171508789, 0.03545395278930664, 0.03550620651245117, 0.03549795150756836, 0.03553484725952148, 0.035530750274658206, 0.03555635070800781, 0.03548160171508789, 0.03550515365600586, 0.035490814208984374, 0.035533824920654294, 0.03554304122924805, 0.03554099273681641, 0.03551129531860352, 0.0355491828918457, 0.03552153778076172, 0.0355676155090332, 0.035593215942382815, 0.0355491828918457, 0.03562803268432617, 0.035606529235839846, 0.03554611206054688, 0.11178189086914063, 0.03534745788574219, 0.0353546257019043, 0.03534131240844727, 0.03535871887207031, 0.035373054504394534, 0.035383296966552735, 0.03536383819580078, 0.035378177642822264, 0.0354252815246582, 0.035519489288330076, 0.03571916961669922, 0.03576422500610352, 0.03552870559692383, 0.03540991973876953, 0.03536076736450195, 0.03539763259887695, 0.03536899185180664, 0.035409889221191405, 0.035404800415039066, 0.03538643264770508, 0.03541600036621094, 0.03538022232055664, 0.03539763259887695, 0.035369983673095705, 0.03540787124633789, 0.0354334716796875, 0.03541299057006836, 0.03540991973876953, 0.03540070343017578, 0.03540377426147461, 0.035422206878662106, 0.035410945892333984, 0.035460094451904296, 0.035471359252929685, 0.03543654251098633, 0.03544473648071289, 0.03546726226806641, 0.0355153923034668, 0.03548160171508789, 0.03549491119384766, 0.03548160171508789, 0.035454975128173825, 0.03548364639282227, 0.03545804977416992, 0.03546828842163086, 0.035460094451904296, 0.03545702362060547, 0.035501056671142575, 0.0354856948852539, 0.035492862701416016, 0.03550310516357422, 0.03550003051757813, 0.03550207901000976, 0.03549798583984375, 0.03550207901000976, 0.03552460861206055, 0.03552166366577148, 0.03552560043334961, 0.035523487091064454, 0.03551232147216797, 0.035522560119628906, 0.035530750274658206, 0.11168358612060547, 0.035366912841796876, 0.035364864349365234, 0.03538227081298828, 0.03534438323974609, 0.0353331184387207, 0.035348480224609374, 0.035350528717041016, 0.03536896133422852, 0.0353546257019043, 0.035351551055908204, 0.035348575592041014, 0.035354526519775394, 0.03535871887207031, 0.0354068489074707, 0.035386367797851564, 0.03538022232055664, 0.035366912841796876, 0.03537510299682617, 0.035399681091308595, 0.03538227081298828, 0.03540991973876953, 0.035381248474121094, 0.03537919998168945, 0.03537510299682617, 0.03539558410644531, 0.03537408065795898, 0.03540787124633789, 0.035454975128173825, 0.03545804977416992, 0.03544575881958008, 0.03542118453979492, 0.03540787124633789, 0.03541299057006836, 0.03546112060546875, 0.035448833465576174, 0.03543040084838867, 0.03547340774536133, 0.03543552017211914, 0.0354334716796875, 0.03545600128173828, 0.035477569580078125, 0.035469249725341795, 0.03548672103881836, 0.03545395278930664, 0.035460094451904296, 0.03547750473022461, 0.03549388885498047, 0.03545395278930664, 0.03549798583984375, 0.03548057556152344, 0.035487743377685545, 0.0355153923034668, 0.03549388885498047, 0.03563212966918945, 0.035939327239990236, 0.03558911895751953, 0.03555942535400391, 0.035661823272705076, 0.03564441680908203, 0.03552972793579102, 0.03553488159179687, 0.035533790588378904, 0.11241884613037109, 0.03542524719238281, 0.03532492828369141, 0.03533942413330078, 0.03540259170532226, 0.03541401672363281, 0.03534643173217773, 0.03531468963623047, 0.035315711975097655, 0.03536896133422852, 0.03537100982666016, 0.03540172958374024, 0.035366912841796876, 0.03535257720947266, 0.035386367797851564, 0.035399681091308595, 0.035451904296875, 0.03539251327514648, 0.03542937469482422, 0.03537612915039062, 0.03539763259887695, 0.03542732620239258, 0.035388416290283206, 0.03540787124633789, 0.035386367797851564, 0.03541196823120117, 0.03544268798828125, 0.0353966064453125, 0.03540172958374024, 0.035451904296875, 0.0354334716796875, 0.03543654251098633, 0.03544678497314453, 0.03542835235595703, 0.03542940902709961, 0.0354365119934082, 0.03543040084838867, 0.035448833465576174, 0.035448833465576174, 0.03546726226806641, 0.0354376335144043, 0.03546310424804688, 0.03543961715698242, 0.035465217590332034, 0.03547750473022461, 0.03547238540649414, 0.03550515365600586, 0.03549593734741211, 0.03547750473022461, 0.03561471939086914, 0.03587071990966797, 0.035509246826171875, 0.03549388885498047, 0.035530750274658206, 0.03550003051757813, 0.035530750274658206, 0.03552665710449219, 0.03551846313476562, 0.03552774429321289, 0.03553068923950195, 0.03551129531860352, 0.03553792190551758, 0.035519550323486325]",tokens/s,27.356528874174575,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/gpt-neo-2.7B,EleutherAI/gpt-neo-2.7B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1209.638912,6568.804352,0.0,5922.357248,5577.085952,s,10,0.3505566749572754,0.03505566749572754,0.00044261111135194864,0.03490660858154297,0.035125981521606446,0.03575150356292724,0.03625192119598389,"[0.03637702560424805, 0.03498697662353516, 0.034847423553466796, 0.03491574478149414, 0.0348974723815918, 0.03497321701049805, 0.034848159790039065, 0.03488582229614258, 0.03489475250244141, 0.03493008041381836]",tokens/s,7302.67081724233,kWh,4.1487517993584944e-07,2.2733230058005766e-07,2.310153894905208e-06,2.9523613754211156e-06,tokens/kWh,86710252.38686608,MB,1209.966592,6568.804352,0.0,5922.357248,5675.07712,s,10,11.677966552734375,1.1677966552734376,0.018037859839655508,1.1660834350585936,1.1882947265625,1.1982243896484375,1.2061681201171874,"[1.208154052734375, 1.147835205078125, 1.1490413818359375, 1.186088134765625, 1.1722734375, 1.1738004150390624, 1.158863037109375, 1.1497440185546874, 1.1721026611328125, 1.160064208984375]",tokens/s,53.947748279214466,kWh,1.3845437320991337e-05,7.5869739852925325e-06,5.359797416489541e-05,7.503038547117925e-05,tokens/kWh,839659.820542967,,s,629,11.972358205795292,0.01903395581207518,0.004195532367416405,0.018300928115844727,0.018993766784667968,0.019183615875244142,0.05357010757446289,"[0.019327999114990235, 0.019203071594238282, 0.01926348876953125, 0.019211263656616212, 0.019227712631225587, 0.019105728149414063, 0.019139583587646485, 0.02003558349609375, 0.019430400848388672, 0.019198976516723632, 0.01926655960083008, 0.01920512008666992, 0.019123199462890626, 0.01904332733154297, 0.01906483268737793, 0.019168256759643554, 0.01900339126586914, 0.019014656066894533, 0.01899622344970703, 0.01903718376159668, 0.018985984802246093, 0.018969600677490234, 0.018915327072143554, 0.018969600677490234, 0.018550783157348632, 0.018328575134277342, 0.019779584884643556, 0.019551231384277345, 0.02165350341796875, 0.02248192024230957, 0.019458047866821288, 0.0192225284576416, 0.01919385528564453, 0.018968576431274413, 0.018970624923706055, 0.018980863571166993, 0.01900748825073242, 0.018905088424682616, 0.018925567626953126, 0.018899967193603515, 0.01882828712463379, 0.018818111419677735, 0.01887225532531738, 0.0188723201751709, 0.01883852767944336, 0.01882111930847168, 0.01883647918701172, 0.01962291145324707, 0.01902796745300293, 0.019104768753051758, 0.018655231475830078, 0.019042303085327148, 0.019153919219970703, 0.01906380844116211, 0.018954303741455077, 0.019004383087158204, 0.01908016014099121, 0.018942975997924806, 0.018957311630249024, 0.018900991439819336, 0.018898944854736328, 0.018947071075439453, 0.05368012619018555, 0.01823744010925293, 0.018373632431030275, 0.01841049575805664, 0.01824051284790039, 0.01821388816833496, 0.018158592224121094, 0.01820979118347168, 0.018136064529418947, 0.018174976348876954, 0.018149375915527344, 0.018157567977905274, 0.01815449523925781, 0.018354175567626953, 0.018378751754760742, 0.01824870491027832, 0.018148351669311523, 0.01820057678222656, 0.018104320526123048, 0.018182144165039063, 0.018149375915527344, 0.018174976348876954, 0.01820159912109375, 0.01819647979736328, 0.018264064788818358, 0.018009088516235353, 0.018339839935302735, 0.018176000595092775, 0.01819545555114746, 0.018382848739624022, 0.018226175308227538, 0.018191360473632814, 0.018147392272949217, 0.018232255935668944, 0.01820262336730957, 0.018229248046875, 0.01816166305541992, 0.01818726348876953, 0.018134016036987305, 0.018152511596679688, 0.018181055068969727, 0.01823744010925293, 0.0182108154296875, 0.018190336227416993, 0.01827840042114258, 0.01820467185974121, 0.018267135620117187, 0.01821286392211914, 0.018241535186767577, 0.018184192657470705, 0.018249727249145507, 0.018192384719848635, 0.01825382423400879, 0.018233343124389647, 0.018223104476928712, 0.01818009567260742, 0.01819647979736328, 0.01821388816833496, 0.018197504043579102, 0.01830611228942871, 0.018174943923950197, 0.018188255310058594, 0.018189311981201172, 0.053553150177001956, 0.01817091178894043, 0.018212831497192383, 0.018242591857910155, 0.018184160232543944, 0.018197504043579102, 0.018102272033691406, 0.01819647979736328, 0.018054208755493163, 0.018260927200317384, 0.018140159606933593, 0.0181790714263916, 0.018194496154785158, 0.01814316749572754, 0.018156543731689453, 0.0181790714263916, 0.018272256851196288, 0.01814630317687988, 0.018301952362060548, 0.01828556823730469, 0.01821388816833496, 0.01820876884460449, 0.018168832778930662, 0.018300928115844727, 0.018172927856445312, 0.018199552536010744, 0.01819241523742676, 0.01826812744140625, 0.018231296539306642, 0.01839308738708496, 0.0183767032623291, 0.018310144424438478, 0.018315263748168945, 0.01827737617492676, 0.01820876884460449, 0.01840742492675781, 0.018318336486816408, 0.01820876884460449, 0.01818623924255371, 0.018273279190063475, 0.01818828773498535, 0.018225151062011717, 0.018229248046875, 0.01820364761352539, 0.01820569610595703, 0.018177024841308592, 0.018214912414550782, 0.018552831649780274, 0.018426912307739258, 0.01825686454772949, 0.018248735427856447, 0.018186208724975585, 0.018249727249145507, 0.01825279998779297, 0.018182144165039063, 0.018177024841308592, 0.018176000595092775, 0.018266111373901366, 0.018280448913574218, 0.018199552536010744, 0.018189311981201172, 0.01817190361022949, 0.018255872726440428, 0.05360639953613281, 0.01820262336730957, 0.018182144165039063, 0.01861529541015625, 0.018779136657714843, 0.018762752532958983, 0.01884160041809082, 0.018817024230957033, 0.018737152099609376, 0.018769920349121092, 0.01882931137084961, 0.018714624404907225, 0.018771968841552734, 0.01880575942993164, 0.018771968841552734, 0.01881395149230957, 0.018760704040527345, 0.018770944595336913, 0.018763807296752928, 0.018777055740356444, 0.018737152099609376, 0.01884160041809082, 0.01887948799133301, 0.018568191528320312, 0.01879859161376953, 0.01884774398803711, 0.018799615859985352, 0.018766847610473633, 0.01881395149230957, 0.018865152359008788, 0.018765823364257812, 0.018893823623657227, 0.01880985641479492, 0.018926591873168946, 0.01886617660522461, 0.01884569549560547, 0.01880678367614746, 0.018747392654418944, 0.018738176345825194, 0.018815999984741212, 0.018898944854736328, 0.018770944595336913, 0.018791423797607423, 0.01887948799133301, 0.01903923225402832, 0.01884774398803711, 0.01884569549560547, 0.018930688858032226, 0.018893823623657227, 0.01880985641479492, 0.019083263397216797, 0.018896896362304686, 0.019604480743408204, 0.019345407485961915, 0.019106815338134766, 0.019056640625, 0.018977792739868164, 0.01881804847717285, 0.01884364891052246, 0.01880678367614746, 0.018902015686035157, 0.018803712844848632, 0.018840576171875, 0.05357670211791992, 0.018185216903686522, 0.018207744598388673, 0.018168832778930662, 0.018167808532714845, 0.0181463680267334, 0.01815750312805176, 0.01827020835876465, 0.018152448654174806, 0.018153472900390624, 0.018166784286499024, 0.01822822380065918, 0.017999872207641602, 0.01799679946899414, 0.018120704650878908, 0.018173952102661133, 0.01822208023071289, 0.018250751495361327, 0.018142208099365235, 0.01818623924255371, 0.018160640716552736, 0.0182609920501709, 0.01815555191040039, 0.018171871185302734, 0.018638879776000976, 0.01891836738586426, 0.01884979248046875, 0.01883135986328125, 0.018840576171875, 0.019160064697265625, 0.018884607315063476, 0.018961408615112304, 0.01880985641479492, 0.019095552444458007, 0.018930688858032226, 0.018909183502197266, 0.01880985641479492, 0.01882521629333496, 0.01880575942993164, 0.01876479911804199, 0.018832384109497072, 0.01884569549560547, 0.018963455200195312, 0.018902015686035157, 0.01886310386657715, 0.01879654312133789, 0.018993152618408202, 0.01886720085144043, 0.018662399291992187, 0.018522111892700196, 0.018785280227661134, 0.01888768005371094, 0.018883583068847656, 0.018908159255981445, 0.018896896362304686, 0.018906112670898437, 0.018884607315063476, 0.018917375564575196, 0.018925567626953126, 0.018811904907226562, 0.018902015686035157, 0.018865152359008788, 0.018908159255981445, 0.05421875381469726, 0.018626560211181642, 0.018890752792358398, 0.01895427131652832, 0.01926243209838867, 0.018911231994628908, 0.018777088165283205, 0.01882624053955078, 0.01824460792541504, 0.018117631912231445, 0.018125823974609375, 0.018177024841308592, 0.018280448913574218, 0.018126848220825196, 0.018177024841308592, 0.018259967803955078, 0.018157567977905274, 0.018199552536010744, 0.018173952102661133, 0.01821388816833496, 0.01820467185974121, 0.018168832778930662, 0.018149375915527344, 0.018249727249145507, 0.018266111373901366, 0.01816268730163574, 0.018536447525024414, 0.01889587211608887, 0.01879347229003906, 0.01882521629333496, 0.018811904907226562, 0.018752511978149415, 0.01887027168273926, 0.01883852767944336, 0.018775039672851563, 0.018832384109497072, 0.018899967193603515, 0.018975744247436522, 0.018807807922363282, 0.018869247436523438, 0.018784255981445314, 0.018919424057006837, 0.01886720085144043, 0.018840576171875, 0.01883647918701172, 0.018840576171875, 0.01887027168273926, 0.018886655807495118, 0.018981887817382814, 0.018927616119384767, 0.018881536483764647, 0.018876415252685547, 0.018964479446411133, 0.018920448303222655, 0.018939903259277344, 0.01881088066101074, 0.018868223190307617, 0.018874368667602538, 0.01881497573852539, 0.018129919052124025, 0.01827123260498047, 0.01814531135559082, 0.018183135986328126, 0.05347532653808594, 0.01821183967590332, 0.018144256591796876, 0.018199552536010744, 0.018157567977905274, 0.018145280838012694, 0.018059263229370116, 0.018084863662719726, 0.017967103958129883, 0.018067487716674803, 0.01826812744140625, 0.018226175308227538, 0.018118656158447266, 0.018163711547851562, 0.018157567977905274, 0.018151424407958985, 0.018115583419799804, 0.018121728897094725, 0.018106367111206053, 0.018135040283203126, 0.01863884735107422, 0.01886518478393555, 0.018824159622192384, 0.018974720001220705, 0.018918399810791017, 0.018868223190307617, 0.01882009506225586, 0.018913280487060546, 0.0189040641784668, 0.018928672790527342, 0.018789344787597657, 0.018885631561279297, 0.01880473518371582, 0.01906585693359375, 0.018952192306518553, 0.018888704299926756, 0.018851839065551757, 0.01876479911804199, 0.018951168060302736, 0.0188590087890625, 0.01819647979736328, 0.018132991790771484, 0.018160640716552736, 0.01823744010925293, 0.018145280838012694, 0.018190336227416993, 0.01815449523925781, 0.018198528289794923, 0.018199552536010744, 0.018123775482177733, 0.01823232078552246, 0.018166784286499024, 0.01822719955444336, 0.018225151062011717, 0.01820569610595703, 0.01815763282775879, 0.01812575912475586, 0.01816985511779785, 0.018150400161743165, 0.018259967803955078, 0.01818726348876953, 0.01813408088684082, 0.018194368362426758, 0.05366067123413086, 0.018185216903686522, 0.018207744598388673, 0.018160640716552736, 0.01823027229309082, 0.018117631912231445, 0.01827020835876465, 0.018319360733032225, 0.018158592224121094, 0.018143232345581056, 0.01820057678222656, 0.018173952102661133, 0.018128896713256838, 0.018241535186767577, 0.01816985511779785, 0.018183231353759766, 0.01820460891723633, 0.018189311981201172, 0.018135040283203126, 0.018156543731689453, 0.018206720352172853, 0.018132991790771484, 0.018223104476928712, 0.018143232345581056, 0.018250751495361327, 0.01823027229309082, 0.01829478454589844, 0.018155519485473632, 0.01820467185974121, 0.01823744010925293, 0.018290687561035156, 0.018335744857788085, 0.018136064529418947, 0.01824563217163086, 0.01822719955444336, 0.018239488601684572, 0.018166784286499024, 0.018173952102661133, 0.01821286392211914, 0.01823232078552246, 0.018229312896728515, 0.018246591567993162, 0.018234367370605468, 0.018215936660766603, 0.018351104736328124, 0.01820979118347168, 0.01824460792541504, 0.01825798416137695, 0.018214847564697264, 0.018189311981201172, 0.01828659248352051, 0.018207744598388673, 0.01820467185974121, 0.018312192916870116, 0.018217983245849608, 0.0181790714263916, 0.018581504821777343, 0.018970624923706055, 0.018490367889404297, 0.018447359085083007, 0.018305023193359374, 0.018320383071899413, 0.018185216903686522, 0.05359308624267578, 0.01820159912109375, 0.01816268730163574, 0.018241535186767577, 0.018059263229370116, 0.017971200942993162, 0.01810534477233887, 0.0182609920501709, 0.018121728897094725, 0.018181119918823242, 0.018164735794067383, 0.018361343383789062, 0.01822105598449707, 0.018241535186767577, 0.01828659248352051, 0.01823744010925293, 0.01822208023071289, 0.01821388816833496, 0.018329599380493163, 0.01822412872314453, 0.018292736053466797, 0.018207744598388673, 0.018198528289794923, 0.018214912414550782, 0.01823744010925293, 0.01818115234375, 0.01824867248535156, 0.0182108154296875, 0.018739200592041014, 0.018925567626953126, 0.0188723201751709, 0.018792448043823243, 0.01880268859863281, 0.018826303482055665, 0.018812864303588868, 0.018873344421386717, 0.01883852767944336, 0.019351615905761718, 0.019053504943847655, 0.018876415252685547, 0.018792448043823243, 0.019075071334838867, 0.01885491180419922, 0.018800640106201173, 0.01884671974182129, 0.01885593605041504, 0.01880575942993164, 0.01882009506225586, 0.019090431213378906, 0.01940275192260742, 0.019088384628295898, 0.01887129592895508, 0.01902387237548828, 0.018933759689331055, 0.018912256240844725, 0.018808832168579103, 0.018922496795654296, 0.01884671974182129, 0.01885081672668457, 0.0188590087890625, 0.018868223190307617, 0.01886310386657715, 0.01887846374511719, 0.053754878997802735, 0.018142208099365235, 0.01826304054260254, 0.018185216903686522, 0.01820876884460449, 0.01819647979736328, 0.018239488601684572, 0.018229248046875, 0.018163711547851562, 0.01829478454589844, 0.018212896347045898, 0.018233312606811523, 0.01822822380065918, 0.018257919311523436, 0.018272256851196288, 0.018267135620117187, 0.018198528289794923, 0.01821183967590332, 0.01822105598449707, 0.01827840042114258, 0.018197504043579102, 0.01822412872314453, 0.01820364761352539, 0.018167808532714845, 0.018239488601684572, 0.018300928115844727, 0.01824665641784668, 0.018177024841308592, 0.018148351669311523, 0.01819443130493164, 0.01819647979736328, 0.01823539161682129, 0.018214912414550782, 0.01823744010925293, 0.018198528289794923, 0.018207744598388673, 0.018197504043579102, 0.01826304054260254, 0.018361343383789062, 0.018298879623413086, 0.018288639068603514, 0.01825484848022461, 0.018502656936645507, 0.01823641586303711, 0.018172927856445312, 0.018335744857788085, 0.018948095321655273, 0.018919424057006837, 0.018554880142211915, 0.018637823104858398, 0.018960384368896483, 0.01885491180419922, 0.01887948799133301, 0.018856960296630858, 0.018949119567871094, 0.01886207962036133, 0.01899830436706543, 0.01885385513305664, 0.018889728546142577, 0.01887846374511719, 0.018880512237548826, 0.018851839065551757, 0.018906175613403322]",tokens/s,52.53768632611819,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,1,1,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/1/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66932615-38c04a4254bdbf08664913ba;a2fcbc9d-4860-4e15-87c8-428e3c740c3f) Repository Not Found for url: https://huggingface.co/1/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: 1 is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-7b,tiiuae/falcon-7b,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ self.load_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 250, in load_model_with_no_weights self.pretrained_model = self.automodel_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 559, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3704, in from_pretrained config = cls._autoset_attn_implementation( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1481, in _autoset_attn_implementation cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1572, in _check_and_enable_flash_attn_2 raise ValueError( ValueError: FalconForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmp70sbh58x/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,meta-llama/Llama-2-7b-hf,meta-llama/Llama-2-7b-hf,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.1,,,,MB,1229.29152,14563.147776,0.0,13916.700672,13738.740736,s,10,0.9365172424316406,0.09365172424316406,0.0003831550521236713,0.09352369689941406,0.09386207656860351,0.09430012550354004,0.09465056465148926,"[0.09473817443847657, 0.09376473236083985, 0.09349996948242187, 0.0934306869506836, 0.0936256332397461, 0.0933544921875, 0.09341878509521484, 0.09370111846923829, 0.09343622589111328, 0.09354742431640625]",tokens/s,2733.5321593791823,kWh,1.1075711711421068e-06,6.06895488795727e-07,6.176783238411304e-06,7.89124989834914e-06,tokens/kWh,32440995.190578815,MB,1229.29152,14563.147776,0.0,13916.700672,13889.779712,s,10,22.609111328125,2.2609111328125,0.003172867868557072,2.2603369140625,2.2642384277343752,2.2664495361328125,2.2682184228515623,"[2.26866064453125, 2.261553466796875, 2.2599931640625, 2.2606806640625, 2.2575, 2.2637470703125, 2.259510986328125, 2.259595947265625, 2.2569189453125, 2.260950439453125]",tokens/s,27.864872301119618,kWh,2.660855508868816e-05,1.4582375692265763e-05,0.00011762609358158787,0.00015881702436254178,tokens/kWh,396682.9138933234,,s,629,23.414120468139664,0.037224356865086884,0.011184660459204559,0.03586969757080078,0.03600281524658203,0.036149658203125,0.12989068054199218,"[0.03647078323364258, 0.036483070373535154, 0.0363694076538086, 0.03647385787963867, 0.036511745452880856, 0.036380672454833986, 0.036311038970947264, 0.03642675018310547, 0.036318206787109376, 0.03628441619873047, 0.03628851318359375, 0.03620044708251953, 0.036380672454833986, 0.03591884613037109, 0.03579596710205078, 0.035778560638427735, 0.035829761505126956, 0.035724288940429685, 0.03582463836669922, 0.035800064086914066, 0.03583180618286133, 0.035796993255615236, 0.03595161437988281, 0.035829761505126956, 0.03587891387939453, 0.035827713012695314, 0.03586764907836914, 0.03584819030761719, 0.035885055541992186, 0.035817470550537106, 0.03578572845458984, 0.03580108642578125, 0.035860481262207033, 0.03581644821166992, 0.035833854675292966, 0.035869728088378905, 0.036013023376464844, 0.03584307098388672, 0.03584307098388672, 0.03581955337524414, 0.035857376098632814, 0.035844097137451174, 0.0359106559753418, 0.03596492767333984, 0.03592192077636719, 0.03591372680664062, 0.03595673751831055, 0.03588608169555664, 0.03590860748291016, 0.0358809585571289, 0.0362690544128418, 0.035980289459228515, 0.035975166320800785, 0.03589017486572266, 0.035922943115234376, 0.03596492767333984, 0.03599359893798828, 0.03597721481323242, 0.03604889678955078, 0.03594956970214844, 0.03599052810668945, 0.035950592041015625, 0.13054464721679687, 0.03573763275146485, 0.03571196746826172, 0.035732479095458985, 0.03571916961669922, 0.03576934432983398, 0.03575296020507813, 0.03574169540405273, 0.035814399719238284, 0.03579391860961914, 0.0358389778137207, 0.035804191589355466, 0.035837921142578125, 0.035850238800048825, 0.03586457443237305, 0.03586054229736328, 0.03586553573608398, 0.035827713012695314, 0.03580416107177734, 0.03578879928588867, 0.03584102249145508, 0.03586150360107422, 0.03589120101928711, 0.03589427185058594, 0.03599359893798828, 0.036275199890136715, 0.036157440185546875, 0.03592499160766602, 0.03583283233642578, 0.03583692932128906, 0.035855358123779296, 0.03587276840209961, 0.035884033203125, 0.035866622924804685, 0.03589734268188476, 0.03585638427734375, 0.03589324951171875, 0.03587891387939453, 0.035888126373291016, 0.035904544830322266, 0.03588911819458008, 0.035888126373291016, 0.035932159423828124, 0.03587481689453125, 0.03589324951171875, 0.0359106559753418, 0.035947521209716796, 0.035912704467773435, 0.03593523025512695, 0.03592192077636719, 0.03593523025512695, 0.03594956970214844, 0.03596492767333984, 0.035963905334472655, 0.035945472717285154, 0.03590041732788086, 0.035931137084960936, 0.03593523025512695, 0.035947521209716796, 0.03593830490112305, 0.03595161437988281, 0.03593625640869141, 0.03601203155517578, 0.12995277404785155, 0.03571200180053711, 0.03569049453735352, 0.035691520690917966, 0.0358205451965332, 0.03574169540405273, 0.035719200134277346, 0.035788768768310546, 0.03573660659790039, 0.03571094512939453, 0.035726398468017576, 0.03572320175170898, 0.0358021125793457, 0.03582566452026367, 0.03578265762329102, 0.03579596710205078, 0.03585945510864258, 0.035866622924804685, 0.03590758514404297, 0.03583795166015625, 0.03580825424194336, 0.03582880020141602, 0.035839935302734376, 0.03584819030761719, 0.03590348815917969, 0.03579084777832031, 0.03583078384399414, 0.03581542587280274, 0.03580928039550781, 0.03582361602783203, 0.035939327239990236, 0.03591372680664062, 0.03597107315063477, 0.03589017486572266, 0.035896320343017575, 0.035850238800048825, 0.03586870574951172, 0.03588399887084961, 0.03588608169555664, 0.03590963363647461, 0.03587583923339844, 0.035899391174316404, 0.03587587356567383, 0.0358768310546875, 0.03589324951171875, 0.035901439666748046, 0.035920894622802735, 0.03590655899047852, 0.03593523025512695, 0.03591372680664062, 0.03594137573242188, 0.03593523025512695, 0.0359659538269043, 0.035942401885986325, 0.03592704010009766, 0.035942401885986325, 0.03597312164306641, 0.0359106559753418, 0.03591372680664062, 0.035985408782958986, 0.03607756805419922, 0.03601510238647461, 0.03595468902587891, 0.12990675354003905, 0.035695552825927734, 0.03566387176513672, 0.03580416107177734, 0.035775489807128906, 0.035716094970703126, 0.035710975646972655, 0.035732479095458985, 0.03572019195556641, 0.035705856323242184, 0.035727359771728515, 0.035743743896484374, 0.03579904174804688, 0.03573964691162109, 0.0357498893737793, 0.03574784088134766, 0.0357314567565918, 0.0357498893737793, 0.03579395294189453, 0.03577446365356445, 0.035773406982421874, 0.03599155044555664, 0.03611443328857422, 0.035833854675292966, 0.03580928039550781, 0.03583078384399414, 0.03580928039550781, 0.03580825424194336, 0.03583692932128906, 0.035800064086914066, 0.03587481689453125, 0.036178943634033206, 0.03601919937133789, 0.0358809585571289, 0.03584204864501953, 0.0358922233581543, 0.03588608169555664, 0.03589734268188476, 0.03587481689453125, 0.035896320343017575, 0.03599257659912109, 0.03589120101928711, 0.03592192077636719, 0.03593011093139648, 0.035947521209716796, 0.035969024658203126, 0.03593523025512695, 0.035910686492919924, 0.03590140914916992, 0.03603251266479492, 0.03600588989257812, 0.035937278747558594, 0.03593625640869141, 0.035988479614257815, 0.03592601776123047, 0.035963905334472655, 0.03609702301025391, 0.03599359893798828, 0.03597107315063477, 0.036021247863769534, 0.03594137573242188, 0.0359628791809082, 0.03604172897338867, 0.13021798706054688, 0.0357130241394043, 0.03567411041259766, 0.035659774780273434, 0.03568230438232422, 0.035691520690917966, 0.035678207397460936, 0.03567001724243164, 0.03571916961669922, 0.035692543029785154, 0.03569356918334961, 0.03573452758789063, 0.03574784088134766, 0.03575091171264649, 0.03572531127929687, 0.035697662353515625, 0.035683326721191407, 0.03580723190307617, 0.03569465637207031, 0.03572012710571289, 0.03569868850708008, 0.03575500869750976, 0.03575603103637695, 0.03580416107177734, 0.03570892715454101, 0.03577753448486328, 0.03576422500610352, 0.035781631469726564, 0.03577241516113281, 0.035794944763183595, 0.03579904174804688, 0.035811328887939455, 0.03580928039550781, 0.035833854675292966, 0.03587583923339844, 0.03580416107177734, 0.035775489807128906, 0.03581542587280274, 0.035947521209716796, 0.03588710403442383, 0.03584204864501953, 0.03597824096679687, 0.03592601776123047, 0.03594137573242188, 0.03587075042724609, 0.035931102752685545, 0.035922943115234376, 0.03594649505615234, 0.03585331344604492, 0.03591372680664062, 0.03597107315063477, 0.035952640533447267, 0.035855358123779296, 0.035889217376708984, 0.0359884147644043, 0.03595775985717774, 0.03590553665161133, 0.03595775985717774, 0.03594137573242188, 0.035988479614257815, 0.035963905334472655, 0.03599359893798828, 0.03613798522949219, 0.12984934997558595, 0.03579391860961914, 0.035789825439453124, 0.035800064086914066, 0.03576115036010742, 0.035757057189941405, 0.03580108642578125, 0.03582668685913086, 0.035827713012695314, 0.036446208953857424, 0.03589734268188476, 0.03581849670410156, 0.03586969757080078, 0.03593523025512695, 0.03596291351318359, 0.03602735900878906, 0.03582668685913086, 0.03579391860961914, 0.035806209564208984, 0.03579084777832031, 0.03578777694702148, 0.03581644821166992, 0.035781631469726564, 0.03577958297729492, 0.03580518341064453, 0.03579904174804688, 0.035817470550537106, 0.03581644821166992, 0.03583795166015625, 0.0358656005859375, 0.03590963363647461, 0.03588710403442383, 0.03602329635620117, 0.03606528091430664, 0.035970046997070314, 0.03586457443237305, 0.035914752960205076, 0.03593830490112305, 0.03594137573242188, 0.035953662872314454, 0.0359659538269043, 0.036013057708740234, 0.03628851318359375, 0.0360816650390625, 0.03597619247436523, 0.03597926330566406, 0.035934207916259765, 0.035983360290527344, 0.03592192077636719, 0.03594854354858398, 0.03591987228393555, 0.03596799850463867, 0.03599052810668945, 0.03600281524658203, 0.035975166320800785, 0.03603251266479492, 0.03599257659912109, 0.03601510238647461, 0.03600588989257812, 0.036111358642578126, 0.03600281524658203, 0.03604275131225586, 0.036036609649658206, 0.12991078186035157, 0.03579084777832031, 0.03583795166015625, 0.035783679962158206, 0.035746814727783204, 0.0358021125793457, 0.03577958297729492, 0.03573452758789063, 0.03574784088134766, 0.03579289627075195, 0.03574272155761719, 0.03575091171264649, 0.03575603103637695, 0.035727359771728515, 0.03577036666870117, 0.035829761505126956, 0.03576934432983398, 0.03576115036010742, 0.035798015594482424, 0.03583488082885742, 0.0358389778137207, 0.03585945510864258, 0.03578265762329102, 0.03604275131225586, 0.035899391174316404, 0.03582156753540039, 0.035806209564208984, 0.03585331344604492, 0.035800064086914066, 0.03582774353027344, 0.035929054260253904, 0.03584204864501953, 0.03582156753540039, 0.035844097137451174, 0.035833854675292966, 0.035860481262207033, 0.03587788772583008, 0.03587276840209961, 0.035862529754638675, 0.03587071990966797, 0.03584921646118164, 0.03586969757080078, 0.035929088592529294, 0.035866622924804685, 0.03587891387939453, 0.0358851203918457, 0.03589011383056641, 0.035901439666748046, 0.03594035339355469, 0.036003902435302736, 0.0359167366027832, 0.03589427185058594, 0.03590553665161133, 0.03595161437988281, 0.035952640533447267, 0.03590553665161133, 0.03595673751831055, 0.03591785430908203, 0.03597001647949219, 0.03592192077636719, 0.03595775985717774, 0.03595775985717774, 0.035939327239990236, 0.13015347290039062, 0.03570995330810547, 0.0356864013671875, 0.03569561767578125, 0.03570073699951172, 0.03574272155761719, 0.03575296020507813, 0.035692543029785154, 0.035716094970703126, 0.03573660659790039, 0.03574166488647461, 0.03574272155761719, 0.03578060913085938, 0.03581235122680664, 0.03578675079345703, 0.03576015853881836, 0.03575497436523437, 0.03577753448486328, 0.035773441314697264, 0.03580518341064453, 0.035840065002441406, 0.03578976058959961, 0.03580518341064453, 0.035783679962158206, 0.035784702301025394, 0.035789825439453124, 0.03581235122680664, 0.03583180618286133, 0.0358287353515625, 0.035811328887939455, 0.03581235122680664, 0.03582259368896484, 0.03582259368896484, 0.03585126495361328, 0.035888126373291016, 0.03584102249145508, 0.03584716796875, 0.03583590316772461, 0.03586355209350586, 0.03589427185058594, 0.03591884613037109, 0.0358922233581543, 0.03589017486572266, 0.03588198471069336, 0.03589529418945313, 0.03607961654663086, 0.03590457534790039, 0.036052928924560544, 0.03657830429077148, 0.03599462509155273, 0.03592192077636719, 0.03595577621459961, 0.03595872116088867, 0.03593830490112305, 0.03592806243896484, 0.03593523025512695, 0.03593523025512695, 0.03596492767333984, 0.03599871826171875, 0.03595980834960937, 0.03597312164306641, 0.035953662872314454, 0.035950592041015625, 0.12984524536132813, 0.035692543029785154, 0.03567103958129883, 0.0357212142944336, 0.0357212142944336, 0.0357386245727539, 0.03568742370605469, 0.035710975646972655, 0.035705856323242184, 0.035745792388916016, 0.0357212142944336, 0.035697662353515625, 0.03572633743286133, 0.03591987228393555, 0.035901439666748046, 0.03575094223022461, 0.03571811294555664, 0.0357325439453125, 0.03571807861328125, 0.03572633743286133, 0.03575398254394531, 0.03574784088134766, 0.035776512145996094, 0.035776512145996094, 0.035765247344970705, 0.03577139282226562, 0.035781631469726564, 0.03579391860961914, 0.03576531219482422, 0.035805118560791015, 0.035827713012695314, 0.03582668685913086, 0.0358205451965332, 0.03579904174804688, 0.03579804611206055, 0.03583894348144531, 0.03582668685913086, 0.0358205451965332, 0.03584716796875, 0.03590758514404297, 0.03586150360107422, 0.03584511947631836, 0.035912704467773435, 0.0358656005859375, 0.03585331344604492, 0.03587891387939453, 0.0358656005859375, 0.03589529418945313, 0.03586867141723633, 0.03587481689453125, 0.03588608169555664, 0.03589734268188476, 0.03594649505615234, 0.03591680145263672, 0.03589017486572266, 0.035884033203125, 0.03588198471069336, 0.03587788772583008, 0.035931137084960936, 0.035912704467773435, 0.035917823791503906, 0.03593523025512695, 0.03587583923339844, 0.13006745910644532, 0.03587891387939453, 0.03574172973632812, 0.03573244857788086, 0.03573452758789063, 0.035740673065185545, 0.03576115036010742, 0.035746814727783204, 0.035732479095458985, 0.03576627349853516, 0.03574169540405273, 0.03575398254394531, 0.03577040100097656, 0.035770336151123044, 0.03574476623535156, 0.03579084777832031, 0.03576115036010742, 0.035760128021240234, 0.035781631469726564, 0.03580723190307617, 0.035846145629882815, 0.03581951904296875, 0.03581235122680664, 0.03585126495361328, 0.03580931091308594, 0.03585327911376953, 0.035958782196044925, 0.03589023971557617, 0.035875774383544924, 0.03590348815917969, 0.03608575820922852, 0.036278270721435545, 0.03591680145263672, 0.03585126495361328, 0.03583795166015625, 0.035862529754638675, 0.03585228729248047, 0.035884033203125, 0.035846145629882815, 0.03587788772583008, 0.03585945510864258, 0.03587788772583008, 0.03585945510864258, 0.035937278747558594, 0.035879936218261715, 0.035884033203125, 0.035899391174316404, 0.03593830490112305, 0.03593427276611328, 0.035962814331054686, 0.03592704010009766, 0.03592192077636719, 0.035914752960205076, 0.035947521209716796, 0.035934207916259765, 0.035952640533447267, 0.0359444465637207, 0.03602841567993164, 0.03624556732177735, 0.036317119598388674, 0.035947521209716796, 0.03597107315063477, 0.0359628791809082]",tokens/s,26.864131021103297,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-rw-1b,tiiuae/falcon-rw-1b,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ self.load_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 250, in load_model_with_no_weights self.pretrained_model = self.automodel_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 559, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3704, in from_pretrained config = cls._autoset_attn_implementation( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1481, in _autoset_attn_implementation cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1572, in _check_and_enable_flash_attn_2 raise ValueError( ValueError: FalconForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpe1ji5rg6/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,8,8,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/8/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66932447-0b7d2c3e2535d40106640e55;b6bd33b1-056a-43b6-9104-5b9c16eab2ef) Repository Not Found for url: https://huggingface.co/8/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: 8 is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-4B,Qwen/Qwen1.5-4B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.3,,0.32.1,,,,1.21.1,,,,0.11.1,,,,MB,1293.668352,10131.8656,0.0,9485.418496,8972.092928,s,10,0.5807290840148925,0.05807290840148925,0.001402282591350963,0.05735932731628418,0.05929040069580078,0.06057876739501953,0.06160946075439453,"[0.06186713409423828, 0.057156158447265626, 0.0572591667175293, 0.05745948791503906, 0.0572275505065918, 0.05900409698486328, 0.0572108154296875, 0.057171840667724606, 0.057850048065185546, 0.05852278518676758]",tokens/s,4408.251748476834,kWh,6.789205051092657e-07,3.719078228184542e-07,3.5717269953103468e-06,4.6225553232380665e-06,tokens/kWh,55380624.37307378,MB,1293.668352,10131.8656,0.0,9485.418496,8972.095488,s,10,23.134972167968748,2.3134972167968746,0.008936641160474967,2.3165758056640624,2.3240864746093752,2.32569619140625,2.32698396484375,"[2.323728759765625, 2.305009521484375, 2.31708203125, 2.30321533203125, 2.318229248046875, 2.327305908203125, 2.303187255859375, 2.30162109375, 2.3195234375, 2.316069580078125]",tokens/s,27.231500233756886,kWh,2.7408579812119902e-05,1.501913217833268e-05,8.763642930448968e-05,0.00013006414129494224,tokens/kWh,484376.39592865907,,s,630,23.13293516921997,0.03671894471304757,0.0003843954034334146,0.03653529739379883,0.03714928474426269,0.03735096340179443,0.0377902173614502,"[0.03707494354248047, 0.03703705596923828, 0.03671859359741211, 0.03719372940063476, 0.03697356796264648, 0.03722956848144531, 0.03670220947265625, 0.036765697479248044, 0.037222400665283206, 0.03748147201538086, 0.037100543975830076, 0.03711897659301758, 0.0370513916015625, 0.036782081604003904, 0.037085182189941404, 0.03700838470458984, 0.03704115295410156, 0.036890625, 0.03667251205444336, 0.03627212905883789, 0.03643084716796875, 0.036279296875, 0.03624652862548828, 0.03629260635375976, 0.03644518280029297, 0.03641548919677735, 0.03723571014404297, 0.03710771179199219, 0.037026817321777344, 0.03808051300048828, 0.037269504547119144, 0.03695513534545898, 0.03703705596923828, 0.036770816802978515, 0.036947006225585936, 0.0377138557434082, 0.03731763076782227, 0.03707904052734375, 0.037043201446533204, 0.03703398513793945, 0.03708927917480469, 0.03689471817016601, 0.03676876831054687, 0.0369244155883789, 0.03713945770263672, 0.03703910446166992, 0.0370247688293457, 0.03692134475708008, 0.037176319122314457, 0.037703678131103514, 0.037294078826904296, 0.036923393249511716, 0.036393985748291016, 0.036296703338623046, 0.036316158294677735, 0.03652096176147461, 0.03650969696044922, 0.036378623962402344, 0.03640934371948242, 0.03635302352905274, 0.036465663909912106, 0.036393985748291016, 0.03645132827758789, 0.03628851318359375, 0.036162559509277346, 0.03625881576538086, 0.036327423095703124, 0.036410369873046876, 0.03633561706542969, 0.036365310668945314, 0.03639295959472656, 0.03640115356445312, 0.03697356796264648, 0.036980735778808595, 0.03698483276367188, 0.036944896697998046, 0.036913150787353514, 0.036939777374267575, 0.03743948745727539, 0.037085182189941404, 0.037190654754638675, 0.03704729461669922, 0.037012481689453126, 0.037372928619384765, 0.03662438583374023, 0.036395008087158204, 0.0363581428527832, 0.036435966491699216, 0.036380672454833986, 0.03644211196899414, 0.0363765754699707, 0.036087806701660154, 0.03633152008056641, 0.03638784027099609, 0.03639910507202149, 0.036465663909912106, 0.03637350463867187, 0.03625471878051758, 0.036211711883544925, 0.036383743286132815, 0.03634380722045898, 0.036421630859375, 0.03632332611083984, 0.036337665557861325, 0.036383743286132815, 0.03645644760131836, 0.03643801498413086, 0.03666636657714844, 0.03633049774169922, 0.03703398513793945, 0.03693670272827149, 0.036975616455078124, 0.03699814224243164, 0.03702783966064453, 0.03708108901977539, 0.036997119903564454, 0.03623526382446289, 0.03641548919677735, 0.03642060852050781, 0.03640729522705078, 0.03632844924926758, 0.036424705505371094, 0.03665203094482422, 0.036572158813476564, 0.03637247848510742, 0.036490238189697266, 0.03654860687255859, 0.03633561706542969, 0.03633663940429688, 0.03646771240234375, 0.036354049682617184, 0.0373135986328125, 0.03712812805175781, 0.03710976028442383, 0.03710566329956055, 0.036988929748535154, 0.03705036926269531, 0.03713740921020508, 0.03683737564086914, 0.03637964630126953, 0.03650969696044922, 0.03643392181396484, 0.03642879867553711, 0.03641446304321289, 0.036465663909912106, 0.036364288330078126, 0.03647590255737305, 0.03642675018310547, 0.03641548919677735, 0.03652096176147461, 0.03644211196899414, 0.036819969177246094, 0.03710259246826172, 0.03702169418334961, 0.03698483276367188, 0.037026817321777344, 0.0370882568359375, 0.036945919036865234, 0.03703398513793945, 0.03688857650756836, 0.0369541130065918, 0.03705548858642578, 0.037029888153076174, 0.036999168395996096, 0.037085182189941404, 0.0369879035949707, 0.036994049072265625, 0.03705241775512695, 0.036934654235839845, 0.036326400756835936, 0.03645951843261719, 0.03639091110229492, 0.03642367935180664, 0.036416511535644534, 0.036446208953857424, 0.03641241455078125, 0.03673702239990234, 0.03646361541748047, 0.03646361541748047, 0.03724288177490234, 0.03710259246826172, 0.036772865295410156, 0.036421630859375, 0.03654655838012695, 0.037648384094238284, 0.03733708953857422, 0.037182464599609374, 0.036972545623779295, 0.03712409591674805, 0.03647283172607422, 0.036365310668945314, 0.03618815994262695, 0.03621376037597656, 0.03645542526245117, 0.036299774169921875, 0.03640422439575195, 0.036408321380615234, 0.03637964630126953, 0.036441089630126954, 0.03679846572875976, 0.03642879867553711, 0.036452350616455076, 0.03635302352905274, 0.03634688186645508, 0.03645747375488281, 0.036550655364990234, 0.03643699264526367, 0.03643392181396484, 0.036410369873046876, 0.03639091110229492, 0.037466110229492186, 0.03674009704589844, 0.03653529739379883, 0.03646976089477539, 0.03657113647460938, 0.036367359161376955, 0.03639910507202149, 0.037269504547119144, 0.03694079971313476, 0.036511745452880856, 0.0362239990234375, 0.03630182266235352, 0.036334590911865236, 0.0364400634765625, 0.03638272094726563, 0.03635609436035156, 0.03631923294067383, 0.03653222274780273, 0.03642879867553711, 0.036408321380615234, 0.036329471588134765, 0.036375553131103515, 0.03640934371948242, 0.036342784881591796, 0.036316192626953125, 0.03646051025390625, 0.036318206787109376, 0.03715891265869141, 0.037084159851074217, 0.036689918518066404, 0.036485118865966795, 0.0363694076538086, 0.036380672454833986, 0.036757503509521484, 0.038344768524169924, 0.036758464813232423, 0.03632128143310547, 0.036173824310302735, 0.03758694458007812, 0.03671244812011719, 0.036942848205566405, 0.03700121688842774, 0.036429824829101565, 0.03649638366699219, 0.036803585052490234, 0.03705344009399414, 0.03641139221191406, 0.03647999954223633, 0.03646054458618164, 0.0364400634765625, 0.03703807830810547, 0.03701145553588867, 0.03699507141113281, 0.03748659133911133, 0.03642060852050781, 0.03668889617919922, 0.03706265640258789, 0.037154815673828126, 0.036790271759033204, 0.036713470458984376, 0.03707904052734375, 0.03687116622924805, 0.03644518280029297, 0.03644313430786133, 0.03643801498413086, 0.0365219841003418, 0.03645951843261719, 0.03638988876342773, 0.03647795104980469, 0.03639910507202149, 0.03628441619873047, 0.036400127410888675, 0.036446208953857424, 0.036808704376220705, 0.03711385726928711, 0.036959232330322264, 0.036983806610107424, 0.03683225631713867, 0.036432895660400394, 0.03634995269775391, 0.037144577026367184, 0.03746201705932617, 0.03717222213745117, 0.037064704895019535, 0.037477375030517575, 0.036982784271240236, 0.039373825073242184, 0.037351425170898435, 0.03712716674804688, 0.03650764846801758, 0.03684249496459961, 0.037943294525146484, 0.03743129730224609, 0.03695001602172852, 0.036465663909912106, 0.03646054458618164, 0.03657625579833984, 0.036490238189697266, 0.03652608108520508, 0.03642675018310547, 0.036466686248779294, 0.0365219841003418, 0.03635200119018555, 0.036340736389160154, 0.03649126434326172, 0.036997119903564454, 0.03756755065917969, 0.03702880096435547, 0.03693670272827149, 0.037176319122314457, 0.037082111358642575, 0.03676671981811523, 0.036408321380615234, 0.03645644760131836, 0.03638272094726563, 0.03662643051147461, 0.03705344009399414, 0.037000190734863284, 0.03697049713134765, 0.036992000579833983, 0.03702169418334961, 0.0363694076538086, 0.036395008087158204, 0.036370433807373044, 0.03684659194946289, 0.037054462432861326, 0.036959232330322264, 0.036569087982177735, 0.03656192016601562, 0.03703910446166992, 0.03668889617919922, 0.0370513916015625, 0.03701862335205078, 0.037698558807373043, 0.037353473663330077, 0.037106689453125, 0.036601856231689454, 0.03728998565673828, 0.03700940704345703, 0.03707289505004883, 0.03705651092529297, 0.03700735855102539, 0.0371486701965332, 0.03699507141113281, 0.03696230316162109, 0.03696025466918945, 0.03716403198242187, 0.036961280822753906, 0.03692031860351563, 0.036953086853027346, 0.03707596969604492, 0.03698995208740234, 0.036746238708496096, 0.03638476943969727, 0.03640729522705078, 0.037082111358642575, 0.03707699203491211, 0.03710771179199219, 0.03711078262329102, 0.037220352172851565, 0.03788083267211914, 0.037031936645507815, 0.03700326538085937, 0.037168128967285156, 0.03693056106567383, 0.03706982421875, 0.03670937728881836, 0.036446208953857424, 0.03648819351196289, 0.03643084716796875, 0.03645849609375, 0.036462593078613284, 0.0363765754699707, 0.03648819351196289, 0.03697868728637695, 0.03701760101318359, 0.03668070220947266, 0.03634483337402344, 0.03642265701293945, 0.0364400634765625, 0.036495361328125, 0.03631411361694336, 0.03647590255737305, 0.0363694076538086, 0.036416511535644534, 0.036495361328125, 0.03764633560180664, 0.037326847076416016, 0.03706880187988281, 0.036429824829101565, 0.0363694076538086, 0.0365291519165039, 0.036483070373535154, 0.03634380722045898, 0.03641958236694336, 0.03646156692504883, 0.03634995269775391, 0.03646771240234375, 0.036400127410888675, 0.03638169479370117, 0.03642675018310547, 0.03646054458618164, 0.03638886260986328, 0.03649638366699219, 0.03640934371948242, 0.03645644760131836, 0.03641241455078125, 0.03684044647216797, 0.03642060852050781, 0.03649945449829101, 0.03645439910888672, 0.03643801498413086, 0.03645542526245117, 0.037182464599609374, 0.03709030532836914, 0.0370513916015625, 0.036918270111083985, 0.03647078323364258, 0.036587520599365236, 0.036483070373535154, 0.036446208953857424, 0.03648614501953125, 0.036395008087158204, 0.036708351135253905, 0.03644416046142578, 0.03645644760131836, 0.03640627288818359, 0.036483070373535154, 0.03641446304321289, 0.03646156692504883, 0.036501502990722655, 0.03645542526245117, 0.036431873321533206, 0.03666227340698242, 0.03649228668212891, 0.036432895660400394, 0.036377601623535157, 0.0363397102355957, 0.036342784881591796, 0.03713228988647461, 0.037378047943115236, 0.03711180877685547, 0.03695820617675781, 0.03699302291870117, 0.036983806610107424, 0.03704012680053711, 0.036961280822753906, 0.036808704376220705, 0.03610726547241211, 0.03637452697753906, 0.036377601623535157, 0.03611852645874023, 0.036342784881591796, 0.03639091110229492, 0.03633363342285156, 0.03636115264892578, 0.036342784881591796, 0.036375553131103515, 0.036397056579589845, 0.036370433807373044, 0.036362239837646484, 0.03632332611083984, 0.0363612174987793, 0.03642265701293945, 0.036291584014892575, 0.03637350463867187, 0.03632128143310547, 0.037582847595214845, 0.03700735855102539, 0.03695718383789062, 0.037187583923339845, 0.036431873321533206, 0.036410369873046876, 0.036483070373535154, 0.036478977203369144, 0.0363581428527832, 0.03639708709716797, 0.03639094543457031, 0.03634272003173828, 0.03630387115478516, 0.036424705505371094, 0.03626291275024414, 0.03621785736083984, 0.03664896011352539, 0.036547584533691405, 0.03638272094726563, 0.03644518280029297, 0.03639091110229492, 0.036378623962402344, 0.03644927978515625, 0.03641139221191406, 0.03650969696044922, 0.036367359161376955, 0.03640524673461914, 0.03711897659301758, 0.03653222274780273, 0.03641548919677735, 0.0364400634765625, 0.03756851196289063, 0.037407745361328126, 0.03670732879638672, 0.03640627288818359, 0.03644313430786133, 0.037250049591064455, 0.03701964950561523, 0.03701356887817383, 0.03692639923095703, 0.036898815155029296, 0.03704217529296875, 0.03698995208740234, 0.036947967529296875, 0.03709235382080078, 0.03704934310913086, 0.036945919036865234, 0.037005313873291014, 0.03704524612426758, 0.03697151947021484, 0.03706163024902344, 0.036975616455078124, 0.03653529739379883, 0.03754393768310547, 0.03695820617675781, 0.03659571075439453, 0.03673907089233398, 0.03706265640258789, 0.037054462432861326, 0.037043201446533204, 0.03692236709594727, 0.036961280822753906, 0.03679948806762695, 0.03628543853759766, 0.03619430541992188, 0.03625267028808594, 0.03640524673461914, 0.0363612174987793, 0.03643392181396484, 0.036410369873046876, 0.036383743286132815, 0.036421630859375, 0.03642367935180664, 0.03643084716796875, 0.03649433517456055, 0.03638784027099609, 0.03643084716796875, 0.03641753768920898, 0.03647488021850586, 0.036327423095703124, 0.03785830307006836, 0.03727155303955078, 0.0370063362121582, 0.03681382369995117, 0.03663974380493164, 0.03644828796386719, 0.03782140731811524, 0.03735039901733399, 0.037013504028320314, 0.03705753707885742, 0.03727974319458008, 0.03707699203491211, 0.03689779281616211, 0.03700838470458984, 0.03696025466918945, 0.03684659194946289, 0.03693260955810547, 0.036779006958007815, 0.036657150268554685, 0.03634483337402344, 0.03633561706542969, 0.03666227340698242, 0.0375551986694336, 0.037364734649658206, 0.037190654754638675, 0.037130241394042966, 0.03686809539794922, 0.03676876831054687, 0.036424705505371094, 0.036482048034667966, 0.036615169525146485, 0.036604927062988284, 0.036790271759033204, 0.03683635330200195, 0.03642572784423828, 0.03630284881591797, 0.03669401550292969, 0.036342784881591796, 0.036354049682617184, 0.03618815994262695, 0.03640217590332031, 0.036528129577636716, 0.037108734130859376, 0.03707494354248047, 0.03704524612426758, 0.037070846557617186, 0.03699609756469727, 0.03694694519042969, 0.0370063362121582, 0.036853759765625, 0.036482048034667966, 0.036383743286132815, 0.03639910507202149, 0.03770163345336914, 0.03722444915771484, 0.03691622543334961, 0.03696230316162109, 0.03704012680053711, 0.03701145553588867, 0.037125118255615236, 0.03692031860351563, 0.03694899368286133, 0.03646054458618164, 0.03640217590332031, 0.03646156692504883, 0.03649638366699219, 0.036410369873046876, 0.0364769287109375, 0.03640524673461914, 0.03650457763671875, 0.03647795104980469, 0.03643084716796875, 0.036468734741210936]",tokens/s,27.233898136638548,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,internlm/internlm2-20b,internlm/internlm2-20b,cuda,0,42,,,True,,,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",internlm2,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,i,i,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/i/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669474cd-2309aaf353ce01945d8df075;b31070f1-0b41-4df9-b71b-3a1b815aec8c) Repository Not Found for url: https://huggingface.co/i/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: i is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Deci/DeciLM-7B,Deci/DeciLM-7B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1302.409216,14990.966784,0.0,14344.51968,14290.684928,s,10,0.8082631759643555,0.08082631759643555,0.0003152849928885365,0.0806977310180664,0.08102052764892578,0.0813618797302246,0.08163496139526366,"[0.08170323181152343, 0.08091744232177735, 0.08079734039306641, 0.08071981048583984, 0.08064323425292969, 0.08067565155029296, 0.08066537475585937, 0.08094467163085937, 0.08064502716064453, 0.0805513916015625]",tokens/s,3167.285206264174,kWh,9.559818039444614e-07,5.238325339823567e-07,6.380359046935491e-06,7.860173384862309e-06,tokens/kWh,32569256.104836483,MB,1302.409216,14990.966784,0.0,14344.51968,14290.687488,s,10,22.69752978515625,2.269752978515625,0.003293547886505128,2.2687360839843747,2.2713015869140625,2.275230627441406,2.2783738598632812,"[2.27915966796875, 2.268241943359375, 2.270428466796875, 2.268171630859375, 2.269941650390625, 2.268164306640625, 2.269230224609375, 2.26937451171875, 2.26678369140625, 2.26803369140625]",tokens/s,27.75632440901159,kWh,2.6737787862939218e-05,1.465318988313723e-05,0.00011754990676386413,0.00015894088450994063,tokens/kWh,396373.785097816,,s,629,23.38711042785646,0.03718141562457304,0.009649119302834136,0.03599564743041992,0.03607531585693359,0.03620618133544922,0.11718672546386719,"[0.03689471817016601, 0.03677798461914063, 0.03669708633422852, 0.03671142578125, 0.03671039962768555, 0.036706302642822264, 0.03671756744384766, 0.03684249496459961, 0.03672371292114258, 0.0365753288269043, 0.036450206756591795, 0.03651891326904297, 0.036604927062988284, 0.036193279266357424, 0.03601510238647461, 0.03600691223144531, 0.0359659538269043, 0.03598438262939453, 0.03603148651123047, 0.03603251266479492, 0.03599359893798828, 0.03602841567993164, 0.036087806701660154, 0.03597926330566406, 0.036021247863769534, 0.03599359893798828, 0.0359659538269043, 0.0360079345703125, 0.03598950576782227, 0.03598950576782227, 0.03599257659912109, 0.03598950576782227, 0.03600076675415039, 0.035983360290527344, 0.03597926330566406, 0.03599359893798828, 0.036001792907714845, 0.035996673583984375, 0.036016128540039063, 0.03600998306274414, 0.03601408004760742, 0.03610214233398437, 0.03616767883300781, 0.03604377746582031, 0.03603046417236328, 0.03600076675415039, 0.036001792907714845, 0.03601203155517578, 0.036038654327392575, 0.036025409698486326, 0.03600479888916015, 0.03602329635620117, 0.03606220626831055, 0.036021247863769534, 0.036013057708740234, 0.036036609649658206, 0.03604889678955078, 0.03604889678955078, 0.03601919937133789, 0.03600998306274414, 0.036018177032470705, 0.03598438262939453, 0.11738829040527343, 0.035939327239990236, 0.036018177032470705, 0.035972095489501955, 0.03594137573242188, 0.03593011093139648, 0.03594956970214844, 0.035992607116699216, 0.0359403190612793, 0.03601510238647461, 0.03615948867797852, 0.03602431869506836, 0.035969024658203126, 0.035961856842041014, 0.035975166320800785, 0.035980289459228515, 0.03599052810668945, 0.03597420883178711, 0.036001792907714845, 0.03595974349975586, 0.03594342422485351, 0.03597824096679687, 0.03599564743041992, 0.036001792907714845, 0.03598950576782227, 0.03600691223144531, 0.03600492858886719, 0.03599763107299805, 0.035983360290527344, 0.03595775985717774, 0.035988479614257815, 0.035972095489501955, 0.03595673751831055, 0.03597107315063477, 0.035952640533447267, 0.035953697204589845, 0.03598640060424805, 0.03601510238647461, 0.03598643112182617, 0.0359659538269043, 0.03600281524658203, 0.036094974517822266, 0.03604991912841797, 0.03604377746582031, 0.03604684829711914, 0.035972095489501955, 0.03607449722290039, 0.03598950576782227, 0.03600076675415039, 0.03599359893798828, 0.03600288009643555, 0.03600377655029297, 0.03600281524658203, 0.03599462509155273, 0.03603353500366211, 0.036004863739013675, 0.03604582214355469, 0.03603251266479492, 0.036034561157226565, 0.03602329635620117, 0.036034561157226565, 0.036029441833496094, 0.03597721481323242, 0.1173391342163086, 0.03602227020263672, 0.03595468902587891, 0.036021247863769534, 0.03594137573242188, 0.036185089111328124, 0.03646054458618164, 0.036094974517822266, 0.03598233413696289, 0.03598233413696289, 0.03598643112182617, 0.03599564743041992, 0.03602025604248047, 0.03595363235473633, 0.0359659538269043, 0.036001792907714845, 0.03600384140014649, 0.03598745727539063, 0.03598950576782227, 0.03603558349609375, 0.03598233413696289, 0.036021247863769534, 0.03596806335449219, 0.0360437126159668, 0.035986465454101564, 0.035980255126953124, 0.03598745727539063, 0.035985408782958986, 0.0360816650390625, 0.036190208435058595, 0.03604582214355469, 0.03600896072387695, 0.0359741439819336, 0.03598950576782227, 0.03605401611328125, 0.03628134536743164, 0.036155391693115234, 0.036029441833496094, 0.03601100921630859, 0.03601510238647461, 0.035983360290527344, 0.03600076675415039, 0.03602227020263672, 0.03601100921630859, 0.036070400238037106, 0.03610214233398437, 0.03603353500366211, 0.03604582214355469, 0.035999744415283204, 0.03600998306274414, 0.03598950576782227, 0.03598233413696289, 0.03600896072387695, 0.03601510238647461, 0.035970046997070314, 0.03604995346069336, 0.03627004623413086, 0.036001792907714845, 0.03599871826171875, 0.0360079345703125, 0.03597619247436523, 0.03599568176269531, 0.035996639251708984, 0.11719987487792968, 0.03598748779296875, 0.035952606201171876, 0.03596492767333984, 0.03594137573242188, 0.035945472717285154, 0.03592806243896484, 0.03595980834960937, 0.036004863739013675, 0.036036609649658206, 0.035983360290527344, 0.03609702301025391, 0.03605401611328125, 0.03595161437988281, 0.03597721481323242, 0.036001792907714845, 0.035961856842041014, 0.0359598388671875, 0.03594953536987305, 0.03596492767333984, 0.035980289459228515, 0.035975166320800785, 0.0359444465637207, 0.036021247863769534, 0.036013057708740234, 0.03598438262939453, 0.03600998306274414, 0.03600076675415039, 0.03598643112182617, 0.036016128540039063, 0.03603353500366211, 0.035988479614257815, 0.0359628791809082, 0.03597312164306641, 0.03597312164306641, 0.03597926330566406, 0.035980289459228515, 0.03604172897338867, 0.035999744415283204, 0.03599052810668945, 0.035969024658203126, 0.03599462509155273, 0.03601408004760742, 0.03605299377441406, 0.03602431869506836, 0.03599977493286133, 0.03602428817749023, 0.03603251266479492, 0.03596492767333984, 0.03601408004760742, 0.0360263671875, 0.03604684829711914, 0.03600076675415039, 0.036040702819824216, 0.03604787063598633, 0.03603148651123047, 0.03600998306274414, 0.035980289459228515, 0.03599155044555664, 0.03598745727539063, 0.03599052810668945, 0.036001792907714845, 0.03599769592285156, 0.1172674560546875, 0.035996673583984375, 0.03598643112182617, 0.03598643112182617, 0.035958782196044925, 0.03593318557739258, 0.035958782196044925, 0.035920894622802735, 0.03601715087890625, 0.03606016159057617, 0.03597011184692383, 0.036011966705322265, 0.035999744415283204, 0.03600182342529297, 0.036175838470458986, 0.03604377746582031, 0.03602739334106445, 0.03599871826171875, 0.035961856842041014, 0.036001792907714845, 0.03603046417236328, 0.03603353500366211, 0.03602534484863281, 0.03599257659912109, 0.03597926330566406, 0.036034561157226565, 0.03603558349609375, 0.03604684829711914, 0.03604172897338867, 0.0359813117980957, 0.036013057708740234, 0.03599257659912109, 0.035983360290527344, 0.0360079345703125, 0.03603353500366211, 0.03602022552490235, 0.03604889678955078, 0.036078590393066406, 0.03604787063598633, 0.03605811309814453, 0.03599052810668945, 0.03599052810668945, 0.036009025573730466, 0.036036575317382816, 0.036012001037597656, 0.03602227020263672, 0.035999744415283204, 0.03598438262939453, 0.03598950576782227, 0.03601919937133789, 0.03599462509155273, 0.03603353500366211, 0.036063232421875, 0.03608575820922852, 0.036018177032470705, 0.03602022552490235, 0.03600588989257812, 0.03605811309814453, 0.03603046417236328, 0.036190208435058595, 0.03621478271484375, 0.036037631988525394, 0.03604275131225586, 0.11716607666015624, 0.03600691223144531, 0.0360263671875, 0.03601408004760742, 0.03598342514038086, 0.03597715377807617, 0.03593625640869141, 0.0359659538269043, 0.035976222991943356, 0.03595977783203125, 0.0359813117980957, 0.035955711364746096, 0.035980289459228515, 0.03591167831420899, 0.035945472717285154, 0.03596083068847656, 0.03596495819091797, 0.03595568084716797, 0.03597926330566406, 0.035969024658203126, 0.0359741439819336, 0.0359628791809082, 0.035953662872314454, 0.0359659538269043, 0.035952640533447267, 0.036135936737060545, 0.03600281524658203, 0.03601103973388672, 0.03595670318603516, 0.03595161437988281, 0.03596492767333984, 0.03596799850463867, 0.03598233413696289, 0.03600384140014649, 0.036001792907714845, 0.035999744415283204, 0.03599155044555664, 0.03601100921630859, 0.036021247863769534, 0.03601103973388672, 0.035996639251708984, 0.036119552612304685, 0.036029441833496094, 0.03601100921630859, 0.03605606460571289, 0.036001792907714845, 0.03601408004760742, 0.03599257659912109, 0.035980289459228515, 0.03599871826171875, 0.03598137664794922, 0.03597206497192383, 0.03598947143554688, 0.03603148651123047, 0.03599257659912109, 0.03602841567993164, 0.03602841567993164, 0.036040702819824216, 0.0360447998046875, 0.03598745727539063, 0.03604172897338867, 0.03612057495117187, 0.035980289459228515, 0.11719475555419921, 0.035963905334472655, 0.03594956970214844, 0.035966976165771485, 0.035934207916259765, 0.035972095489501955, 0.035961856842041014, 0.035958782196044925, 0.035955711364746096, 0.03596905517578125, 0.035968990325927735, 0.035975166320800785, 0.035970046997070314, 0.035917823791503906, 0.03594649505615234, 0.035934207916259765, 0.03595673751831055, 0.03597619247436523, 0.035970046997070314, 0.035963905334472655, 0.03595673751831055, 0.036001792907714845, 0.03637452697753906, 0.03617587280273438, 0.035966976165771485, 0.03596083068847656, 0.0359628791809082, 0.03595468902587891, 0.03595673751831055, 0.035969024658203126, 0.03598233413696289, 0.035950592041015625, 0.035947521209716796, 0.03599052810668945, 0.035980289459228515, 0.03597824096679687, 0.03602739334106445, 0.03601408004760742, 0.03601919937133789, 0.036037631988525394, 0.035980289459228515, 0.03595775985717774, 0.03596492767333984, 0.03598950576782227, 0.03599462509155273, 0.03613695907592773, 0.03628851318359375, 0.03608268737792969, 0.03597619247436523, 0.035985408782958986, 0.035988479614257815, 0.036219905853271485, 0.03650559997558594, 0.03608883285522461, 0.036063232421875, 0.036016128540039063, 0.036018177032470705, 0.03600691223144531, 0.03600281524658203, 0.036016128540039063, 0.03600588989257812, 0.03601715087890625, 0.0360079345703125, 0.11749478149414062, 0.03599769592285156, 0.03597721481323242, 0.035993663787841794, 0.03600070571899414, 0.03599052810668945, 0.03595980834960937, 0.03602739334106445, 0.03602022552490235, 0.03599871826171875, 0.036016128540039063, 0.03599257659912109, 0.03599564743041992, 0.035996673583984375, 0.03599359893798828, 0.03598745727539063, 0.036055038452148434, 0.035972095489501955, 0.036001792907714845, 0.03601100921630859, 0.03600588989257812, 0.036004863739013675, 0.03597721481323242, 0.03598543930053711, 0.03599257659912109, 0.036044769287109375, 0.03601203155517578, 0.03597312164306641, 0.0360079345703125, 0.03603046417236328, 0.03599462509155273, 0.036029441833496094, 0.03599769592285156, 0.0360447998046875, 0.03606118392944336, 0.03604889678955078, 0.036013057708740234, 0.0360079345703125, 0.036001792907714845, 0.036038654327392575, 0.03599359893798828, 0.036051998138427736, 0.03610518264770508, 0.03599462509155273, 0.03599871826171875, 0.03603558349609375, 0.03602329635620117, 0.036001792907714845, 0.035963905334472655, 0.036004863739013675, 0.03598438262939453, 0.03601408004760742, 0.036004863739013675, 0.03604787063598633, 0.03608063888549805, 0.036078590393066406, 0.03600384140014649, 0.036021247863769534, 0.03606220626831055, 0.036004863739013675, 0.036013057708740234, 0.0360447998046875, 0.03602329635620117, 0.11722444915771485, 0.03605811309814453, 0.036168704986572264, 0.036013057708740234, 0.035958782196044925, 0.035969024658203126, 0.0359628791809082, 0.035955711364746096, 0.03597929763793945, 0.035952606201171876, 0.036021247863769534, 0.035975166320800785, 0.03598745727539063, 0.03592704010009766, 0.035942401885986325, 0.03599052810668945, 0.035972095489501955, 0.03606022262573242, 0.035967937469482424, 0.03599564743041992, 0.0359659538269043, 0.03597721481323242, 0.035966976165771485, 0.035947521209716796, 0.035914752960205076, 0.035942401885986325, 0.03597312164306641, 0.03596083068847656, 0.03596499252319336, 0.03595872116088867, 0.035980289459228515, 0.03597312164306641, 0.035963905334472655, 0.0359659538269043, 0.03596083068847656, 0.03597926330566406, 0.03600281524658203, 0.03595673751831055, 0.035945472717285154, 0.03597824096679687, 0.035945472717285154, 0.035931137084960936, 0.035947521209716796, 0.03595980834960937, 0.03598233413696289, 0.035967041015625, 0.03594847869873047, 0.035945472717285154, 0.03596492767333984, 0.035972095489501955, 0.035947521209716796, 0.03598438262939453, 0.03597110366821289, 0.03596796798706055, 0.03597312164306641, 0.035988479614257815, 0.03597619247436523, 0.0359628791809082, 0.035947521209716796, 0.03597107315063477, 0.035958782196044925, 0.035969024658203126, 0.03596083068847656, 0.11708927917480469, 0.03594854354858398, 0.03594035339355469, 0.03596799850463867, 0.03596499252319336, 0.035951553344726564, 0.03592601776123047, 0.03596083068847656, 0.03595161437988281, 0.03591884613037109, 0.03600691223144531, 0.0363059196472168, 0.036122623443603515, 0.03596492767333984, 0.03595775985717774, 0.03595161437988281, 0.035932159423828124, 0.03592601776123047, 0.036127742767333985, 0.03628339385986328, 0.036016128540039063, 0.035975166320800785, 0.03598438262939453, 0.03598438262939453, 0.03597312164306641, 0.035983360290527344, 0.036103168487548826, 0.03593523025512695, 0.035947521209716796, 0.03594553756713867, 0.0359617919921875, 0.036036609649658206, 0.03595468902587891, 0.03599564743041992, 0.035980289459228515, 0.03600281524658203, 0.035970046997070314, 0.03600384140014649, 0.03597107315063477, 0.035953662872314454, 0.0359741439819336, 0.03596799850463867, 0.03599871826171875, 0.036021247863769534, 0.036004863739013675, 0.03595980834960937, 0.035963905334472655, 0.03599257659912109, 0.03597107315063477, 0.03597926330566406, 0.03598745727539063, 0.03597312164306641, 0.03594956970214844, 0.03598643112182617, 0.035988479614257815, 0.03600588989257812, 0.035958782196044925, 0.03597824096679687, 0.03606425476074219, 0.03601408004760742, 0.035988479614257815, 0.03601203155517578, 0.03598745727539063]",tokens/s,26.89515671208345,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,stabilityai/stablelm-3b-4e1t,stabilityai/stablelm-3b-4e1t,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1252.458496,6715.604992,0.0,6069.157888,5763.863552,s,10,0.3416221122741699,0.034162211227416986,0.00027871521922884717,0.03406932830810547,0.034227484130859374,0.03460892601013183,0.0349140795135498,"[0.034990367889404295, 0.03406399917602539, 0.03401990509033203, 0.03410099029541016, 0.034064960479736325, 0.0341097297668457, 0.03400771331787109, 0.034142719268798825, 0.03407369613647461, 0.03404803085327148]",tokens/s,7493.660123339628,kWh,4.0440601203131506e-07,2.2159552568480977e-07,2.472081044784978e-06,3.0980825825011027e-06,tokens/kWh,82631754.70078318,MB,1252.458496,6736.576512,0.0,6090.129408,5863.102976,s,10,15.611999633789063,1.5611999633789062,0.01563306749427106,1.559531982421875,1.5835751342773436,1.5853139953613282,1.5867050842285157,"[1.5870528564453126, 1.5604825439453125, 1.583188720703125, 1.5585814208984374, 1.5727518310546875, 1.541266845703125, 1.546939697265625, 1.5688944091796875, 1.5516402587890625, 1.5412010498046875]",tokens/s,40.353575120287,kWh,1.8315486804379813e-05,1.003699735311458e-05,6.100135862281457e-05,8.935384278030897e-05,tokens/kWh,705062.0100905542,,s,629,15.89164443016052,0.025264935501050113,0.0040845115960043855,0.02488012886047363,0.025364888763427733,0.025515622329711916,0.05883666015625,"[0.025825279235839844, 0.025340927124023437, 0.02527948760986328, 0.025300991058349608, 0.02531123161315918, 0.025244672775268553, 0.02555084800720215, 0.025560064315795897, 0.02532454490661621, 0.025298944473266603, 0.025265151977539063, 0.025250816345214845, 0.02530303955078125, 0.02530816078186035, 0.025243648529052733, 0.025222143173217772, 0.024252416610717774, 0.02414182472229004, 0.024221696853637696, 0.024163328170776367, 0.0241408634185791, 0.025174976348876953, 0.02596249580383301, 0.025833471298217774, 0.025203712463378908, 0.02527743911743164, 0.02528358459472656, 0.025290752410888673, 0.025317375183105468, 0.025227264404296876, 0.025312255859375, 0.025265151977539063, 0.025250816345214845, 0.02528563117980957, 0.025288703918457032, 0.02526924705505371, 0.02528358459472656, 0.02533683204650879, 0.025222143173217772, 0.025404415130615234, 0.025363456726074218, 0.024204288482666016, 0.024233983993530273, 0.024217599868774413, 0.024600576400756836, 0.025312255859375, 0.0253439998626709, 0.02497331237792969, 0.025028608322143556, 0.02532044792175293, 0.02490982437133789, 0.025075712203979493, 0.025414655685424805, 0.025456640243530275, 0.025383935928344727, 0.025273344039916993, 0.025413631439208984, 0.02488012886047363, 0.025205759048461913, 0.025392127990722657, 0.02530201530456543, 0.025384960174560548, 0.05883596801757812, 0.02430463981628418, 0.024239103317260743, 0.02429439926147461, 0.02424934387207031, 0.024367103576660155, 0.025218048095703126, 0.025383935928344727, 0.025333759307861328, 0.024826879501342772, 0.0249487361907959, 0.025232383728027344, 0.02531942367553711, 0.025152511596679687, 0.025288703918457032, 0.02526924705505371, 0.02531532859802246, 0.025203712463378908, 0.02567475128173828, 0.026589183807373046, 0.02554982376098633, 0.025398271560668945, 0.02526924705505371, 0.025250816345214845, 0.025268224716186522, 0.025189376831054686, 0.025255935668945313, 0.025226240158081056, 0.025450496673583983, 0.025424896240234376, 0.025409536361694338, 0.025280511856079102, 0.025433088302612306, 0.0255467529296875, 0.024375295639038085, 0.024284160614013672, 0.02429132843017578, 0.02405580711364746, 0.024246271133422852, 0.02429542350769043, 0.0242872314453125, 0.02430259132385254, 0.02433126449584961, 0.024236032485961914, 0.024366079330444337, 0.024327167510986326, 0.02436403274536133, 0.024308736801147462, 0.024236032485961914, 0.024240127563476564, 0.02438041687011719, 0.024345600128173828, 0.0243189754486084, 0.024308736801147462, 0.02434867286682129, 0.024266752243041992, 0.024375295639038085, 0.024376319885253905, 0.024358911514282225, 0.02411008071899414, 0.024073215484619142, 0.024398880004882814, 0.024391647338867187, 0.058840065002441405, 0.024376319885253905, 0.024243200302124023, 0.024259584426879883, 0.02426572799682617, 0.0243240966796875, 0.024255487442016603, 0.02428313636779785, 0.02424527931213379, 0.024454111099243163, 0.0242872314453125, 0.024262655258178712, 0.02472755241394043, 0.02529689598083496, 0.025366527557373047, 0.025284608840942382, 0.025284608840942382, 0.025298944473266603, 0.025409536361694338, 0.025977855682373048, 0.02570444869995117, 0.025264127731323242, 0.02525491142272949, 0.02533580780029297, 0.025381887435913086, 0.025425920486450194, 0.025385984420776365, 0.025249792098999024, 0.02533990478515625, 0.025273344039916993, 0.02536038398742676, 0.02531328010559082, 0.025268224716186522, 0.02527846336364746, 0.025250816345214845, 0.02528156852722168, 0.025315296173095702, 0.02529996871948242, 0.02525491142272949, 0.02522319984436035, 0.025271263122558594, 0.025332735061645507, 0.025349119186401366, 0.02527948760986328, 0.025252864837646483, 0.0252457275390625, 0.0254289608001709, 0.025249792098999024, 0.02528767967224121, 0.025265151977539063, 0.025242656707763673, 0.025332704544067382, 0.025289728164672853, 0.025298944473266603, 0.025410560607910155, 0.025442304611206053, 0.025289728164672853, 0.025200639724731445, 0.025249792098999024, 0.02529484748840332, 0.025289728164672853, 0.025359359741210938, 0.025359359741210938, 0.0588554573059082, 0.024250335693359375, 0.024266752243041992, 0.024189952850341798, 0.024200191497802736, 0.02430771255493164, 0.024205312728881836, 0.024246271133422852, 0.024209407806396483, 0.024231935501098634, 0.02508799934387207, 0.025364479064941405, 0.025267263412475587, 0.025310144424438477, 0.025260032653808592, 0.025297920227050782, 0.025289728164672853, 0.025274368286132814, 0.025200639724731445, 0.02533580780029297, 0.025333759307861328, 0.025282560348510744, 0.025267200469970705, 0.02527948760986328, 0.025192447662353516, 0.02588057518005371, 0.024392704010009765, 0.02427903938293457, 0.02424934387207031, 0.024237056732177735, 0.024198144912719727, 0.024268800735473633, 0.02429132843017578, 0.024521728515625, 0.02437222480773926, 0.024267776489257813, 0.024473600387573242, 0.024576000213623047, 0.024243200302124023, 0.024226816177368164, 0.024225791931152343, 0.024304704666137697, 0.02419910430908203, 0.024328191757202147, 0.02428313636779785, 0.024308736801147462, 0.02431078338623047, 0.024218624114990234, 0.024172544479370117, 0.024303615570068358, 0.024270847320556642, 0.024250368118286132, 0.026025983810424806, 0.025309183120727538, 0.025181184768676756, 0.02530713653564453, 0.02528358459472656, 0.025284608840942382, 0.025306112289428712, 0.025297920227050782, 0.02531942367553711, 0.025333759307861328, 0.02540447998046875, 0.05883692932128906, 0.02433945655822754, 0.02429439926147461, 0.024178688049316405, 0.024201215744018553, 0.024244224548339844, 0.024195072174072265, 0.024185855865478514, 0.024182783126831055, 0.024194047927856444, 0.024177663803100585, 0.024254463195800782, 0.02411827278137207, 0.024265792846679686, 0.024220672607421875, 0.024144832611083984, 0.02425753593444824, 0.024573951721191405, 0.02470911979675293, 0.025168895721435547, 0.025182207107543944, 0.025187328338623048, 0.025222143173217772, 0.02535424041748047, 0.025194496154785157, 0.025196544647216795, 0.025197568893432616, 0.025178112030029298, 0.025224191665649414, 0.025288703918457032, 0.025200639724731445, 0.025082880020141602, 0.02513817596435547, 0.025801727294921875, 0.02513100814819336, 0.025187328338623048, 0.025083904266357423, 0.025350143432617187, 0.025274368286132814, 0.025624576568603515, 0.025183231353759765, 0.025189376831054686, 0.025195520401000978, 0.025297983169555664, 0.025491392135620117, 0.02551807975769043, 0.025226240158081056, 0.025232383728027344, 0.02511564826965332, 0.025208831787109375, 0.025178112030029298, 0.025189376831054686, 0.025195520401000978, 0.025193471908569336, 0.025025535583496093, 0.0255467529296875, 0.02527539253234863, 0.025240575790405274, 0.02528665542602539, 0.025347072601318358, 0.025238527297973632, 0.02530816078186035, 0.025327680587768554, 0.05893318557739258, 0.024217599868774413, 0.024159231185913087, 0.02426572799682617, 0.024242176055908202, 0.02426982307434082, 0.024260608673095704, 0.024268800735473633, 0.024207359313964845, 0.024226816177368164, 0.0243056640625, 0.024242240905761717, 0.02423289680480957, 0.024220672607421875, 0.02432204818725586, 0.024204288482666016, 0.02424115180969238, 0.024250368118286132, 0.024199167251586915, 0.024253440856933595, 0.024214527130126954, 0.024137727737426756, 0.02425651168823242, 0.02428313636779785, 0.024242176055908202, 0.02412441635131836, 0.024509439468383788, 0.025201663970947266, 0.025194496154785157, 0.025199615478515625, 0.025221120834350585, 0.025290752410888673, 0.025205759048461913, 0.025135103225708007, 0.02524575996398926, 0.025338815689086913, 0.025220096588134764, 0.025332735061645507, 0.025274368286132814, 0.02510438346862793, 0.02409369659423828, 0.023999488830566407, 0.024221759796142578, 0.024348608016967775, 0.02424115180969238, 0.02419820785522461, 0.024258495330810547, 0.024345600128173828, 0.024185855865478514, 0.02428211212158203, 0.02427494430541992, 0.024147968292236328, 0.02431590461730957, 0.024344575881958007, 0.024393728256225586, 0.024255487442016603, 0.024245248794555665, 0.02429030418395996, 0.02432512092590332, 0.024327167510986326, 0.02448486328125, 0.02426982307434082, 0.024329248428344726, 0.05880316925048828, 0.02429542350769043, 0.024261632919311524, 0.0241582088470459, 0.024185855865478514, 0.024189952850341798, 0.024154111862182616, 0.0243189754486084, 0.024238079071044923, 0.02429030418395996, 0.024201215744018553, 0.024220672607421875, 0.024155168533325194, 0.024227807998657227, 0.024235008239746093, 0.02435481643676758, 0.024237056732177735, 0.024222719192504884, 0.024263679504394533, 0.024838144302368165, 0.025402368545532225, 0.02531839942932129, 0.025192447662353516, 0.025511936187744142, 0.025205759048461913, 0.02531839942932129, 0.025235456466674806, 0.02527846336364746, 0.02550681686401367, 0.02530508804321289, 0.025205759048461913, 0.0252938232421875, 0.025288703918457032, 0.025248767852783204, 0.025240575790405274, 0.025351167678833008, 0.025206783294677734, 0.025206783294677734, 0.024212480545043946, 0.024226816177368164, 0.024195135116577147, 0.024299455642700196, 0.02425651168823242, 0.02426470375061035, 0.024185855865478514, 0.02428108787536621, 0.024125471115112304, 0.024230880737304686, 0.0241213436126709, 0.024203327178955077, 0.02420012855529785, 0.024211456298828125, 0.024245248794555665, 0.024275968551635742, 0.024137727737426756, 0.02429952049255371, 0.024212480545043946, 0.024216575622558592, 0.024151039123535157, 0.024156160354614258, 0.02411827278137207, 0.024240127563476564, 0.02471833610534668, 0.05885440063476562, 0.024155136108398437, 0.02428620719909668, 0.02409881591796875, 0.024195072174072265, 0.02425040054321289, 0.024308704376220704, 0.024231935501098634, 0.024197120666503907, 0.024145952224731447, 0.024257503509521484, 0.02429132843017578, 0.024176639556884767, 0.02429542350769043, 0.024275968551635742, 0.024131584167480468, 0.024276992797851563, 0.024210432052612304, 0.024137792587280275, 0.02421139144897461, 0.024228864669799805, 0.024190975189208985, 0.024253440856933595, 0.02429132843017578, 0.02532352066040039, 0.02531635284423828, 0.025249792098999024, 0.02515660858154297, 0.025239551544189453, 0.025149440765380858, 0.02512384033203125, 0.025282560348510744, 0.025159679412841796, 0.02517094421386719, 0.025282560348510744, 0.025300991058349608, 0.025593856811523437, 0.02627276802062988, 0.025371648788452147, 0.025249824523925782, 0.02536240005493164, 0.02526924705505371, 0.0252938232421875, 0.025227264404296876, 0.025272319793701172, 0.025223167419433593, 0.025227264404296876, 0.02530816078186035, 0.02529177665710449, 0.025342975616455078, 0.025270303726196288, 0.02537366485595703, 0.02547302436828613, 0.02534604835510254, 0.025211904525756838, 0.025268224716186522, 0.025255935668945313, 0.025332735061645507, 0.025410560607910155, 0.025419776916503906, 0.025206783294677734, 0.02527539253234863, 0.025233407974243165, 0.058929153442382816, 0.024190975189208985, 0.024169471740722655, 0.024213504791259766, 0.024246271133422852, 0.024195072174072265, 0.024219648361206055, 0.02424934387207031, 0.02425651168823242, 0.024177663803100585, 0.024275968551635742, 0.024250368118286132, 0.024189952850341798, 0.024237056732177735, 0.024236032485961914, 0.024268800735473633, 0.024300544738769532, 0.024311807632446288, 0.02411520004272461, 0.024154111862182616, 0.024201215744018553, 0.024113216400146485, 0.024233919143676758, 0.024311807632446288, 0.024403968811035157, 0.024231935501098634, 0.024203264236450195, 0.024193023681640623, 0.024377344131469726, 0.024246271133422852, 0.025273344039916993, 0.025273344039916993, 0.025178112030029298, 0.025169919967651368, 0.025228288650512694, 0.025255935668945313, 0.025621503829956056, 0.02631884765625, 0.02552422332763672, 0.02534502410888672, 0.025233407974243165, 0.02529484748840332, 0.025221120834350585, 0.025233407974243165, 0.025234432220458985, 0.025238527297973632, 0.025332767486572264, 0.025312223434448243, 0.025200639724731445, 0.02529484748840332, 0.025243648529052733, 0.025247743606567383, 0.024227840423583984, 0.02427187156677246, 0.024229888916015626, 0.024236032485961914, 0.024211456298828125, 0.024218624114990234, 0.024268800735473633, 0.02428927993774414, 0.024263679504394533, 0.02425651168823242, 0.02429747200012207, 0.05894041442871094, 0.024202239990234374, 0.024246271133422852, 0.02415007972717285, 0.0249496955871582, 0.025239551544189453, 0.025111551284790038, 0.025202688217163087, 0.025280511856079102, 0.02429644775390625, 0.024219648361206055, 0.02438860893249512, 0.024162303924560546, 0.024222719192504884, 0.024261695861816406, 0.024246208190917967, 0.024242176055908202, 0.02424115180969238, 0.024188928604125977, 0.024244224548339844, 0.024213504791259766, 0.024200191497802736, 0.024209407806396483, 0.024176639556884767, 0.024206335067749024, 0.024370176315307617, 0.024276992797851563, 0.024216575622558592, 0.024170495986938476, 0.024311807632446288, 0.02412851142883301, 0.02428211212158203, 0.02429747200012207, 0.02424934387207031, 0.024206335067749024, 0.024382463455200197, 0.024181760787963868, 0.024359935760498046, 0.024262655258178712, 0.024436735153198243, 0.024237056732177735, 0.024373247146606446, 0.024183807373046876, 0.02429952049255371, 0.024227840423583984, 0.024235008239746093, 0.024224767684936522, 0.024201248168945314, 0.024249311447143554, 0.024297504425048827, 0.024258527755737305, 0.02434662437438965, 0.02424934387207031, 0.02488217544555664, 0.024400896072387695, 0.024255487442016603, 0.025127935409545898, 0.025779199600219727, 0.025259008407592775, 0.025236480712890624, 0.025250816345214845, 0.025267232894897462, 0.025256927490234376]",tokens/s,39.5805483041283,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,huggyllama/llama-13b,huggyllama/llama-13b,cuda,0,42,,,True,,,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",llama,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-2.7b,,cuda,0,42,,,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,d35829e539df8480b726c647eeabf91e41eae047,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,,MB,1185.927168,6415.712256,0.0,5769.265152,5440.254976,s,11,0.3183155212402343,0.02893777465820312,0.00024440343790442406,0.028865312576293944,0.028916160583496095,0.029311344146728514,0.029627490997314453,"[0.029706527709960937, 0.028916160583496095, 0.02884147262573242, 0.02885798454284668, 0.02887718391418457, 0.028840864181518554, 0.028873600006103516, 0.028865312576293944, 0.028827871322631836, 0.028881824493408204, 0.02882671928405762]",tokens/s,8846.568301250854,kWh,3.4261609909039193e-07,1.877335526960846e-07,1.842044146742026e-06,2.3723937985285025e-06,tokens/kWh,107907886.18600596,MB,1185.927168,6415.712256,0.0,5769.265152,5540.548096,s,11,10.149310302734376,0.9226645729758524,0.004762304900469307,0.9212511596679688,0.9227883911132813,0.9300670776367188,0.9358900268554687,"[0.9373457641601562, 0.9220174560546875, 0.9227883911132813, 0.9211010131835937, 0.9195780639648438, 0.9221399536132813, 0.9212511596679688, 0.9197085571289062, 0.9199305419921875, 0.9225245971679688, 0.9209248046875]",tokens/s,68.28050176111923,kWh,1.0808960003291616e-05,5.922850363901451e-06,4.616753971543917e-05,6.289935008263223e-05,tokens/kWh,1001600.1742026833,,s,692,10.422560778617846,0.015061504015343726,0.0034901749508221744,0.014611984252929687,0.014728191661834715,0.014953076696395874,0.0438364058303833,"[0.01529651165008545, 0.015035391807556153, 0.015278079986572265, 0.014949376106262208, 0.015134719848632813, 0.01509273624420166, 0.015023103713989258, 0.015120384216308593, 0.015137791633605957, 0.01501696014404297, 0.015230976104736327, 0.014863360404968261, 0.014832639694213867, 0.014895135879516602, 0.014941184043884278, 0.014957599639892578, 0.0148274564743042, 0.014936063766479492, 0.015036479949951172, 0.014972864151000976, 0.014844927787780762, 0.014842880249023438, 0.014833663940429688, 0.01490944004058838, 0.015264767646789551, 0.015022080421447754, 0.015358976364135742, 0.01501696014404297, 0.01480191993713379, 0.014784511566162109, 0.014873600006103516, 0.014765055656433105, 0.015286335945129395, 0.01509881591796875, 0.015408127784729005, 0.014867456436157226, 0.014768128395080566, 0.01459609603881836, 0.014683135986328125, 0.01458892822265625, 0.014629887580871583, 0.014625791549682618, 0.014639103889465332, 0.014708736419677734, 0.01461350440979004, 0.014702591896057129, 0.014608448028564452, 0.014644160270690917, 0.014659584045410156, 0.014667807579040527, 0.01465238380432129, 0.014627840042114258, 0.014658559799194336, 0.01466374397277832, 0.014699456214904785, 0.014600192070007324, 0.014698592185974121, 0.014697376251220704, 0.014653440475463866, 0.014702591896057129, 0.014631936073303223, 0.014686207771301269, 0.043905025482177736, 0.01461350440979004, 0.014574591636657714, 0.014592063903808595, 0.014669759750366211, 0.014537728309631348, 0.014637056350708008, 0.014544896125793457, 0.01456332778930664, 0.014578687667846679, 0.01458073616027832, 0.014647328376770019, 0.014566368103027343, 0.014585856437683106, 0.01452236843109131, 0.014577664375305176, 0.01459712028503418, 0.01457151985168457, 0.01468518352508545, 0.014534655570983887, 0.014642175674438476, 0.01459609603881836, 0.014614527702331542, 0.014648320198059082, 0.014589952468872071, 0.014639103889465332, 0.014590975761413574, 0.014816255569458007, 0.01460428810119629, 0.014631936073303223, 0.014627840042114258, 0.01466476821899414, 0.014684096336364746, 0.014617600440979005, 0.014653440475463866, 0.014621696472167968, 0.014680064201354981, 0.014607359886169433, 0.014791680335998534, 0.014671872138977051, 0.014631936073303223, 0.01468518352508545, 0.014611455917358398, 0.014645248413085938, 0.014632960319519044, 0.014607359886169433, 0.014629887580871583, 0.014659584045410156, 0.014599167823791503, 0.014667776107788086, 0.014673919677734374, 0.014624768257141114, 0.014658559799194336, 0.014710783958435059, 0.014608384132385254, 0.014628864288330079, 0.01468723201751709, 0.014630911827087402, 0.014644224166870118, 0.014686207771301269, 0.014630911827087402, 0.014674943923950195, 0.014658559799194336, 0.04384665679931641, 0.014584832191467285, 0.014599167823791503, 0.015581184387207032, 0.014691328048706055, 0.014572607994079589, 0.014601152420043945, 0.014574591636657714, 0.014562303543090821, 0.014622719764709472, 0.014575615882873535, 0.01457254409790039, 0.01464627170562744, 0.014609408378601075, 0.014601216316223145, 0.014564352035522461, 0.014610431671142577, 0.014611455917358398, 0.01458892822265625, 0.014623744010925293, 0.014586879730224609, 0.0146626558303833, 0.014583807945251465, 0.014576640129089356, 0.014603263854980468, 0.014584832191467285, 0.014599167823791503, 0.014619647979736328, 0.01459615993499756, 0.014598079681396484, 0.014706687927246094, 0.014568448066711426, 0.014618623733520507, 0.014611488342285156, 0.014610400199890137, 0.014648320198059082, 0.01459712028503418, 0.014666751861572265, 0.014611455917358398, 0.01459712028503418, 0.014618623733520507, 0.0146810884475708, 0.014619647979736328, 0.014635007858276367, 0.014797823905944824, 0.014641216278076172, 0.014645183563232422, 0.014671872138977051, 0.01467903995513916, 0.014639103889465332, 0.014657535552978516, 0.014639103889465332, 0.014648320198059082, 0.014662688255310058, 0.014663647651672364, 0.014625823974609375, 0.014646240234375, 0.014682111740112304, 0.014637056350708008, 0.014671872138977051, 0.014665727615356445, 0.014655488014221191, 0.014686207771301269, 0.04385177612304687, 0.014581760406494141, 0.014616576194763184, 0.014592000007629394, 0.014536704063415527, 0.014797823905944824, 0.014636032104492188, 0.014543871879577636, 0.01459712028503418, 0.014559231758117675, 0.014566399574279786, 0.01458073616027832, 0.01457049560546875, 0.014578687667846679, 0.01458073616027832, 0.014577664375305176, 0.014565407752990723, 0.014581727981567382, 0.014560256004333496, 0.01460428810119629, 0.014776320457458495, 0.01457254409790039, 0.014639103889465332, 0.014585856437683106, 0.014575615882873535, 0.014593024253845215, 0.014585856437683106, 0.014574591636657714, 0.01460428810119629, 0.01461350440979004, 0.014584832191467285, 0.014638079643249511, 0.01458073616027832, 0.014593024253845215, 0.014601216316223145, 0.01464627170562744, 0.014625791549682618, 0.014620672225952149, 0.014642175674438476, 0.01459609603881836, 0.014621696472167968, 0.014635007858276367, 0.014644224166870118, 0.01461350440979004, 0.014610431671142577, 0.014652416229248047, 0.014607359886169433, 0.014641152381896973, 0.014637056350708008, 0.014663680076599121, 0.014608384132385254, 0.014643199920654297, 0.014599167823791503, 0.01461350440979004, 0.014671872138977051, 0.014647295951843262, 0.014645248413085938, 0.014647295951843262, 0.014618656158447266, 0.014647263526916504, 0.01468723201751709, 0.014664704322814942, 0.014638079643249511, 0.04392652893066406, 0.014581760406494141, 0.014547967910766601, 0.01458073616027832, 0.014640128135681153, 0.01458073616027832, 0.014569472312927247, 0.014601216316223145, 0.014583807945251465, 0.014576640129089356, 0.014568448066711426, 0.01460739231109619, 0.014598112106323242, 0.01459712028503418, 0.014610527992248535, 0.01457859230041504, 0.014615551948547363, 0.01458790397644043, 0.014599167823791503, 0.014584832191467285, 0.014574591636657714, 0.0145797119140625, 0.014603263854980468, 0.01487667179107666, 0.014709759712219238, 0.01460223960876465, 0.014592000007629394, 0.01459404754638672, 0.014558208465576173, 0.014527520179748536, 0.01453769588470459, 0.014511103630065919, 0.01457158374786377, 0.01457759952545166, 0.014544896125793457, 0.014610431671142577, 0.014564352035522461, 0.014564352035522461, 0.014610431671142577, 0.014543871879577636, 0.014559231758117675, 0.01458790397644043, 0.014585951805114745, 0.014576543807983398, 0.014584832191467285, 0.014561280250549317, 0.01460223960876465, 0.01457151985168457, 0.014566399574279786, 0.014589952468872071, 0.014567423820495605, 0.014636032104492188, 0.01459712028503418, 0.014578687667846679, 0.01459609603881836, 0.014582783699035644, 0.014573568344116212, 0.014601216316223145, 0.014576640129089356, 0.014630911827087402, 0.014592000007629394, 0.01459609603881836, 0.01458892822265625, 0.04382207870483398, 0.014517248153686524, 0.014498815536499024, 0.014503935813903808, 0.014535679817199706, 0.014516223907470703, 0.014516223907470703, 0.014504960060119629, 0.014501952171325683, 0.014517184257507323, 0.014520319938659668, 0.014509056091308594, 0.01455513572692871, 0.014640128135681153, 0.014567423820495605, 0.014665727615356445, 0.014560256004333496, 0.014568448066711426, 0.01469542407989502, 0.014589952468872071, 0.014575615882873535, 0.014543871879577636, 0.014560256004333496, 0.014600192070007324, 0.01469542407989502, 0.014649344444274901, 0.014598143577575684, 0.014862336158752442, 0.014589952468872071, 0.014668800354003907, 0.01457151985168457, 0.01459609603881836, 0.01464627170562744, 0.014616576194763184, 0.014643199920654297, 0.014630911827087402, 0.014693375587463378, 0.01458892822265625, 0.01465446376800537, 0.01459609603881836, 0.014590975761413574, 0.014693375587463378, 0.014647295951843262, 0.014676992416381835, 0.01462889575958252, 0.014666720390319824, 0.01460428810119629, 0.014680064201354981, 0.014635007858276367, 0.014620736122131347, 0.014718912124633788, 0.015427583694458008, 0.014797823905944824, 0.014637056350708008, 0.014699520111083985, 0.014660608291625977, 0.014658559799194336, 0.014683135986328125, 0.014649344444274901, 0.014708736419677734, 0.014664704322814942, 0.014730239868164062, 0.014653440475463866, 0.043905025482177736, 0.014606335639953612, 0.014595071792602539, 0.014629887580871583, 0.014528544425964355, 0.014614496231079101, 0.014863360404968261, 0.014936063766479492, 0.014631936073303223, 0.014568448066711426, 0.014643232345581054, 0.014589920043945312, 0.01458790397644043, 0.014578687667846679, 0.014603263854980468, 0.014583871841430665, 0.014586848258972168, 0.014577631950378417, 0.014576640129089356, 0.014611455917358398, 0.014566399574279786, 0.014583871841430665, 0.014538687705993652, 0.014530559539794922, 0.014638079643249511, 0.014603263854980468, 0.014578687667846679, 0.014623744010925293, 0.014539775848388671, 0.014592000007629394, 0.01468825626373291, 0.014541824340820313, 0.01461350440979004, 0.014641152381896973, 0.014558208465576173, 0.014616576194763184, 0.014608384132385254, 0.014640128135681153, 0.014637056350708008, 0.014595104217529296, 0.014623711585998535, 0.014635007858276367, 0.014543871879577636, 0.014564352035522461, 0.014586943626403809, 0.014604224205017089, 0.014642175674438476, 0.014617600440979005, 0.014635071754455567, 0.014645183563232422, 0.0145797119140625, 0.014592000007629394, 0.014659584045410156, 0.014627903938293456, 0.014643136024475098, 0.014642175674438476, 0.014644224166870118, 0.014623744010925293, 0.01467084789276123, 0.014671872138977051, 0.014633983612060546, 0.01466163158416748, 0.014655488014221191, 0.0438548469543457, 0.014529536247253418, 0.014584832191467285, 0.014550016403198243, 0.014544896125793457, 0.014581760406494141, 0.014599167823791503, 0.01458073616027832, 0.014595071792602539, 0.014543871879577636, 0.014530559539794922, 0.014528512001037597, 0.014511103630065919, 0.014545920372009278, 0.014569472312927247, 0.014545920372009278, 0.01460531234741211, 0.014562303543090821, 0.014564352035522461, 0.014584832191467285, 0.014623744010925293, 0.014541824340820313, 0.01455513572692871, 0.014516223907470703, 0.014534655570983887, 0.014511103630065919, 0.014498847961425781, 0.014545887947082519, 0.014540800094604492, 0.01457151985168457, 0.01452236843109131, 0.014589952468872071, 0.014798848152160645, 0.014534655570983887, 0.014589952468872071, 0.014844991683959962, 0.014692288398742676, 0.014604351997375488, 0.014627776145935058, 0.014620672225952149, 0.014585856437683106, 0.01457151985168457, 0.014576640129089356, 0.01455513572692871, 0.01459404754638672, 0.014709759712219238, 0.01457151985168457, 0.014599167823791503, 0.014648320198059082, 0.014601216316223145, 0.014648320198059082, 0.014593024253845215, 0.014630911827087402, 0.014631936073303223, 0.014611455917358398, 0.01459712028503418, 0.01464627170562744, 0.014630911827087402, 0.014635007858276367, 0.014653440475463866, 0.014636032104492188, 0.014636032104492188, 0.014666848182678223, 0.043805599212646484, 0.014519295692443847, 0.014513216018676757, 0.014550975799560547, 0.014515199661254884, 0.014600192070007324, 0.014564352035522461, 0.014542847633361817, 0.014612480163574219, 0.014523391723632812, 0.014537728309631348, 0.014593024253845215, 0.014548992156982422, 0.01461350440979004, 0.014592000007629394, 0.014560256004333496, 0.014585856437683106, 0.014615551948547363, 0.014556159973144531, 0.014607359886169433, 0.01457049560546875, 0.01458073616027832, 0.01460223960876465, 0.014553088188171387, 0.01460531234741211, 0.014592000007629394, 0.014535679817199706, 0.014530559539794922, 0.014559231758117675, 0.014530624389648437, 0.014601152420043945, 0.014561280250549317, 0.014559231758117675, 0.014630911827087402, 0.01461350440979004, 0.014609408378601075, 0.014620672225952149, 0.014667776107788086, 0.014655488014221191, 0.014622719764709472, 0.014626815795898437, 0.014611455917358398, 0.014619647979736328, 0.014624768257141114, 0.014610431671142577, 0.014629887580871583, 0.014631936073303223, 0.014633983612060546, 0.014627840042114258, 0.014636032104492188, 0.014623744010925293, 0.014630911827087402, 0.014652416229248047, 0.01465446376800537, 0.014649344444274901, 0.014581760406494141, 0.014601216316223145, 0.01458073616027832, 0.014575615882873535, 0.0146810884475708, 0.014627840042114258, 0.014628864288330079, 0.014683135986328125, 0.043908096313476565, 0.014556159973144531, 0.014530559539794922, 0.014667776107788086, 0.014729215621948242, 0.014968832015991211, 0.014741503715515136, 0.014599167823791503, 0.014551039695739745, 0.014601216316223145, 0.014569472312927247, 0.01460223960876465, 0.014589952468872071, 0.014592000007629394, 0.01458073616027832, 0.01459712028503418, 0.014545920372009278, 0.014535679817199706, 0.014568448066711426, 0.014520319938659668, 0.014537728309631348, 0.014577664375305176, 0.01449062442779541, 0.014554112434387208, 0.014529600143432617, 0.014536640167236329, 0.015199232101440429, 0.014745599746704101, 0.014556159973144531, 0.014615551948547363, 0.014600192070007324, 0.014556159973144531, 0.014614527702331542, 0.014598143577575684, 0.014609408378601075, 0.014628864288330079, 0.014565376281738282, 0.014561280250549317, 0.01458790397644043, 0.014592000007629394, 0.014564352035522461, 0.014896127700805664, 0.014610431671142577, 0.014623744010925293, 0.014621696472167968, 0.014638079643249511, 0.014610431671142577, 0.014763008117675782, 0.014996479988098145, 0.014694399833679199, 0.014710783958435059, 0.014622719764709472, 0.014641152381896973, 0.014659584045410156, 0.01467801570892334, 0.014625791549682618, 0.01468825626373291, 0.014626815795898437, 0.01467084789276123, 0.014718976020812988, 0.014643199920654297, 0.014680064201354981, 0.014656512260437012, 0.043835391998291014, 0.014551039695739745, 0.01455513572692871, 0.014650367736816406, 0.014582783699035644, 0.014550016403198243, 0.014584896087646484, 0.014553024291992187, 0.014592000007629394, 0.014606335639953612, 0.014599167823791503, 0.01459712028503418, 0.014640128135681153, 0.014569472312927247, 0.01460531234741211, 0.01454694366455078, 0.014599167823791503, 0.014618623733520507, 0.014567423820495605, 0.01460223960876465, 0.014581791877746582, 0.014625760078430175, 0.014577664375305176, 0.014589952468872071, 0.014556159973144531, 0.014620672225952149, 0.014663680076599121, 0.014557184219360352, 0.01458073616027832, 0.014559295654296875, 0.014573504447937011, 0.01455513572692871, 0.014593088150024414, 0.014607295989990235, 0.014617600440979005, 0.014639103889465332, 0.014611455917358398, 0.014668864250183105, 0.014622655868530274, 0.014611455917358398, 0.014631936073303223, 0.014633983612060546, 0.014624832153320312, 0.014614527702331542, 0.014614463806152344, 0.014626815795898437, 0.014649344444274901, 0.014584832191467285, 0.014657535552978516, 0.014638079643249511, 0.014622719764709472, 0.014683135986328125, 0.014632991790771484, 0.014646240234375, 0.014659584045410156, 0.014649344444274901, 0.01459609603881836, 0.014668800354003907, 0.014653440475463866, 0.014631936073303223, 0.014716927528381347, 0.014657535552978516, 0.014644224166870118]",tokens/s,66.39443172350265,,,main,False,False,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,microsoft/phi-1_5,microsoft/phi-1_5,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1294.721024,3855.089664,0.0,3208.64256,2982.448128,s,10,0.24245904159545897,0.024245904159545896,0.0002649227581284578,0.024169376373291016,0.024296899223327638,0.024665985774993897,0.024961255016326903,"[0.025035072326660156, 0.024180896759033205, 0.024166976928710938, 0.02409539222717285, 0.02413792037963867, 0.024214879989624023, 0.024171775817871093, 0.02415292739868164, 0.024178016662597657, 0.02412518310546875]",tokens/s,10558.484365665934,kWh,2.878512477159307e-07,1.5772920276132244e-07,1.318625683177613e-06,1.764206133654866e-06,tokens/kWh,145107759.86797562,MB,1294.721024,3855.089664,0.0,3208.64256,3022.612992,s,10,11.844922973632812,1.1844922973632812,0.011623514036406584,1.1784940795898438,1.1996446166992187,1.2022169006347656,1.204274727783203,"[1.2047891845703125, 1.1773560791015625, 1.173236328125, 1.174958251953125, 1.1882401123046875, 1.179632080078125, 1.1747027587890626, 1.1738843994140624, 1.199072998046875, 1.19905078125]",tokens/s,53.187344603455905,kWh,1.3827396296599198e-05,7.5771007332187664e-06,3.788754457062299e-05,5.929204160044096e-05,tokens/kWh,1062537.202286714,,s,629,12.042174470901486,0.01914495146407232,0.0029360874802776744,0.01864499282836914,0.01933066215515137,0.019551865768432618,0.043126497955322265,"[0.01968639945983887, 0.019614719390869142, 0.01925734329223633, 0.0194150390625, 0.019418111801147463, 0.019386367797851564, 0.01965670394897461, 0.019408960342407227, 0.01934124755859375, 0.019248159408569335, 0.01917027282714844, 0.01984102439880371, 0.01978265571594238, 0.01938739204406738, 0.019314687728881837, 0.0193832950592041, 0.01951955223083496, 0.019243967056274413, 0.019334144592285156, 0.019619840621948242, 0.019519487380981446, 0.019475456237792968, 0.01923891258239746, 0.01925529670715332, 0.01901158332824707, 0.019768320083618163, 0.019911712646484375, 0.019864543914794922, 0.019698688507080078, 0.019636224746704102, 0.019408895492553712, 0.019294208526611328, 0.01882624053955078, 0.01858252716064453, 0.018610176086425782, 0.018727935791015626, 0.01845145606994629, 0.01845248031616211, 0.01840640068054199, 0.01864396858215332, 0.01863577651977539, 0.01859174346923828, 0.01863167953491211, 0.01860095977783203, 0.01862451171875, 0.0187064323425293, 0.01858870315551758, 0.01865212821960449, 0.018589696884155273, 0.01861529541015625, 0.01863987159729004, 0.01886726379394531, 0.019306432723999022, 0.019272703170776367, 0.019325952529907226, 0.019159040451049804, 0.018610176086425782, 0.01861529541015625, 0.0186296329498291, 0.01865727996826172, 0.018601984024047852, 0.01862144088745117, 0.04322099304199219, 0.01860201644897461, 0.019056608200073242, 0.019260416030883788, 0.02003865623474121, 0.018938880920410156, 0.018670591354370117, 0.018738176345825194, 0.01863167953491211, 0.01857535934448242, 0.0186562557220459, 0.01864908790588379, 0.018610176086425782, 0.01861529541015625, 0.01863577651977539, 0.018778112411499022, 0.01863577651977539, 0.018531328201293946, 0.01860505676269531, 0.018661376953125, 0.01865830421447754, 0.01864396858215332, 0.018753536224365236, 0.018679807662963867, 0.018597888946533202, 0.01864499282836914, 0.01865830421447754, 0.0186296329498291, 0.018693119049072265, 0.018714624404907225, 0.018651136398315428, 0.018679807662963867, 0.018748416900634765, 0.01862860870361328, 0.01863376045227051, 0.018447328567504882, 0.018378751754760742, 0.018581504821777343, 0.01863065528869629, 0.01884364891052246, 0.01860915184020996, 0.01884774398803711, 0.018774015426635742, 0.01864192008972168, 0.018640960693359375, 0.018587583541870116, 0.01866035270690918, 0.01868185615539551, 0.018421760559082033, 0.018463743209838866, 0.018587648391723634, 0.018634752273559572, 0.018756671905517577, 0.018641855239868162, 0.01865727996826172, 0.018671615600585938, 0.01862451171875, 0.01862041664123535, 0.018686975479125977, 0.01861631965637207, 0.018689023971557618, 0.018549760818481444, 0.018479103088378905, 0.043186176300048826, 0.018572288513183592, 0.018564096450805666, 0.01859071922302246, 0.01863270378112793, 0.01859686470031738, 0.01860403251647949, 0.01866649627685547, 0.01865830421447754, 0.018676736831665038, 0.0186562557220459, 0.01836851119995117, 0.01863167953491211, 0.018709503173828124, 0.018754560470581053, 0.018675712585449217, 0.01860505676269531, 0.018549760818481444, 0.01865216064453125, 0.01841971206665039, 0.01833782386779785, 0.018559968948364258, 0.01869004821777344, 0.018610176086425782, 0.018679807662963867, 0.01860403251647949, 0.018573312759399413, 0.01862246322631836, 0.018626560211181642, 0.01866854476928711, 0.0186746883392334, 0.01863065528869629, 0.018626560211181642, 0.01860915184020996, 0.01863270378112793, 0.018693119049072265, 0.01858355140686035, 0.018601984024047852, 0.0186296329498291, 0.018655231475830078, 0.01859174346923828, 0.01863167953491211, 0.01861222457885742, 0.018579456329345705, 0.01861324882507324, 0.01862348747253418, 0.018592767715454102, 0.018700288772583007, 0.01864396858215332, 0.01861222457885742, 0.01863065528869629, 0.018581504821777343, 0.018647039413452148, 0.01863270378112793, 0.018569215774536133, 0.018603008270263673, 0.01866854476928711, 0.018588672637939452, 0.018685951232910156, 0.01865318489074707, 0.018593791961669923, 0.01863270378112793, 0.018683904647827147, 0.04312678527832031, 0.01857535934448242, 0.018894847869873048, 0.019122175216674805, 0.018553855895996094, 0.01862348747253418, 0.018542591094970702, 0.01862758445739746, 0.01864396858215332, 0.01862246322631836, 0.018565120697021483, 0.018579456329345705, 0.018610176086425782, 0.018741247177124023, 0.018752511978149415, 0.018535423278808593, 0.0185743350982666, 0.018579456329345705, 0.01855897521972656, 0.018581504821777343, 0.01860403251647949, 0.01859686470031738, 0.018588672637939452, 0.01863167953491211, 0.018572288513183592, 0.01861324882507324, 0.018570240020751954, 0.018646015167236327, 0.018618368148803712, 0.01864192008972168, 0.018655231475830078, 0.01861324882507324, 0.018704383850097657, 0.018576383590698242, 0.01863065528869629, 0.018589696884155273, 0.0186429443359375, 0.01862451171875, 0.01864297676086426, 0.018641887664794923, 0.01860812759399414, 0.01861324882507324, 0.0185743350982666, 0.01860403251647949, 0.018630687713623046, 0.018619359970092772, 0.018635839462280274, 0.01892038345336914, 0.01861123275756836, 0.018631647109985352, 0.01859584045410156, 0.01862451171875, 0.01862451171875, 0.01859584045410156, 0.01862860870361328, 0.018654207229614257, 0.018532352447509767, 0.019187711715698243, 0.01880166435241699, 0.01869824028015137, 0.01865727996826172, 0.01862144088745117, 0.018588672637939452, 0.04312575912475586, 0.018586624145507814, 0.018687999725341797, 0.018717695236206054, 0.019082239151000976, 0.01838591957092285, 0.018372608184814454, 0.01860095977783203, 0.01860710334777832, 0.01860710334777832, 0.018661376953125, 0.018524160385131837, 0.018469888687133788, 0.01863680076599121, 0.01879756736755371, 0.019760128021240234, 0.020015104293823242, 0.019538944244384765, 0.019330047607421876, 0.01939455986022949, 0.01930342483520508, 0.01916316795349121, 0.01927881622314453, 0.018948095321655273, 0.018718719482421875, 0.018589696884155273, 0.01862758445739746, 0.01864908790588379, 0.018626623153686524, 0.018642879486083983, 0.018637887954711913, 0.018603967666625976, 0.018593791961669923, 0.01862451171875, 0.01866035270690918, 0.01864089584350586, 0.01862758445739746, 0.018362367630004883, 0.018542591094970702, 0.01861631965637207, 0.018603008270263673, 0.018669567108154296, 0.01861631965637207, 0.01867366409301758, 0.018601984024047852, 0.018637823104858398, 0.01876479911804199, 0.018699296951293944, 0.018776031494140626, 0.01926655960083008, 0.0192675838470459, 0.019156991958618166, 0.01919692802429199, 0.01918976020812988, 0.01921843147277832, 0.019132415771484376, 0.019151872634887695, 0.019188735961914064, 0.019145727157592773, 0.019154943466186524, 0.0192174072265625, 0.019076095581054688, 0.01864499282836914, 0.04472422409057617, 0.019483648300170898, 0.019268608093261717, 0.0189901123046875, 0.01914057540893555, 0.01921433639526367, 0.019182592391967773, 0.019169279098510742, 0.018737152099609376, 0.01866854476928711, 0.01859891128540039, 0.018626560211181642, 0.018671615600585938, 0.018646015167236327, 0.01862246322631836, 0.01863065528869629, 0.01862860870361328, 0.018684928894042968, 0.018663423538208008, 0.018361343383789062, 0.018618368148803712, 0.018700288772583007, 0.01859174346923828, 0.01866444778442383, 0.018584575653076172, 0.01861631965637207, 0.018637823104858398, 0.018682912826538087, 0.01864393615722656, 0.01869004821777344, 0.01866035270690918, 0.01864806365966797, 0.018397184371948243, 0.018487295150756835, 0.01861427116394043, 0.018766847610473633, 0.01863884735107422, 0.0186060791015625, 0.01845248031616211, 0.018493440628051756, 0.018685951232910156, 0.018505727767944336, 0.018637823104858398, 0.01864192008972168, 0.018589696884155273, 0.018594816207885743, 0.01862860870361328, 0.018695167541503906, 0.01861734390258789, 0.018737152099609376, 0.01861734390258789, 0.018646015167236327, 0.018563072204589845, 0.01879347229003906, 0.018676799774169923, 0.018655168533325196, 0.01865936088562012, 0.018711551666259766, 0.01871356773376465, 0.018682880401611326, 0.018634752273559572, 0.01865318489074707, 0.0186296329498291, 0.043216896057128903, 0.01869004821777344, 0.018710559844970703, 0.01868387222290039, 0.018671615600585938, 0.018744319915771485, 0.01863270378112793, 0.01868083190917969, 0.018700288772583007, 0.018494464874267577, 0.018482175827026368, 0.018594816207885743, 0.01864192008972168, 0.018661376953125, 0.01863372802734375, 0.018547712326049806, 0.01865727996826172, 0.01883340835571289, 0.01863372802734375, 0.018676736831665038, 0.01868083190917969, 0.01862553596496582, 0.018765823364257812, 0.018571264266967775, 0.018603008270263673, 0.01866035270690918, 0.018717695236206054, 0.018597888946533202, 0.018919424057006837, 0.018626560211181642, 0.018611200332641603, 0.018729984283447267, 0.0186296329498291, 0.018581504821777343, 0.018655231475830078, 0.01859891128540039, 0.01859584045410156, 0.01864396858215332, 0.018569280624389648, 0.018589632034301758, 0.018628671646118165, 0.018622400283813477, 0.018611200332641603, 0.018646080017089842, 0.018692031860351562, 0.018569215774536133, 0.018710527420043945, 0.018719743728637696, 0.018618368148803712, 0.018594816207885743, 0.018670591354370117, 0.018565120697021483, 0.01863577651977539, 0.01863372802734375, 0.01860403251647949, 0.0186060791015625, 0.01859071922302246, 0.018620447158813478, 0.018638816833496094, 0.018561023712158203, 0.018585599899291993, 0.01861529541015625, 0.018573312759399413, 0.043245567321777346, 0.018637823104858398, 0.018581504821777343, 0.01863577651977539, 0.018576383590698242, 0.018584575653076172, 0.018572288513183592, 0.018588672637939452, 0.01866854476928711, 0.01864089584350586, 0.018589696884155273, 0.018572288513183592, 0.018565120697021483, 0.01862246322631836, 0.018547712326049806, 0.018579456329345705, 0.018637823104858398, 0.018547712326049806, 0.018693119049072265, 0.018549760818481444, 0.01863065528869629, 0.01862451171875, 0.018589696884155273, 0.01862451171875, 0.018661376953125, 0.018737152099609376, 0.018678783416748047, 0.018587648391723634, 0.0186265926361084, 0.018593759536743165, 0.01863065528869629, 0.018564096450805666, 0.01864806365966797, 0.018576383590698242, 0.018566144943237304, 0.01866547203063965, 0.018663423538208008, 0.018678783416748047, 0.01867263984680176, 0.018581504821777343, 0.018594816207885743, 0.018647039413452148, 0.0186060791015625, 0.01859584045410156, 0.018749439239501953, 0.018714624404907225, 0.01863372802734375, 0.01868185615539551, 0.018684959411621092, 0.018751455307006837, 0.018741247177124023, 0.01844326400756836, 0.01859891128540039, 0.018675712585449217, 0.018707456588745116, 0.0186060791015625, 0.0186562557220459, 0.01866649627685547, 0.018618368148803712, 0.018586624145507814, 0.01860915184020996, 0.018644031524658204, 0.018650047302246092, 0.04323328018188476, 0.01875660705566406, 0.019636224746704102, 0.02063667106628418, 0.019869695663452147, 0.019645503997802734, 0.019401664733886718, 0.019336191177368164, 0.019396608352661132, 0.01942425537109375, 0.019360767364501954, 0.019305471420288087, 0.01924710464477539, 0.019560480117797853, 0.0194682559967041, 0.01945599937438965, 0.018988031387329102, 0.018634752273559572, 0.019355648040771483, 0.019726335525512697, 0.019252223968505858, 0.019300352096557616, 0.018969600677490234, 0.01864806365966797, 0.0186296329498291, 0.019569664001464843, 0.01926655960083008, 0.01922867202758789, 0.01925632095336914, 0.019297279357910157, 0.019153919219970703, 0.01925324821472168, 0.018780160903930664, 0.018741247177124023, 0.018655231475830078, 0.01845145606994629, 0.018471935272216796, 0.0186562557220459, 0.01862860870361328, 0.01861631965637207, 0.018589696884155273, 0.01860710334777832, 0.01863065528869629, 0.018610176086425782, 0.01921023941040039, 0.019200000762939453, 0.019288063049316406, 0.019276800155639647, 0.01899929618835449, 0.018892799377441406, 0.019208192825317383, 0.01864089584350586, 0.01868185615539551, 0.018588672637939452, 0.01863167953491211, 0.01867263984680176, 0.01857846450805664, 0.018790367126464844, 0.0186429443359375, 0.018537471771240235, 0.018490367889404297, 0.01863884735107422, 0.01860812759399414, 0.04307251358032226, 0.01856716728210449, 0.018584575653076172, 0.018569215774536133, 0.018589696884155273, 0.018562047958374024, 0.018651136398315428, 0.018749439239501953, 0.019147775650024415, 0.019188735961914064, 0.01918671989440918, 0.019208160400390625, 0.019184640884399414, 0.019430400848388672, 0.019330047607421876, 0.019113983154296875, 0.019257408142089844, 0.01914361572265625, 0.018745344161987306, 0.018603008270263673, 0.018777088165283205, 0.018974720001220705, 0.01945907211303711, 0.020965375900268556, 0.019904512405395508, 0.0194201602935791, 0.01948876762390137, 0.019381248474121093, 0.019333120346069335, 0.019307519912719725, 0.019315711975097655, 0.019327999114990235, 0.019342336654663086, 0.019075071334838867, 0.0192993278503418, 0.019098623275756836, 0.01900441551208496, 0.01922559928894043, 0.019316736221313476, 0.019375104904174805, 0.019317760467529296, 0.019319807052612305, 0.0192675838470459, 0.019284992218017577, 0.019272703170776367, 0.018692096710205077, 0.01862144088745117, 0.01864396858215332, 0.018697216033935548, 0.01859993553161621, 0.018601984024047852, 0.0186429443359375, 0.01860508728027344, 0.018572256088256835, 0.01861529541015625, 0.01861020851135254, 0.018591712951660157, 0.018646015167236327, 0.018704383850097657, 0.018619392395019533, 0.018972671508789063, 0.0190382080078125, 0.018964479446411133]",tokens/s,52.23309141716101,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,google/recurrentgemma-2b,google/recurrentgemma-2b,cuda,0,42,,,True,,,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",recurrent_gemma,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/gpt-j-6b,EleutherAI/gpt-j-6b,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1274.322944,13248.233472,0.0,12601.786368,12440.741376,s,10,0.9218367080688477,0.09218367080688476,0.001666277690247654,0.09160566711425781,0.09234149627685546,0.09475706710815429,0.09668952377319336,"[0.09717263793945312, 0.09150502777099609, 0.09163021087646485, 0.09155305480957031, 0.09150707244873046, 0.09158112335205078, 0.0915707550048828, 0.09170790100097656, 0.09180422210693359, 0.09180470275878906]",tokens/s,2777.064503498602,kWh,1.0860646172036455e-06,5.94903035503903e-07,5.722828227596333e-06,7.403795880303882e-06,tokens/kWh,34576858.16015402,MB,1274.646528,13248.233472,0.0,12601.786368,12548.118016,s,10,28.16623950195313,2.8166239501953125,0.010036860751375876,2.8111727294921875,2.8310316162109377,2.833289855957031,2.8350964477539065,"[2.810013916015625, 2.811123779296875, 2.8112216796875, 2.80408203125, 2.808963134765625, 2.835548095703125, 2.817852783203125, 2.81043408203125, 2.83052978515625, 2.82647021484375]",tokens/s,22.367203117630027,kWh,3.308196343610653e-05,1.812979639462739e-05,0.00012616358283700318,0.00017737534266773714,tokens/kWh,355179.0178526268,,s,629,28.94576739883422,0.04601870810625473,0.010922542204483092,0.044611583709716796,0.04518645629882813,0.04530380706787109,0.1363861962890625,"[0.04497100830078125, 0.045134849548339843, 0.04494745635986328, 0.04431872177124024, 0.04427264022827149, 0.044349441528320314, 0.044267520904541016, 0.044478462219238284, 0.04492697525024414, 0.04489318466186523, 0.04562739181518555, 0.04513491058349609, 0.044500926971435546, 0.04443852615356445, 0.044343360900878905, 0.04437088012695312, 0.044335105895996096, 0.04443852615356445, 0.044375038146972655, 0.04437606430053711, 0.04444364929199219, 0.044521472930908204, 0.044695552825927735, 0.044478462219238284, 0.044461055755615236, 0.044439552307128906, 0.0443955192565918, 0.04448153686523437, 0.044652542114257815, 0.04460134506225586, 0.044483585357666014, 0.04436991882324219, 0.04453580856323242, 0.0444661750793457, 0.04442726516723633, 0.0445296630859375, 0.04453171157836914, 0.04452044677734375, 0.044461055755615236, 0.04435763168334961, 0.04455014419555664, 0.044690433502197265, 0.044510208129882815, 0.04458393478393555, 0.04439244842529297, 0.04441600036621094, 0.04456243133544922, 0.044424190521240234, 0.04448255920410156, 0.0445030403137207, 0.04456447982788086, 0.04451225662231445, 0.04458911895751953, 0.04448659133911133, 0.044568641662597656, 0.044650432586669925, 0.04460031890869141, 0.044598270416259765, 0.04492287826538086, 0.04570828628540039, 0.04528537750244141, 0.04461260986328125, 0.13629542541503906, 0.04419379043579102, 0.04425523376464844, 0.04421017456054688, 0.04428799819946289, 0.04433414459228516, 0.04434220886230469, 0.04443142318725586, 0.04435756683349609, 0.0448276481628418, 0.04432691192626953, 0.04438118362426758, 0.04428902435302735, 0.04441292953491211, 0.0443873291015625, 0.04428287887573242, 0.044203006744384765, 0.04436684799194336, 0.04436172866821289, 0.04433100891113281, 0.0447907829284668, 0.04475913619995117, 0.04466473770141602, 0.044801025390625, 0.04458803176879883, 0.04434124755859375, 0.04435459136962891, 0.04444054412841797, 0.044437503814697264, 0.044450817108154295, 0.04444160079956055, 0.048661502838134765, 0.04487372970581055, 0.044641281127929686, 0.044537857055664064, 0.04458291244506836, 0.044611583709716796, 0.04482867050170898, 0.044572673797607425, 0.04459929656982422, 0.044483585357666014, 0.04474367904663086, 0.04449689483642578, 0.04455526351928711, 0.044507137298583986, 0.04459929656982422, 0.044655616760253904, 0.044668926239013675, 0.044644351959228515, 0.044682239532470705, 0.044846080780029295, 0.04474163055419922, 0.04452044677734375, 0.044649471282958986, 0.044633087158203126, 0.0451143684387207, 0.04516556930541992, 0.04469247817993164, 0.04461875152587891, 0.04474879837036133, 0.04469247817993164, 0.04469760131835938, 0.044649471282958986, 0.1364920654296875, 0.044385246276855465, 0.04425830459594727, 0.04436479949951172, 0.04433715057373047, 0.04531609725952149, 0.044355583190917966, 0.04439039993286133, 0.04433926391601563, 0.0443309440612793, 0.044278785705566405, 0.04424703979492187, 0.04441497421264649, 0.044439552307128906, 0.0444139518737793, 0.044371967315673826, 0.044631038665771484, 0.04453273773193359, 0.04431769561767578, 0.04440576171875, 0.044440574645996093, 0.04440883255004883, 0.04455737686157227, 0.04449068832397461, 0.044622848510742184, 0.04438016128540039, 0.04445695877075195, 0.04444364929199219, 0.04460236740112305, 0.045246463775634765, 0.04511948776245117, 0.04500275039672851, 0.04450406265258789, 0.044624897003173826, 0.044493824005126956, 0.04451023864746094, 0.04460131072998047, 0.04456755065917969, 0.044507137298583986, 0.044565502166748046, 0.044505088806152344, 0.04453376007080078, 0.04456243133544922, 0.04465868759155273, 0.044616703033447266, 0.04464025497436523, 0.04461568069458008, 0.04463513565063477, 0.044609535217285154, 0.04460543823242188, 0.04462182235717774, 0.044824607849121095, 0.04683875274658203, 0.04523212814331055, 0.04506828689575195, 0.04496486282348633, 0.044693504333496094, 0.044639232635498044, 0.044624897003173826, 0.0446484489440918, 0.044627967834472655, 0.04460748672485351, 0.04475289535522461, 0.1363968048095703, 0.04435148620605469, 0.044306430816650394, 0.04427980804443359, 0.044262401580810545, 0.04429926300048828, 0.044298240661621094, 0.044300289154052735, 0.044295169830322265, 0.04432896041870117, 0.04438016128540039, 0.04432076644897461, 0.044336128234863284, 0.04436787033081055, 0.044339199066162106, 0.044281856536865234, 0.044380191802978516, 0.044777473449707034, 0.04493308639526367, 0.04489011383056641, 0.04485836791992188, 0.044371967315673826, 0.044386302947998044, 0.04450201416015625, 0.04449689483642578, 0.044477439880371096, 0.04446003341674805, 0.044447742462158206, 0.04445183944702148, 0.044491775512695314, 0.04445491027832031, 0.04444876861572265, 0.0444846076965332, 0.04451942443847656, 0.044467201232910154, 0.04452556610107422, 0.04439039993286133, 0.04446822357177734, 0.044469249725341796, 0.04449587249755859, 0.04449587249755859, 0.04452249526977539, 0.04457164764404297, 0.04454195022583008, 0.04453478240966797, 0.0445665283203125, 0.04451436614990235, 0.04457363128662109, 0.04455526351928711, 0.04462387084960937, 0.044614654541015625, 0.04455219268798828, 0.04458803176879883, 0.04462080001831055, 0.044560382843017575, 0.04458092880249023, 0.04462176132202148, 0.04469145584106445, 0.04470988845825195, 0.044677120208740234, 0.04472940826416016, 0.0447077751159668, 0.04466483306884766, 0.13635891723632812, 0.04417638397216797, 0.04423372650146484, 0.044459007263183595, 0.04421017456054688, 0.044409854888916016, 0.044251136779785157, 0.04432588958740234, 0.04428902435302735, 0.04430745697021484, 0.04434329605102539, 0.044377086639404296, 0.04434739303588867, 0.04435968017578125, 0.044352512359619144, 0.04452761459350586, 0.04432998275756836, 0.04428902435302735, 0.04441702270507813, 0.04434329605102539, 0.044421119689941405, 0.04442828750610352, 0.04451225662231445, 0.04456959915161133, 0.045592575073242186, 0.044593151092529294, 0.04449689483642578, 0.04448057556152344, 0.0448603515625, 0.044633087158203126, 0.04491980743408203, 0.04465356826782227, 0.045041664123535156, 0.04501606369018555, 0.0449699821472168, 0.04516352081298828, 0.04485734558105469, 0.04458803176879883, 0.04455526351928711, 0.044639232635498044, 0.04458803176879883, 0.04453273773193359, 0.04453580856323242, 0.04451430511474609, 0.04457990264892578, 0.04466886520385742, 0.04464025497436523, 0.04455321502685547, 0.04456140899658203, 0.04472422409057617, 0.044706817626953124, 0.04459724807739258, 0.04469145584106445, 0.04463206481933594, 0.044560382843017575, 0.044608512878417966, 0.044627967834472655, 0.044665855407714845, 0.04467507171630859, 0.04469657516479492, 0.04484505462646484, 0.0449699821472168, 0.04468735885620117, 0.1364172821044922, 0.0454769287109375, 0.044771263122558594, 0.044848129272460936, 0.044837886810302735, 0.044813312530517575, 0.0443955192565918, 0.044767230987548826, 0.04502732849121094, 0.044862464904785154, 0.044899391174316405, 0.044424129486083985, 0.044407806396484374, 0.04434636688232422, 0.04489011383056641, 0.045020160675048826, 0.044935169219970705, 0.0450621452331543, 0.04482252883911133, 0.044483585357666014, 0.044542976379394535, 0.04496384048461914, 0.04487372970581055, 0.04444364929199219, 0.04499251174926758, 0.044805118560791016, 0.04503859329223633, 0.045058048248291016, 0.04501606369018555, 0.04504883193969727, 0.04544921493530273, 0.04500582504272461, 0.04475904083251953, 0.045859840393066405, 0.045992961883544924, 0.04516147232055664, 0.04503551864624023, 0.045071361541748046, 0.04507852935791016, 0.045085697174072265, 0.045146110534667966, 0.04518297576904297, 0.0451512336730957, 0.04508364868164062, 0.04512460708618164, 0.045262847900390625, 0.0451143684387207, 0.04512255859375, 0.045139968872070314, 0.04520959854125976, 0.04516556930541992, 0.045126655578613284, 0.045126655578613284, 0.04516150283813476, 0.045107166290283204, 0.045224960327148435, 0.0451409912109375, 0.04574924850463867, 0.04531916809082031, 0.0452863998413086, 0.04473548889160156, 0.0446453742980957, 0.045018112182617184, 0.13650534057617186, 0.044286975860595705, 0.04424703979492187, 0.04424703979492187, 0.044352512359619144, 0.04434739303588867, 0.0442531852722168, 0.044391422271728515, 0.04432281494140625, 0.04441088104248047, 0.04442521667480469, 0.0444139518737793, 0.044913665771484375, 0.04489215850830078, 0.04494028854370117, 0.044316673278808595, 0.04531302261352539, 0.04507340621948242, 0.04500172805786133, 0.04493414306640625, 0.045071361541748046, 0.04483686447143555, 0.04498332977294922, 0.044930015563964844, 0.04496281433105469, 0.04494950485229492, 0.04503551864624023, 0.04482355117797852, 0.0449699821472168, 0.04499660873413086, 0.04489318466186523, 0.04455942535400391, 0.04464838409423828, 0.04453376007080078, 0.04496486282348633, 0.045050880432128904, 0.04499967956542969, 0.04501094436645508, 0.045050880432128904, 0.044777473449707034, 0.04452249526977539, 0.04449484634399414, 0.04450099182128906, 0.04449894332885742, 0.04451430511474609, 0.04440991973876953, 0.04465248107910156, 0.04453376007080078, 0.04457369613647461, 0.04456143951416015, 0.04456854248046875, 0.04465868759155273, 0.045300735473632815, 0.045227008819580077, 0.045088768005371094, 0.04501606369018555, 0.044652542114257815, 0.04456755065917969, 0.04467814254760742, 0.04463616180419922, 0.04479487991333008, 0.044723201751708984, 0.04468735885620117, 0.137059326171875, 0.044777473449707034, 0.04430950546264648, 0.04432179260253906, 0.04434124755859375, 0.04437609481811523, 0.04422857666015625, 0.0442081298828125, 0.04429619216918945, 0.04433817672729492, 0.0444323844909668, 0.04436787033081055, 0.04431257629394531, 0.044388351440429685, 0.04421529769897461, 0.044213249206542966, 0.04440883255004883, 0.04432076644897461, 0.04437913513183594, 0.04439961624145508, 0.04443340682983398, 0.04447641754150391, 0.04437811279296875, 0.044267520904541016, 0.04432998275756836, 0.04432998275756836, 0.04434841537475586, 0.04441190338134766, 0.04449280166625977, 0.04449894332885742, 0.044393470764160156, 0.044631038665771484, 0.04516966247558594, 0.04500172805786133, 0.04513075256347656, 0.045052928924560545, 0.04500172805786133, 0.04504576110839844, 0.04503859329223633, 0.04455014419555664, 0.04459110260009766, 0.044507137298583986, 0.04442428970336914, 0.04450191879272461, 0.044624897003173826, 0.04467609786987305, 0.044469249725341796, 0.04444467163085938, 0.04452761459350586, 0.04460748672485351, 0.04459212875366211, 0.04467507171630859, 0.04565606307983398, 0.045257728576660154, 0.0451962890625, 0.04526387023925781, 0.04463206481933594, 0.04465356826782227, 0.0446453742980957, 0.044685310363769534, 0.04471295928955078, 0.04487680053710937, 0.045246463775634765, 0.13659852600097655, 0.04443852615356445, 0.04432179260253906, 0.04510105514526367, 0.044870655059814454, 0.044854270935058595, 0.044336128234863284, 0.0443504638671875, 0.044335105895996096, 0.044267520904541016, 0.044326976776123045, 0.044315582275390626, 0.04480716705322266, 0.044900352478027344, 0.04500172805786133, 0.045018112182617184, 0.045009918212890625, 0.044921855926513675, 0.04482048034667969, 0.04440883255004883, 0.04437299346923828, 0.04437811279296875, 0.04436787033081055, 0.04441292953491211, 0.045018112182617184, 0.0451860466003418, 0.044767230987548826, 0.04453887939453125, 0.04445798492431641, 0.0444846076965332, 0.04501203155517578, 0.04508870315551758, 0.04510310363769531, 0.0449617919921875, 0.044724254608154296, 0.04514096069335938, 0.045096958160400394, 0.045021183013916014, 0.045058048248291016, 0.045055999755859374, 0.04510003280639648, 0.045176830291748044, 0.045142017364501956, 0.0451932144165039, 0.04509798431396484, 0.04510105514526367, 0.045282302856445314, 0.0451860466003418, 0.04518809509277344, 0.045175807952880856, 0.04527206420898437, 0.04521984100341797, 0.04516864013671875, 0.045267967224121096, 0.045174785614013675, 0.0452863998413086, 0.04537548828125, 0.04534067153930664, 0.04535398483276367, 0.04525568008422851, 0.04523519897460938, 0.04530585479736328, 0.046170112609863284, 0.1374433288574219, 0.04499353790283203, 0.044829696655273435, 0.045532161712646485, 0.04500172805786133, 0.04492902374267578, 0.04449484634399414, 0.044333057403564455, 0.04431872177124024, 0.04419071960449219, 0.044114944458007815, 0.0442716178894043, 0.04472115325927734, 0.04476518249511719, 0.044424190521240234, 0.044388351440429685, 0.044352512359619144, 0.04432595062255859, 0.044305343627929684, 0.04434022521972656, 0.04519116973876953, 0.045257728576660154, 0.04493107223510742, 0.04503756713867187, 0.04443033599853516, 0.044674049377441405, 0.045061119079589845, 0.04499871826171875, 0.044963775634765626, 0.04441190338134766, 0.044409854888916016, 0.044461055755615236, 0.04448153686523437, 0.0444139518737793, 0.044499969482421874, 0.04451225662231445, 0.04469145584106445, 0.04493414306640625, 0.04514918518066406, 0.04515327835083008, 0.04503859329223633, 0.044467201232910154, 0.04493619155883789, 0.045284351348876956, 0.04563046264648438, 0.04656332778930664, 0.04493414306640625, 0.045036544799804686, 0.04520243072509766, 0.04527718353271484, 0.04510003280639648, 0.045125633239746096, 0.04523417663574219, 0.04510924911499024, 0.04473855972290039, 0.04482457733154297, 0.044930049896240234, 0.0450621452331543, 0.0451512336730957, 0.045243392944335936, 0.045314048767089846, 0.045192192077636716, 0.0452147216796875]",tokens/s,21.730292768998503,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,stabilityai/stablelm-2-1_6b,stabilityai/stablelm-2-1_6b,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1272.25856,4276.617216,0.0,3632.267264,3408.132096,s,10,0.21382742500305174,0.021382742500305175,0.00040380614256645505,0.021255472183227538,0.02143521595001221,0.02201244020462036,0.022474219608306884,"[0.022589664459228515, 0.021306943893432618, 0.021212192535400392, 0.02128316879272461, 0.021218847274780274, 0.021263967514038085, 0.021209951400756835, 0.021289920806884764, 0.02120579147338867, 0.021246976852416992]",tokens/s,11972.271564152557,kWh,2.526427775706321e-07,1.3843550400918338e-07,1.2562159859785875e-06,1.6472942675584028e-06,tokens/kWh,155406356.3758039,MB,1272.25856,4278.714368,0.0,3632.267264,3474.14784,s,10,12.003229980468753,1.200322998046875,0.009540185117779846,1.2029205322265626,1.2098683227539062,1.2106496154785156,1.211274649658203,"[1.206054443359375, 1.1821016845703125, 1.1959776611328126, 1.20736474609375, 1.211430908203125, 1.19922900390625, 1.19978662109375, 1.206357177734375, 1.1852330322265625, 1.2096947021484374]",tokens/s,52.48587263804115,kWh,1.3756805777210114e-05,7.5383188242675455e-06,3.8057648798421595e-05,5.935277339989927e-05,tokens/kWh,1061449.9776704102,,s,629,12.17444862937927,0.019355244243846218,0.002585969156717962,0.019136512756347656,0.01934458808898926,0.019548159408569337,0.04027453338623047,"[0.019579967498779296, 0.01917024040222168, 0.019124223709106446, 0.01923072052001953, 0.019338239669799806, 0.019154943466186524, 0.01930342483520508, 0.019219455718994142, 0.01925119972229004, 0.019072000503540038, 0.019138559341430664, 0.019336191177368164, 0.019397632598876953, 0.01919385528564453, 0.019192832946777344, 0.019117055892944337, 0.01942118453979492, 0.01923583984375, 0.01924608039855957, 0.019359807968139648, 0.01912006378173828, 0.019132415771484376, 0.018818080902099608, 0.01891529655456543, 0.018658367156982422, 0.018852800369262696, 0.01965772819519043, 0.02002841567993164, 0.019401727676391603, 0.0196177921295166, 0.019132415771484376, 0.01921843147277832, 0.018778175354003907, 0.018218944549560547, 0.018255872726440428, 0.018367488861083983, 0.01835212707519531, 0.018981887817382814, 0.01927577590942383, 0.019293184280395507, 0.019075071334838867, 0.019713151931762696, 0.019009408950805665, 0.018917375564575196, 0.01899622344970703, 0.019100671768188478, 0.019843072891235353, 0.01965363121032715, 0.01923993682861328, 0.019124223709106446, 0.018924543380737305, 0.018934783935546876, 0.01882316780090332, 0.019082239151000976, 0.019115007400512696, 0.01903615951538086, 0.0188590087890625, 0.01902079963684082, 0.018898944854736328, 0.01880985641479492, 0.01906892776489258, 0.019330047607421876, 0.040850433349609375, 0.01903923225402832, 0.019086336135864256, 0.019113983154296875, 0.01855897521972656, 0.018543615341186523, 0.018523136138916017, 0.018457599639892578, 0.01848422431945801, 0.018448383331298827, 0.018490367889404297, 0.01844326400756836, 0.018428928375244142, 0.01845043182373047, 0.018496511459350586, 0.01846784019470215, 0.018446367263793947, 0.01846883201599121, 0.018951168060302736, 0.019107839584350587, 0.01906073570251465, 0.019155967712402345, 0.01908121681213379, 0.019148799896240236, 0.018694143295288086, 0.0184586238861084, 0.01859993553161621, 0.018546688079833985, 0.01846886444091797, 0.01863270378112793, 0.01841971206665039, 0.018463743209838866, 0.0184453125, 0.018408447265625, 0.01843404769897461, 0.01839923286437988, 0.0184586238861084, 0.019177471160888672, 0.019204095840454103, 0.01947238349914551, 0.019080192565917968, 0.01925017547607422, 0.01903308868408203, 0.01920921516418457, 0.019306495666503908, 0.018524160385131837, 0.018524160385131837, 0.018536447525024414, 0.018594816207885743, 0.019140607833862306, 0.019133440017700197, 0.019115007400512696, 0.019005439758300782, 0.01904435157775879, 0.019096576690673828, 0.018472959518432617, 0.01847091293334961, 0.018512895584106445, 0.01861529541015625, 0.018514944076538087, 0.01922764778137207, 0.018951168060302736, 0.018513919830322266, 0.04029747009277344, 0.019301376342773437, 0.019135488510131835, 0.01906380844116211, 0.019135488510131835, 0.01909663963317871, 0.019165119171142577, 0.019080192565917968, 0.019080255508422853, 0.01914566421508789, 0.019142656326293944, 0.019163135528564454, 0.018508895874023438, 0.01861008071899414, 0.018505727767944336, 0.018519039154052733, 0.018461696624755858, 0.019074047088623047, 0.01841868782043457, 0.018594816207885743, 0.018481151580810547, 0.018499584197998048, 0.018512895584106445, 0.018518016815185546, 0.018479135513305663, 0.01847702407836914, 0.018485248565673826, 0.01881395149230957, 0.019264511108398438, 0.018555904388427736, 0.018511903762817382, 0.018492383956909178, 0.01845452880859375, 0.019542015075683594, 0.019335168838500977, 0.019180543899536134, 0.019479551315307618, 0.019132415771484376, 0.01924198341369629, 0.01905971145629883, 0.019186752319335938, 0.01917740821838379, 0.019082239151000976, 0.019180543899536134, 0.019131391525268555, 0.019269632339477538, 0.019139583587646485, 0.01920204734802246, 0.019152896881103516, 0.019142656326293944, 0.019191808700561523, 0.019122175216674805, 0.019176448822021484, 0.019168256759643554, 0.019155967712402345, 0.019195903778076173, 0.019148799896240236, 0.019146751403808594, 0.019166208267211913, 0.01924300765991211, 0.01929523277282715, 0.019105791091918945, 0.01940787124633789, 0.04074905776977539, 0.01918060874938965, 0.01916102409362793, 0.018911231994628908, 0.018923519134521484, 0.02000588798522949, 0.019154943466186524, 0.019163135528564454, 0.01923481559753418, 0.019096576690673828, 0.019186687469482423, 0.019180543899536134, 0.01922150421142578, 0.019098623275756836, 0.018791423797607423, 0.018991104125976564, 0.019582975387573243, 0.01945497512817383, 0.01908940887451172, 0.018890752792358398, 0.018949119567871094, 0.019345407485961915, 0.019354623794555666, 0.019177471160888672, 0.019174400329589843, 0.01924710464477539, 0.01919692802429199, 0.01924608039855957, 0.0190382080078125, 0.01921433639526367, 0.019096576690673828, 0.019215360641479492, 0.01926144027709961, 0.019162176132202148, 0.0191343994140625, 0.01920512008666992, 0.01962393569946289, 0.01922150421142578, 0.019074047088623047, 0.019091455459594727, 0.01901158332824707, 0.019084287643432618, 0.019090431213378906, 0.01907302474975586, 0.01903001594543457, 0.01923276710510254, 0.019330047607421876, 0.019177471160888672, 0.019079168319702147, 0.019034143447875976, 0.01910780715942383, 0.019106815338134766, 0.019006464004516603, 0.019153919219970703, 0.019132415771484376, 0.019169279098510742, 0.019096576690673828, 0.01905971145629883, 0.019191808700561523, 0.019149824142456053, 0.019177471160888672, 0.01904025650024414, 0.01903923225402832, 0.04016230392456055, 0.018553855895996094, 0.019129344940185547, 0.01906483268737793, 0.019340288162231444, 0.01925734329223633, 0.019293184280395507, 0.019136512756347656, 0.0192174072265625, 0.019418111801147463, 0.019323904037475585, 0.01922047996520996, 0.019090431213378906, 0.019275808334350587, 0.019354591369628905, 0.019388416290283202, 0.019163135528564454, 0.019185663223266602, 0.019168256759643554, 0.01920204734802246, 0.019187711715698243, 0.019135488510131835, 0.01920102310180664, 0.019292160034179686, 0.01923788833618164, 0.019272703170776367, 0.019150848388671874, 0.01924710464477539, 0.01923993682861328, 0.019141632080078123, 0.019345439910888673, 0.019782623291015624, 0.019552255630493166, 0.019147775650024415, 0.019142816543579102, 0.019079008102416993, 0.01903308868408203, 0.019418111801147463, 0.0192174072265625, 0.019198976516723632, 0.019176448822021484, 0.019112960815429687, 0.01930342483520508, 0.01922662353515625, 0.019190784454345702, 0.01914681625366211, 0.019561439514160155, 0.019230688095092773, 0.01909350395202637, 0.019166208267211913, 0.019145727157592773, 0.01945907211303711, 0.019698688507080078, 0.019168256759643554, 0.01926144027709961, 0.019184640884399414, 0.019570688247680663, 0.01919385528564453, 0.019148799896240236, 0.01919385528564453, 0.019147775650024415, 0.019195903778076173, 0.019122175216674805, 0.04099071884155273, 0.01918976020812988, 0.019175424575805664, 0.019151872634887695, 0.019184640884399414, 0.01920921516418457, 0.01914886474609375, 0.019173311233520507, 0.019143680572509765, 0.01926348876953125, 0.019041471481323242, 0.019021631240844727, 0.01909350395202637, 0.01887846374511719, 0.018938880920410156, 0.018933759689331055, 0.01904128074645996, 0.01904025650024414, 0.018807872772216797, 0.0189736328125, 0.01885798454284668, 0.0190515193939209, 0.018919488906860352, 0.019041215896606446, 0.01901260757446289, 0.019072000503540038, 0.018918399810791017, 0.018931711196899414, 0.01896873664855957, 0.019012575149536134, 0.01906163215637207, 0.018932735443115235, 0.018932735443115235, 0.018969600677490234, 0.01916431999206543, 0.019007328033447266, 0.018977792739868164, 0.01900851249694824, 0.0190086727142334, 0.018988895416259765, 0.018936832427978514, 0.019110912322998046, 0.018998271942138673, 0.018966527938842775, 0.018942975997924806, 0.018918399810791017, 0.018948095321655273, 0.018959360122680666, 0.019208192825317383, 0.01921753692626953, 0.018999231338500976, 0.020249536514282226, 0.01944780731201172, 0.01926144027709961, 0.019272703170776367, 0.019332096099853514, 0.019170303344726563, 0.01906483268737793, 0.01841766357421875, 0.018499584197998048, 0.01841868782043457, 0.018511871337890624, 0.01847500801086426, 0.040215553283691405, 0.018527231216430663, 0.018514944076538087, 0.01849241638183594, 0.01842790412902832, 0.01846784019470215, 0.018454559326171877, 0.019949535369873046, 0.019783679962158202, 0.01948467254638672, 0.019177471160888672, 0.01936591911315918, 0.01913750457763672, 0.019105791091918945, 0.01844326400756836, 0.018493440628051756, 0.018453504562377928, 0.018516992568969725, 0.018497535705566406, 0.018963455200195312, 0.019112960815429687, 0.019363840103149413, 0.019133440017700197, 0.01922764778137207, 0.018481151580810547, 0.018530303955078126, 0.018566144943237304, 0.018463743209838866, 0.019121152877807617, 0.01922764778137207, 0.019156991958618166, 0.019095552444458007, 0.019149824142456053, 0.01943654441833496, 0.01918156814575195, 0.019109888076782225, 0.019160064697265625, 0.019172351837158205, 0.019143680572509765, 0.0192491512298584, 0.019308544158935546, 0.019148799896240236, 0.01921331214904785, 0.01920614433288574, 0.019192832946777344, 0.019138559341430664, 0.019216384887695313, 0.019144704818725586, 0.019216384887695313, 0.019173376083374022, 0.019117055892944337, 0.019161088943481445, 0.019106815338134766, 0.019149824142456053, 0.019188735961914064, 0.019182592391967773, 0.019319807052612305, 0.01918156814575195, 0.019169279098510742, 0.019153919219970703, 0.019122175216674805, 0.019344383239746094, 0.019308544158935546, 0.040787967681884765, 0.01918156814575195, 0.019183616638183593, 0.019006464004516603, 0.01922764778137207, 0.019215360641479492, 0.019294208526611328, 0.01925836753845215, 0.019186687469482423, 0.019216384887695313, 0.019138559341430664, 0.01923379135131836, 0.019155967712402345, 0.019203071594238282, 0.019191808700561523, 0.019105791091918945, 0.01923788833618164, 0.01926553535461426, 0.019198976516723632, 0.019203071594238282, 0.01921023941040039, 0.019508224487304687, 0.019418111801147463, 0.01926246452331543, 0.019134464263916014, 0.019494911193847657, 0.02004684829711914, 0.019304447174072266, 0.019109888076782225, 0.019138559341430664, 0.019304447174072266, 0.019184640884399414, 0.019105791091918945, 0.019187711715698243, 0.019095552444458007, 0.019382272720336914, 0.019153919219970703, 0.019195903778076173, 0.01924505615234375, 0.019096576690673828, 0.01903308868408203, 0.018541568756103514, 0.0184770565032959, 0.01864806365966797, 0.01918976020812988, 0.019088384628295898, 0.019109888076782225, 0.019489791870117186, 0.018503679275512695, 0.019104768753051758, 0.019088384628295898, 0.019169279098510742, 0.01905766487121582, 0.019123199462890626, 0.019313663482666017, 0.019078144073486326, 0.01926655960083008, 0.019108863830566408, 0.018490367889404297, 0.0189040641784668, 0.019147775650024415, 0.018555904388427736, 0.019115007400512696, 0.04064665603637695, 0.01918976020812988, 0.019009536743164062, 0.01927065658569336, 0.019288063049316406, 0.019151935577392577, 0.01928390312194824, 0.019131391525268555, 0.019277824401855468, 0.019309568405151366, 0.019096576690673828, 0.019070976257324217, 0.01906585693359375, 0.01906073570251465, 0.01900441551208496, 0.019084287643432618, 0.018963455200195312, 0.01843097686767578, 0.01942630386352539, 0.01925836753845215, 0.01903615951538086, 0.01841049575805664, 0.01844534492492676, 0.018623455047607422, 0.019499008178710937, 0.01909760093688965, 0.01904742431640625, 0.019096576690673828, 0.01903615951538086, 0.019006464004516603, 0.01909760093688965, 0.019084287643432618, 0.018531328201293946, 0.01843404769897461, 0.01836851119995117, 0.018381824493408205, 0.01836031913757324, 0.018480159759521483, 0.018365440368652345, 0.018444255828857423, 0.01842278480529785, 0.01844326400756836, 0.01846272087097168, 0.018533376693725585, 0.01845964813232422, 0.018404352188110353, 0.01836755180358887, 0.018269119262695314, 0.01826617622375488, 0.018185152053833007, 0.01820467185974121, 0.0183767032623291, 0.01844326400756836, 0.01845248031616211, 0.018513919830322266, 0.01845145606994629, 0.018456575393676757, 0.01848320007324219, 0.01852012825012207, 0.01892038345336914, 0.02003865623474121, 0.019587072372436523, 0.019283967971801756, 0.041128959655761715, 0.019110912322998046, 0.018928640365600585, 0.019177471160888672, 0.018961536407470704, 0.018986879348754884, 0.019083263397216797, 0.019188735961914064, 0.019755008697509766, 0.019857440948486328, 0.01946006393432617, 0.01930342483520508, 0.01927577590942383, 0.019155967712402345, 0.01923276710510254, 0.019219455718994142, 0.019170303344726563, 0.019318784713745117, 0.019143680572509765, 0.01927168083190918, 0.01924710464477539, 0.01922051239013672, 0.019190752029418945, 0.019117055892944337, 0.019200000762939453, 0.01924198341369629, 0.01940995216369629, 0.019125215530395506, 0.01920614433288574, 0.019207168579101562, 0.019102720260620116, 0.019139583587646485, 0.019140607833862306, 0.019056640625, 0.019198976516723632, 0.019171327590942384, 0.01921023941040039, 0.019190784454345702, 0.019149824142456053, 0.01925836753845215, 0.019140607833862306, 0.019187711715698243, 0.0191016960144043, 0.019195968627929688, 0.01909446334838867, 0.01915088081359863, 0.01912931251525879, 0.019137535095214844, 0.019191808700561523, 0.019115007400512696, 0.01920512008666992, 0.019166208267211913, 0.019176448822021484, 0.019146751403808594, 0.019144704818725586, 0.0192225284576416, 0.019155967712402345, 0.019211263656616212, 0.019098623275756836, 0.019151872634887695, 0.019084287643432618, 0.019200000762939453, 0.01909350395202637]",tokens/s,51.665584138414516,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,togethercomputer/RedPajama-INCITE-Base-3B-v1,togethercomputer/RedPajama-INCITE-Base-3B-v1,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1290.743808,6833.045504,0.0,6186.5984,6020.355072,s,10,0.3598095359802245,0.03598095359802246,0.000327616359819926,0.03585665512084961,0.03615181999206543,0.03654773464202881,0.03686446636199951,"[0.03694364929199219, 0.03587587356567383, 0.03584486389160156, 0.036063838958740234, 0.035855487823486326, 0.03585529708862305, 0.03582822418212891, 0.03581388854980469, 0.03587059020996094, 0.03585782241821289]",tokens/s,7114.875354889703,kWh,4.261281898172258e-07,2.3349698723394658e-07,2.2525587564819797e-06,2.9121839335331525e-06,tokens/kWh,87906535.38473883,MB,1290.743808,6833.045504,0.0,6186.5984,6150.076416,s,10,14.46971630859375,1.446971630859375,0.025161149540616375,1.4332265014648438,1.4859286865234376,1.488804248046875,1.491104697265625,"[1.4852896728515625, 1.4916798095703125, 1.4388095703125, 1.4768404541015625, 1.426799072265625, 1.4347098388671875, 1.4234737548828125, 1.429978515625, 1.4303924560546875, 1.4317431640625]",tokens/s,43.53920882511255,kWh,1.7033960242017947e-05,9.332253111065811e-06,5.951512954651741e-05,8.588134289960118e-05,tokens/kWh,733570.271178102,,s,629,14.768255001068102,0.023478942767993824,0.004322459082862367,0.022701055526733398,0.02381783103942871,0.023979827880859374,0.05897809753417969,"[0.02389708709716797, 0.023614463806152345, 0.02374143981933594, 0.023562240600585937, 0.023763967514038087, 0.02350284767150879, 0.02431283187866211, 0.024577024459838868, 0.02473472023010254, 0.02489753532409668, 0.024993791580200195, 0.024582143783569335, 0.024782880783081055, 0.024524768829345702, 0.024769535064697267, 0.023834623336791993, 0.024599552154541016, 0.023565311431884766, 0.023340032577514647, 0.023840768814086914, 0.02469068717956543, 0.02433433532714844, 0.023974912643432617, 0.023983104705810547, 0.023916543960571288, 0.023834623336791993, 0.02413670349121094, 0.024687616348266602, 0.024155136108398437, 0.02381520080566406, 0.023843807220458986, 0.023802879333496094, 0.023795711517333985, 0.023799808502197265, 0.022724607467651366, 0.02269593620300293, 0.02265395164489746, 0.022740991592407226, 0.022673408508300782, 0.022766592025756836, 0.02269696044921875, 0.022793216705322264, 0.0226693115234375, 0.022740991592407226, 0.022781951904296875, 0.022732831954956054, 0.02254742431640625, 0.022795263290405272, 0.022791168212890626, 0.02272870445251465, 0.022777856826782225, 0.022759424209594727, 0.023467071533203127, 0.023888864517211915, 0.024092639923095703, 0.022802463531494142, 0.022794240951538085, 0.022647775650024415, 0.02269491195678711, 0.022641664505004884, 0.022601728439331056, 0.023748607635498048, 0.058893310546875, 0.0224901123046875, 0.022862848281860353, 0.023766016006469725, 0.02369024085998535, 0.023801855087280274, 0.02391551971435547, 0.023760896682739258, 0.023767040252685546, 0.0239554557800293, 0.023773183822631837, 0.023768064498901367, 0.023706623077392578, 0.023770143508911133, 0.02378646469116211, 0.023682048797607422, 0.023773183822631837, 0.023766016006469725, 0.023764991760253908, 0.023569408416748046, 0.023801855087280274, 0.023739391326904297, 0.023744512557983398, 0.023160831451416015, 0.02384588813781738, 0.023434240341186522, 0.02306662368774414, 0.023752704620361328, 0.023441408157348635, 0.023702527999877928, 0.023707679748535156, 0.02371887969970703, 0.0237076473236084, 0.02368511962890625, 0.02373222351074219, 0.02369740867614746, 0.023764991760253908, 0.023750656127929686, 0.023526432037353516, 0.02338915252685547, 0.023546880722045898, 0.02371379280090332, 0.023748672485351563, 0.023665599822998047, 0.023837696075439452, 0.023817216873168946, 0.023798784255981444, 0.023771135330200196, 0.023792640686035156, 0.023777280807495117, 0.023834623336791993, 0.023790592193603514, 0.023829504013061522, 0.023880704879760743, 0.02392780876159668, 0.023774208068847655, 0.023773183822631837, 0.02374553680419922, 0.023805952072143553, 0.023706623077392578, 0.023802879333496094, 0.023775232315063476, 0.023824384689331055, 0.05904383850097656, 0.02270310401916504, 0.022656000137329102, 0.022617151260375976, 0.022664127349853517, 0.022697984695434572, 0.02263859176635742, 0.02266726493835449, 0.022665216445922853, 0.022642688751220705, 0.022649856567382814, 0.022648832321166993, 0.022595584869384764, 0.022683679580688478, 0.022656991958618165, 0.02267852783203125, 0.02262015914916992, 0.02268671989440918, 0.02274508857727051, 0.02271129608154297, 0.02264371109008789, 0.022701055526733398, 0.022739967346191405, 0.02273587226867676, 0.02270515251159668, 0.02265292739868164, 0.02266316795349121, 0.02272051239013672, 0.02266828727722168, 0.022601728439331056, 0.022647808074951172, 0.02268057632446289, 0.022675455093383787, 0.022626304626464845, 0.022738943099975584, 0.022664192199707032, 0.022730752944946288, 0.02270515251159668, 0.022700031280517577, 0.02314035224914551, 0.02368307113647461, 0.02370150375366211, 0.023736320495605468, 0.023764991760253908, 0.023828479766845705, 0.023770111083984375, 0.023796735763549806, 0.023827455520629884, 0.024102912902832032, 0.023197696685791015, 0.022767616271972657, 0.022635520935058592, 0.022733823776245117, 0.022602752685546876, 0.022701055526733398, 0.022577152252197266, 0.02265088081359863, 0.02264371109008789, 0.022401023864746093, 0.022349855422973634, 0.022381536483764647, 0.022433792114257813, 0.022598655700683593, 0.059652095794677736, 0.022591487884521484, 0.022658048629760744, 0.022368255615234374, 0.02244915199279785, 0.022794240951538085, 0.022759424209594727, 0.02241433525085449, 0.022368255615234374, 0.022437887191772463, 0.022311935424804686, 0.022552576065063477, 0.02365951919555664, 0.02346188735961914, 0.02350387191772461, 0.023454816818237304, 0.023383968353271483, 0.02352639961242676, 0.02348236846923828, 0.023430143356323242, 0.023377920150756838, 0.023352319717407227, 0.02334623908996582, 0.02340345573425293, 0.023405567169189453, 0.023669759750366212, 0.02351411247253418, 0.02389504051208496, 0.023983104705810547, 0.023820287704467775, 0.02390220832824707, 0.024028160095214843, 0.023669759750366212, 0.024034303665161134, 0.023782400131225585, 0.023734272003173826, 0.023827455520629884, 0.023744543075561522, 0.023795679092407228, 0.02386534309387207, 0.023770111083984375, 0.023706623077392578, 0.02387046432495117, 0.023718912124633788, 0.023916543960571288, 0.023763967514038087, 0.023718912124633788, 0.023747583389282227, 0.023932928085327147, 0.023440383911132814, 0.02370150375366211, 0.023833599090576172, 0.023767040252685546, 0.023734272003173826, 0.02394726371765137, 0.023848960876464844, 0.023773183822631837, 0.023867391586303712, 0.02371788787841797, 0.023813119888305666, 0.02262835121154785, 0.022700031280517577, 0.022624256134033204, 0.059025409698486325, 0.02266214370727539, 0.022606847763061523, 0.02266111946105957, 0.0227061767578125, 0.022656000137329102, 0.022632448196411133, 0.02262937545776367, 0.022580223083496095, 0.02266726493835449, 0.022640640258789063, 0.022567935943603516, 0.022656000137329102, 0.022571008682250978, 0.022601728439331056, 0.022615039825439453, 0.022595584869384764, 0.022593536376953126, 0.02268876838684082, 0.022600704193115235, 0.02264371109008789, 0.022650943756103516, 0.022594495773315428, 0.022609920501708985, 0.02267852783203125, 0.022647808074951172, 0.022588415145874022, 0.022768640518188478, 0.022725631713867187, 0.02265497589111328, 0.02271334457397461, 0.02264678382873535, 0.02265193557739258, 0.0225382080078125, 0.022714368820190428, 0.022563840866088865, 0.022764543533325195, 0.022610944747924806, 0.02263654327392578, 0.02263654327392578, 0.02268671989440918, 0.02262015914916992, 0.02272870445251465, 0.022617088317871094, 0.02267136001586914, 0.022723583221435546, 0.02251571273803711, 0.022607872009277344, 0.022706239700317384, 0.022386751174926757, 0.02251968002319336, 0.0223242244720459, 0.022586368560791017, 0.022722560882568358, 0.022736000061035155, 0.022624128341674803, 0.0227061767578125, 0.022631423950195313, 0.02263654327392578, 0.02267238426208496, 0.02270412826538086, 0.02261203193664551, 0.022956991195678712, 0.05908889770507812, 0.02271129608154297, 0.022689792633056642, 0.02323967933654785, 0.022937599182128905, 0.022674495697021485, 0.022642623901367186, 0.022681600570678712, 0.022618112564086915, 0.022579200744628908, 0.02243891143798828, 0.02248806381225586, 0.024610815048217775, 0.023311359405517578, 0.02262835121154785, 0.02267136001586914, 0.022648832321166993, 0.022658048629760744, 0.022709247589111328, 0.022777856826782225, 0.022740991592407226, 0.02262937545776367, 0.022700031280517577, 0.02269388771057129, 0.02284339141845703, 0.022977535247802734, 0.022939647674560547, 0.023004159927368165, 0.022811647415161132, 0.023335935592651368, 0.022734848022460938, 0.02264473533630371, 0.022912000656127928, 0.022658111572265625, 0.022656959533691408, 0.02270412826538086, 0.022838272094726563, 0.022748159408569335, 0.0227194881439209, 0.022610944747924806, 0.022732799530029296, 0.022749183654785156, 0.022897663116455077, 0.022675455093383787, 0.022579200744628908, 0.022684671401977538, 0.02269388771057129, 0.0226693115234375, 0.022631423950195313, 0.022616064071655274, 0.022631423950195313, 0.022664192199707032, 0.022658048629760744, 0.022689792633056642, 0.022734848022460938, 0.0227061767578125, 0.022734848022460938, 0.02266316795349121, 0.02274508857727051, 0.022595584869384764, 0.022632448196411133, 0.022787071228027343, 0.0227194881439209, 0.05901107025146484, 0.022639616012573242, 0.022617088317871094, 0.022618112564086915, 0.02367180824279785, 0.022838272094726563, 0.022619136810302733, 0.022578176498413087, 0.022633472442626954, 0.022692863464355468, 0.02262118339538574, 0.02262937545776367, 0.02267238426208496, 0.022578176498413087, 0.022632448196411133, 0.02262015914916992, 0.022603776931762694, 0.022607872009277344, 0.02271232032775879, 0.02264575958251953, 0.02264473533630371, 0.022657024383544923, 0.022573055267333983, 0.02264678382873535, 0.02260479927062988, 0.022573055267333983, 0.02268262481689453, 0.02272051239013672, 0.02268569564819336, 0.022625280380249024, 0.022618112564086915, 0.02271334457397461, 0.022665216445922853, 0.022563840866088865, 0.022607872009277344, 0.022573055267333983, 0.022673408508300782, 0.022676479339599608, 0.02264678382873535, 0.02263859176635742, 0.022674432754516603, 0.02265190315246582, 0.02247065544128418, 0.02254243278503418, 0.02230672073364258, 0.022425600051879883, 0.022452224731445314, 0.022297632217407225, 0.02271331214904785, 0.022339584350585938, 0.02259459114074707, 0.022308832168579103, 0.022419456481933595, 0.022495231628417968, 0.022432767868041992, 0.022329343795776366, 0.02243071937561035, 0.02230067253112793, 0.02249932861328125, 0.022351871490478514, 0.0224399356842041, 0.022395904541015626, 0.02240716743469238, 0.05876838302612305, 0.0224768009185791, 0.022375423431396483, 0.02249830436706543, 0.022388832092285156, 0.02228316879272461, 0.022397951126098634, 0.02243395233154297, 0.0223702392578125, 0.02237225532531738, 0.02307174491882324, 0.02408448028564453, 0.0230645751953125, 0.022642688751220705, 0.022345727920532226, 0.02264678382873535, 0.022676479339599608, 0.022535167694091796, 0.022792192459106447, 0.022700128555297853, 0.02259343910217285, 0.022675455093383787, 0.022549663543701172, 0.02256368064880371, 0.022648832321166993, 0.022687744140625, 0.02266726493835449, 0.02271232032775879, 0.022973440170288087, 0.022657024383544923, 0.022641664505004884, 0.022684671401977538, 0.02267852783203125, 0.02265292739868164, 0.022749183654785156, 0.02265497589111328, 0.022788095474243163, 0.0226375675201416, 0.022625312805175782, 0.022774784088134766, 0.02274505615234375, 0.022564863204956053, 0.022730752944946288, 0.02287718391418457, 0.02272768020629883, 0.022845439910888672, 0.022715391159057616, 0.02269900894165039, 0.022658048629760744, 0.022701055526733398, 0.022770687103271483, 0.022738943099975584, 0.022764543533325195, 0.02266111946105957, 0.022725631713867187, 0.0231014404296875, 0.02266111946105957, 0.022709247589111328, 0.022790143966674805, 0.022674432754516603, 0.022958080291748048, 0.022675455093383787, 0.022807552337646485, 0.05912063980102539, 0.0227061767578125, 0.022684671401977538, 0.02268671989440918, 0.022595584869384764, 0.022754304885864256, 0.022627328872680662, 0.022742015838623047, 0.022815807342529297, 0.022767551422119142, 0.022714368820190428, 0.022640640258789063, 0.022809600830078124, 0.02269081687927246, 0.022593536376953126, 0.0226693115234375, 0.02267750358581543, 0.022600704193115235, 0.022769664764404295, 0.022675455093383787, 0.02265497589111328, 0.023618560791015625, 0.022724607467651366, 0.022915071487426757, 0.022746112823486327, 0.022640640258789063, 0.02264271926879883, 0.022668256759643554, 0.022647872924804687, 0.022655935287475587, 0.0227194881439209, 0.02266316795349121, 0.022640640258789063, 0.02266111946105957, 0.022855680465698244, 0.02263654327392578, 0.02271027183532715, 0.022592512130737305, 0.022724607467651366, 0.022576160430908203, 0.02271023941040039, 0.02265292739868164, 0.02262937545776367, 0.022642688751220705, 0.0226375675201416, 0.022594560623168947, 0.022689792633056642, 0.022674432754516603, 0.02265395164489746, 0.0226375675201416, 0.022697984695434572, 0.022625280380249024, 0.022797311782836914, 0.02282598304748535, 0.02264371109008789, 0.02265497589111328, 0.0227194881439209, 0.022623231887817383, 0.022658048629760744, 0.022565887451171874, 0.022601728439331056, 0.022665216445922853, 0.02266316795349121, 0.05907763290405273, 0.02271334457397461, 0.022686752319335937, 0.022666208267211913, 0.02265907287597656, 0.022639616012573242, 0.022631423950195313, 0.022611967086791994, 0.02266009521484375, 0.022798336029052735, 0.022822912216186524, 0.02272870445251465, 0.022731775283813475, 0.02269081687927246, 0.022754304885864256, 0.022575103759765625, 0.022715391159057616, 0.022684671401977538, 0.022606847763061523, 0.022540288925170897, 0.022549503326416014, 0.0225864315032959, 0.022452159881591795, 0.022485055923461915, 0.02272863960266113, 0.022777856826782225, 0.02266422462463379, 0.022675424575805663, 0.02269900894165039, 0.02260479927062988, 0.022625280380249024, 0.02270412826538086, 0.0226693115234375, 0.022657024383544923, 0.022723583221435546, 0.022800416946411134, 0.022975456237792968, 0.02269593620300293, 0.022673408508300782, 0.022716415405273437, 0.02271129608154297, 0.02272972869873047, 0.022573055267333983, 0.022741024017333984, 0.022423519134521484, 0.022541311264038084, 0.022618112564086915, 0.02265088081359863, 0.02248192024230957, 0.02287615966796875, 0.023621631622314454, 0.023591936111450194, 0.02302463912963867, 0.023005184173583985, 0.023147520065307618, 0.022754304885864256, 0.02265497589111328, 0.022691839218139647, 0.02275225639343262, 0.022615039825439453, 0.022665216445922853, 0.02267238426208496, 0.022602783203125]",tokens/s,42.59135557684422,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,meta-llama/Meta-Llama-3-8B,meta-llama/Meta-Llama-3-8B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.1,,,,MB,1282.953216,17201.364992,0.0,16554.917888,16436.726784,s,10,0.8367364196777345,0.08367364196777345,0.0004923809547797837,0.08345528030395508,0.0842702606201172,0.08460759582519531,0.08487746398925781,"[0.08494493103027344, 0.08346476745605469, 0.08351712036132812, 0.08343762969970703, 0.08329440307617188, 0.08344579315185546, 0.0833201904296875, 0.08337798309326172, 0.08373830413818359, 0.08419529724121094]",tokens/s,3059.5058847635355,kWh,9.892178782158429e-07,5.420425019838148e-07,6.7589799442166245e-06,8.290240324416282e-06,tokens/kWh,30879683.818819214,MB,1282.953216,17201.364992,0.0,16554.917888,16436.729344,s,10,24.17886572265625,2.4178865722656253,0.0025432229852819553,2.4171455078125,2.418894140625,2.4220992919921875,2.4246634130859377,"[2.425304443359375, 2.41773876953125, 2.41712841796875, 2.41661572265625, 2.41716259765625, 2.418181884765625, 2.41741455078125, 2.416422607421875, 2.416042724609375, 2.41685400390625]",tokens/s,26.05581284194291,kWh,2.8481320403516297e-05,1.5608483978958246e-05,0.00012753740527058415,0.00017162720965305868,tokens/kWh,367074.6621549891,,s,629,24.89124155044554,0.03957272106589119,0.010000622787143695,0.03834368133544922,0.038423565673828125,0.03855401611328125,0.12240008026123048,"[0.039119873046875, 0.039169025421142575, 0.039103488922119144, 0.03913523101806641, 0.03925196838378906, 0.03905843353271484, 0.03910246276855469, 0.03912089538574219, 0.03907379150390625, 0.03880448150634765, 0.03872256088256836, 0.03891404724121094, 0.03848191833496094, 0.038319103240966795, 0.03842559814453125, 0.03830374526977539, 0.038338558197021484, 0.03829555130004883, 0.03829145431518555, 0.038307838439941407, 0.038317054748535154, 0.03832217788696289, 0.03834470367431641, 0.038265857696533206, 0.03828224182128906, 0.03828736114501953, 0.03837849426269531, 0.0383375358581543, 0.03832940673828125, 0.03831084823608399, 0.03832831954956055, 0.03835903930664063, 0.03832729721069336, 0.03834368133544922, 0.03836620712280273, 0.03831398391723633, 0.03831398391723633, 0.038296577453613284, 0.03834265518188477, 0.03828736114501953, 0.03833446502685547, 0.03836006546020508, 0.038392833709716793, 0.03834777450561523, 0.038346752166748044, 0.03831398391723633, 0.03839078521728516, 0.038332416534423826, 0.038341663360595704, 0.03833750534057617, 0.038353919982910156, 0.03830476760864258, 0.03837747192382813, 0.03833446502685547, 0.038332416534423826, 0.0383559684753418, 0.0383375358581543, 0.038353919982910156, 0.038400001525878906, 0.03837747192382813, 0.03842355346679688, 0.03834265518188477, 0.1224796142578125, 0.03840204620361328, 0.03835801696777344, 0.03839897537231445, 0.03863347244262695, 0.038959102630615236, 0.03842873764038086, 0.038403072357177735, 0.03841939163208008, 0.03841126251220703, 0.03829555130004883, 0.038296577453613284, 0.03829555130004883, 0.03833651351928711, 0.03830169677734375, 0.03833651351928711, 0.03832115173339844, 0.03836108779907227, 0.038332416534423826, 0.03839897537231445, 0.03829043197631836, 0.038365184783935545, 0.03832012939453125, 0.038317054748535154, 0.03830886459350586, 0.038319103240966795, 0.038338558197021484, 0.03848806381225586, 0.03868672180175781, 0.038458366394042966, 0.03832320022583008, 0.038368255615234374, 0.038319103240966795, 0.03830681610107422, 0.03828940963745117, 0.03834777450561523, 0.03830886459350586, 0.038340606689453126, 0.03829043197631836, 0.03830681610107422, 0.03831500625610352, 0.03830886459350586, 0.03834982299804687, 0.03832320022583008, 0.038329345703125, 0.03830681610107422, 0.03831097412109375, 0.038374336242675784, 0.038365184783935545, 0.038370304107666016, 0.038319103240966795, 0.03834470367431641, 0.03837542343139649, 0.038346752166748044, 0.03833651351928711, 0.038384639739990234, 0.03834780883789062, 0.03834672164916992, 0.03834470367431641, 0.038392833709716793, 0.03835903930664063, 0.038384639739990234, 0.038332416534423826, 0.1225185317993164, 0.03834163284301758, 0.03833139038085937, 0.03838566589355469, 0.038312961578369144, 0.03834163284301758, 0.03832217788696289, 0.038371326446533204, 0.038286334991455076, 0.03836415863037109, 0.038299648284912106, 0.03834777450561523, 0.03828326416015625, 0.038351871490478515, 0.03829452896118164, 0.03836620712280273, 0.0383559684753418, 0.03848089599609375, 0.03849728012084961, 0.038317054748535154, 0.03830681610107422, 0.038365184783935545, 0.03838771057128906, 0.03835289764404297, 0.038296577453613284, 0.038330368041992184, 0.03834470367431641, 0.03834982299804687, 0.03830374526977539, 0.03834265518188477, 0.03834470367431641, 0.038324222564697266, 0.03832115173339844, 0.03834368133544922, 0.03837849426269531, 0.038545406341552735, 0.03833651351928711, 0.03836006546020508, 0.038319103240966795, 0.0383263053894043, 0.03833750534057617, 0.038346782684326175, 0.038442977905273436, 0.03855052947998047, 0.03838873672485352, 0.038386688232421876, 0.03832524871826172, 0.03834470367431641, 0.03832115173339844, 0.03836620712280273, 0.038330368041992184, 0.038403072357177735, 0.038348800659179685, 0.038365184783935545, 0.038329345703125, 0.03835801696777344, 0.03832831954956055, 0.038354942321777344, 0.038346782684326175, 0.03843379211425781, 0.0383927993774414, 0.03836723327636719, 0.0383559684753418, 0.12249190521240234, 0.03837849426269531, 0.03837542343139649, 0.038365184783935545, 0.038419456481933595, 0.03847987365722656, 0.038381568908691405, 0.03832524871826172, 0.03833446502685547, 0.038346752166748044, 0.03838873672485352, 0.03842867279052734, 0.038368255615234374, 0.03832217788696289, 0.03839078521728516, 0.038406143188476564, 0.038307838439941407, 0.03837235260009766, 0.03836108779907227, 0.03830374526977539, 0.03836620712280273, 0.03845119857788086, 0.03833958435058594, 0.038362110137939456, 0.038302753448486326, 0.0383139533996582, 0.038330368041992184, 0.0383559684753418, 0.038302719116210936, 0.038345729827880856, 0.03828224182128906, 0.03836620712280273, 0.03832524871826172, 0.03834163284301758, 0.038307838439941407, 0.03836415863037109, 0.03832012939453125, 0.03835903930664063, 0.03831811141967773, 0.03833135986328125, 0.03829145431518555, 0.038362110137939456, 0.03833446502685547, 0.03833446502685547, 0.03830886459350586, 0.0383631362915039, 0.03830476760864258, 0.03837644958496094, 0.03834163284301758, 0.0383559684753418, 0.03832524871826172, 0.03833446502685547, 0.03831500625610352, 0.03836620712280273, 0.03828121566772461, 0.03836620712280273, 0.038335487365722655, 0.03836620712280273, 0.03829967880249024, 0.03839587020874023, 0.03832217788696289, 0.03833958435058594, 0.038330368041992184, 0.1223004150390625, 0.03837033462524414, 0.03833647918701172, 0.03834163284301758, 0.038362110137939456, 0.03837235260009766, 0.03834163284301758, 0.038386688232421876, 0.038370304107666016, 0.03850342559814453, 0.038329345703125, 0.03835084915161133, 0.03829350280761719, 0.038338558197021484, 0.038345729827880856, 0.03834982299804687, 0.03833343887329101, 0.03837849426269531, 0.038312961578369144, 0.03830988693237305, 0.0382740478515625, 0.03834163284301758, 0.03835084915161133, 0.03833958435058594, 0.03832627105712891, 0.038316032409667966, 0.03829452896118164, 0.0383744010925293, 0.03842047882080078, 0.038346752166748044, 0.03835903930664063, 0.038296577453613284, 0.038286334991455076, 0.03832115173339844, 0.03833651351928711, 0.03839692687988281, 0.03838566589355469, 0.03832115173339844, 0.038296577453613284, 0.038296577453613284, 0.03832320022583008, 0.03832831954956055, 0.038406143188476564, 0.038324222564697266, 0.03830169677734375, 0.038330368041992184, 0.038346752166748044, 0.038389759063720705, 0.03842361450195313, 0.03836723327636719, 0.03836819076538086, 0.03836108779907227, 0.03837644958496094, 0.038365184783935545, 0.038414337158203124, 0.0384634895324707, 0.038406143188476564, 0.03839487838745117, 0.03840921783447265, 0.03836928176879883, 0.03838771057128906, 0.03852799987792969, 0.03838566589355469, 0.12250931549072265, 0.038370304107666016, 0.03839487838745117, 0.03840716934204102, 0.03856588745117188, 0.03874508666992187, 0.03842559814453125, 0.038356990814208985, 0.038302719116210936, 0.038335487365722655, 0.038329345703125, 0.03829862213134766, 0.03831196975708008, 0.03829756927490234, 0.03831500625610352, 0.03836108779907227, 0.03829862213134766, 0.03833958435058594, 0.03828940963745117, 0.038348800659179685, 0.038302719116210936, 0.038335487365722655, 0.03833446502685547, 0.0383109130859375, 0.03828121566772461, 0.03832524871826172, 0.038302719116210936, 0.038384639739990234, 0.038340606689453126, 0.03855567932128906, 0.03855152130126953, 0.03839078521728516, 0.038346752166748044, 0.03832115173339844, 0.038284286499023434, 0.03833958435058594, 0.038332416534423826, 0.03835084915161133, 0.03832320022583008, 0.038317054748535154, 0.038258689880371094, 0.03836006546020508, 0.03834265518188477, 0.038351871490478515, 0.038324222564697266, 0.03834368133544922, 0.03828224182128906, 0.038348800659179685, 0.0383744010925293, 0.03842764663696289, 0.038335487365722655, 0.03833446502685547, 0.03857408142089844, 0.03894892883300781, 0.038529983520507814, 0.038362110137939456, 0.038362110137939456, 0.0383631362915039, 0.03833958435058594, 0.03839487838745117, 0.03832012939453125, 0.03836624145507812, 0.038348766326904295, 0.12237725067138672, 0.038348800659179685, 0.03833033752441406, 0.038330368041992184, 0.03834368133544922, 0.038392833709716793, 0.03836108779907227, 0.03842559814453125, 0.03829660797119141, 0.03843888092041016, 0.03839590454101562, 0.03836108779907227, 0.03834163284301758, 0.03831398391723633, 0.038330368041992184, 0.038354942321777344, 0.038356990814208985, 0.03865497589111328, 0.03849318313598633, 0.03835903930664063, 0.038316032409667966, 0.03838259124755859, 0.0383744010925293, 0.0383559684753418, 0.038308895111083985, 0.03834262466430664, 0.038300670623779294, 0.03832115173339844, 0.03833756637573242, 0.0382852783203125, 0.03830476760864258, 0.03834576034545899, 0.038399967193603515, 0.038317054748535154, 0.038350910186767576, 0.03838252639770508, 0.03834368133544922, 0.03833958435058594, 0.038307838439941407, 0.03834163284301758, 0.03831398391723633, 0.03831193542480469, 0.03838566589355469, 0.0386693115234375, 0.03843379211425781, 0.038345729827880856, 0.03831398391723633, 0.038340606689453126, 0.038353919982910156, 0.0385054702758789, 0.03835084915161133, 0.03834368133544922, 0.03830886459350586, 0.03835084915161133, 0.03834470367431641, 0.03835903930664063, 0.03836006546020508, 0.03830579376220703, 0.03832217788696289, 0.0383744010925293, 0.03831808090209961, 0.03835801696777344, 0.03837542343139649, 0.1224089584350586, 0.03838873672485352, 0.038386688232421876, 0.038421504974365236, 0.03841228866577148, 0.038392833709716793, 0.03838054275512695, 0.03833446502685547, 0.03830579376220703, 0.038370304107666016, 0.03831193542480469, 0.03830169677734375, 0.03829555130004883, 0.038362110137939456, 0.03831500625610352, 0.03833343887329101, 0.03829350280761719, 0.03830988693237305, 0.038296577453613284, 0.03832217788696289, 0.038317054748535154, 0.038397953033447264, 0.03845017623901367, 0.038370304107666016, 0.03829145431518555, 0.0383559684753418, 0.03832012939453125, 0.038354942321777344, 0.03846758270263672, 0.038324222564697266, 0.03831500625610352, 0.038370304107666016, 0.038389759063720705, 0.0383375358581543, 0.03830579376220703, 0.03833958435058594, 0.03831398391723633, 0.038324222564697266, 0.03828224182128906, 0.038384639739990234, 0.03827814483642578, 0.03833446502685547, 0.03834368133544922, 0.038358047485351564, 0.03832112121582031, 0.03839897537231445, 0.03827814483642578, 0.03836108779907227, 0.038348800659179685, 0.038392833709716793, 0.03833446502685547, 0.03832831954956055, 0.038310943603515626, 0.03835184097290039, 0.0383559684753418, 0.03838873672485352, 0.038330368041992184, 0.038324222564697266, 0.0383232307434082, 0.03838460922241211, 0.0383375358581543, 0.03834368133544922, 0.03834163284301758, 0.12276326751708984, 0.038419456481933595, 0.03834982299804687, 0.03838566589355469, 0.03832012939453125, 0.03838566589355469, 0.03836928176879883, 0.038401023864746094, 0.03829555130004883, 0.038354942321777344, 0.038340606689453126, 0.038324222564697266, 0.038319103240966795, 0.03834368133544922, 0.038332416534423826, 0.03835289764404297, 0.03829248046875, 0.038348800659179685, 0.03830169677734375, 0.03830374526977539, 0.03832012939453125, 0.0383744010925293, 0.03829043197631836, 0.038316032409667966, 0.038296577453613284, 0.03833651351928711, 0.038330368041992184, 0.03837747192382813, 0.038370304107666016, 0.03830988693237305, 0.03832320022583008, 0.038362110137939456, 0.038391807556152346, 0.0383375358581543, 0.03828736114501953, 0.03832831954956055, 0.03831500625610352, 0.038324222564697266, 0.03830988693237305, 0.03832115173339844, 0.03833343887329101, 0.038316032409667966, 0.03836108779907227, 0.038307838439941407, 0.038348800659179685, 0.03835193634033203, 0.038287296295166015, 0.03830886459350586, 0.038307838439941407, 0.03837644958496094, 0.03836108779907227, 0.03836723327636719, 0.038307838439941407, 0.038362110137939456, 0.03840512084960938, 0.03834982299804687, 0.03835084915161133, 0.03832729721069336, 0.03832831954956055, 0.03835903930664063, 0.03833343887329101, 0.03836108779907227, 0.03836006546020508, 0.12329065704345703, 0.038396896362304686, 0.038384639739990234, 0.03834777450561523, 0.03840716934204102, 0.038624256134033204, 0.038419456481933595, 0.03839692687988281, 0.03827814483642578, 0.038330368041992184, 0.03828940963745117, 0.03833958435058594, 0.038302719116210936, 0.03834470367431641, 0.038300670623779294, 0.038317054748535154, 0.03829555130004883, 0.038319103240966795, 0.03828940963745117, 0.038299648284912106, 0.03831193542480469, 0.03832524871826172, 0.03832012939453125, 0.038381568908691405, 0.03831808090209961, 0.038351871490478515, 0.038389759063720705, 0.03832320022583008, 0.03831808090209961, 0.03842355346679688, 0.038468608856201174, 0.03839385604858398, 0.03833343887329101, 0.038403072357177735, 0.038302719116210936, 0.03836108779907227, 0.03832012939453125, 0.03837849426269531, 0.03828326416015625, 0.03830169677734375, 0.03829043197631836, 0.038338558197021484, 0.03833756637573242, 0.03834979248046875, 0.03829862213134766, 0.038408191680908206, 0.03827711868286133, 0.038397953033447264, 0.03844300842285156, 0.03839385604858398, 0.03828025436401367, 0.038348766326904295, 0.038348766326904295, 0.03842355346679688, 0.03833446502685547, 0.038381568908691405, 0.03832012939453125, 0.038384639739990234, 0.03833446502685547, 0.03839897537231445, 0.03833139038085937, 0.038338558197021484, 0.03836624145507812]",tokens/s,25.269932748241754,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen2-beta-14B,Qwen/Qwen2-beta-14B,cuda,0,42,,,True,,,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",qwen2,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-32B,Qwen/Qwen1.5-32B,cuda,0,42,,,True,,,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",qwen2,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,0,0,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/0/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669325af-61200470529587d96975dadd;60a074d7-72a6-4564-b6fa-55d0ca7d0695) Repository Not Found for url: https://huggingface.co/0/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: 0 is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/gpt-neo-125m,EleutherAI/gpt-neo-125m,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1211.916288,1084.751872,0.0,438.304768,353.979904,s,21,0.16843772792816164,0.008020844187055317,0.0002709089237060463,0.007919328212738037,0.008172672271728515,0.008186047554016113,0.008893605995178223,"[0.00907049560546875, 0.008113951683044433, 0.008172672271728515, 0.00807795238494873, 0.00808073616027832, 0.00810422420501709, 0.008186047554016113, 0.008108256340026855, 0.008135295867919922, 0.008065119743347168, 0.007819615840911865, 0.007802303791046143, 0.007833055973052979, 0.007879487991333008, 0.007810368061065674, 0.007919328212738037, 0.007848832130432129, 0.007841023921966552, 0.007876800060272218, 0.00781382417678833, 0.007878335952758789]",tokens/s,31916.839927292614,kWh,9.571627311055647e-08,5.2447969149308054e-08,2.3770098252511734e-07,3.8586522478498184e-07,tokens/kWh,663444082.4322859,MB,1211.916288,1084.751872,0.0,438.304768,363.871232,s,21,9.988530548095703,0.47564431181408107,0.008977244660993571,0.4732193908691406,0.4863353271484375,0.48660519409179687,0.48887653198242187,"[0.4894443664550781, 0.48660519409179687, 0.48212466430664064, 0.4860497741699219, 0.4863353271484375, 0.48628439331054685, 0.48400201416015626, 0.4829789428710938, 0.485130126953125, 0.4732193908691406, 0.4726200866699219, 0.4688006591796875, 0.4662080993652344, 0.4653125610351562, 0.4749071960449219, 0.46514337158203123, 0.46465264892578123, 0.46667947387695313, 0.46523233032226563, 0.46647604370117185, 0.47032388305664063]",tokens/s,132.4519150869722,kWh,5.448313286942431e-06,2.9844913679581617e-06,9.698405451201775e-06,1.813121010610237e-05,tokens/kWh,3474671.5542608085,,s,1322,10.139579388618463,0.007669878508788554,0.001044043459577593,0.007547391891479492,0.007727923154830932,0.007775228643417359,0.015717710523605346,"[0.008573951721191407, 0.008546303749084473, 0.008517663955688476, 0.00905519962310791, 0.007686143875122071, 0.007426047801971435, 0.007384064197540283, 0.007418879985809326, 0.007400447845458984, 0.007666687965393066, 0.007672832012176513, 0.0076984319686889645, 0.007664639949798584, 0.007715904235839844, 0.007704512119293213, 0.007732223987579345, 0.0076871681213378906, 0.007704576015472412, 0.007683072090148926, 0.007693312168121338, 0.007762944221496582, 0.007693312168121338, 0.007774208068847656, 0.007671807765960693, 0.007674880027770996, 0.007679999828338623, 0.007676928043365478, 0.007707647800445557, 0.0076902399063110355, 0.0077199358940124516, 0.007731200218200684, 0.007697408199310303, 0.007737343788146973, 0.007768064022064209, 0.007799808025360107, 0.0077916159629821775, 0.007697408199310303, 0.007681024074554443, 0.0076871681213378906, 0.007712768077850342, 0.007759871959686279, 0.007670783996582031, 0.007703551769256592, 0.007677951812744141, 0.007664639949798584, 0.0076943359375, 0.00780083179473877, 0.007712768077850342, 0.007774208068847656, 0.00774451208114624, 0.007669760227203369, 0.0076605439186096195, 0.00765235185623169, 0.0076984319686889645, 0.007651328086853027, 0.007697408199310303, 0.007675903797149658, 0.007682047843933106, 0.007749631881713868, 0.007665664196014404, 0.007674880027770996, 0.007855103969573975, 0.016287744522094725, 0.007701504230499268, 0.007701504230499268, 0.007684095859527588, 0.007665664196014404, 0.007685120105743408, 0.007683072090148926, 0.007642111778259277, 0.007667712211608887, 0.007699456214904785, 0.007705599784851074, 0.007819263935089112, 0.007762944221496582, 0.007714816093444824, 0.007715839862823487, 0.007745535850524903, 0.00780185604095459, 0.007766079902648926, 0.0077751679420471195, 0.007692287921905518, 0.007665664196014404, 0.007675903797149658, 0.007659520149230957, 0.007726079940795898, 0.007748608112335205, 0.0076902399063110355, 0.007664639949798584, 0.007673855781555176, 0.008031231880187988, 0.007861248016357422, 0.0076943359375, 0.007748608112335205, 0.007706624031066894, 0.007692287921905518, 0.007701504230499268, 0.007674880027770996, 0.007686143875122071, 0.00773632001876831, 0.0076984319686889645, 0.007689216136932373, 0.007707647800445557, 0.007717887878417969, 0.007724031925201416, 0.007683072090148926, 0.007711743831634522, 0.007773183822631836, 0.007753727912902832, 0.007725056171417236, 0.007733248233795166, 0.0076871681213378906, 0.007786496162414551, 0.007716864109039307, 0.0076943359375, 0.007679999828338623, 0.00769536018371582, 0.007740416049957275, 0.0076912641525268555, 0.0077199358940124516, 0.0077209601402282715, 0.007733280181884765, 0.007677919864654541, 0.007667776107788086, 0.007656383991241455, 0.016045055389404296, 0.007366655826568603, 0.007414783954620361, 0.007416831970214844, 0.007391232013702393, 0.007420928001403809, 0.007413760185241699, 0.00738099193572998, 0.007370751857757568, 0.007402495861053467, 0.007366720199584961, 0.007389120101928711, 0.007553023815155029, 0.007709695816040039, 0.007718912124633789, 0.007733248233795166, 0.007725056171417236, 0.007664639949798584, 0.007696383953094482, 0.0077281279563903805, 0.007700479984283447, 0.007723008155822754, 0.0077097277641296385, 0.007703519821166992, 0.007675903797149658, 0.007768064022064209, 0.007673855781555176, 0.007700479984283447, 0.0076687359809875484, 0.007671807765960693, 0.007711743831634522, 0.007747615814208984, 0.007689184188842774, 0.007726079940795898, 0.007665664196014404, 0.007681024074554443, 0.007667712211608887, 0.007672832012176513, 0.007755775928497314, 0.007739391803741455, 0.007716864109039307, 0.007776256084442139, 0.007678976058959961, 0.007684095859527588, 0.0076871681213378906, 0.0077036161422729494, 0.007706560134887695, 0.0076984319686889645, 0.0076871681213378906, 0.007753727912902832, 0.007681024074554443, 0.00780291223526001, 0.007710720062255859, 0.0077465281486511234, 0.007723008155822754, 0.007721983909606934, 0.007689216136932373, 0.007686143875122071, 0.007671807765960693, 0.007739424228668213, 0.007704607963562012, 0.00771782398223877, 0.007689216136932373, 0.016323583602905273, 0.007738368034362793, 0.007704576015472412, 0.007689216136932373, 0.00769536018371582, 0.00773632001876831, 0.007862271785736084, 0.0076902399063110355, 0.007661568164825439, 0.007708672046661377, 0.007669760227203369, 0.007663616180419922, 0.007697408199310303, 0.00773632001876831, 0.0076984319686889645, 0.007757855892181397, 0.007748576164245605, 0.007707647800445557, 0.00774348783493042, 0.007702527999877929, 0.007670815944671631, 0.007693280220031738, 0.0076984319686889645, 0.007684095859527588, 0.007707647800445557, 0.007697408199310303, 0.007749631881713868, 0.007721983909606934, 0.007706624031066894, 0.007702527999877929, 0.0076943359375, 0.007746560096740723, 0.007646207809448242, 0.007676928043365478, 0.008143872261047362, 0.0076871681213378906, 0.007693376064300537, 0.007671743869781494, 0.007665664196014404, 0.007764992237091065, 0.007697408199310303, 0.007701504230499268, 0.007717887878417969, 0.007682047843933106, 0.0077281279563903805, 0.00757862377166748, 0.007614463806152344, 0.007542784214019775, 0.007575551986694336, 0.007553023815155029, 0.0076605439186096195, 0.007666687965393066, 0.007708672046661377, 0.007673855781555176, 0.007674880027770996, 0.007988224029541016, 0.007820288181304931, 0.007685120105743408, 0.007780352115631104, 0.007659520149230957, 0.007746560096740723, 0.007673855781555176, 0.007670783996582031, 0.016248832702636717, 0.007948287963867188, 0.008249343872070313, 0.007725056171417236, 0.007675903797149658, 0.007684095859527588, 0.007701504230499268, 0.007648255825042725, 0.007697408199310303, 0.007670783996582031, 0.0076912641525268555, 0.007706624031066894, 0.007679999828338623, 0.007683072090148926, 0.007768064022064209, 0.00780185604095459, 0.0076943359375, 0.00773632001876831, 0.007686143875122071, 0.007813183784484864, 0.007715775966644287, 0.007685120105743408, 0.007703551769256592, 0.0076871681213378906, 0.007684095859527588, 0.007704576015472412, 0.007676928043365478, 0.007692287921905518, 0.007703551769256592, 0.007704576015472412, 0.007683072090148926, 0.007662591934204102, 0.007704576015472412, 0.007726079940795898, 0.007666687965393066, 0.007716864109039307, 0.007808000087738037, 0.007706624031066894, 0.007685120105743408, 0.0077199358940124516, 0.007740447998046875, 0.007689184188842774, 0.007696415901184082, 0.007699423789978027, 0.007682047843933106, 0.007704576015472412, 0.0076687359809875484, 0.0076687359809875484, 0.007675903797149658, 0.007716864109039307, 0.0076902399063110355, 0.007718976020812988, 0.007698368072509765, 0.007684095859527588, 0.007663648128509521, 0.00771888017654419, 0.007675903797149658, 0.00773529577255249, 0.007665664196014404, 0.007679999828338623, 0.007656447887420655, 0.007664639949798584, 0.007706624031066894, 0.01619353675842285, 0.007712768077850342, 0.007676928043365478, 0.007701504230499268, 0.007662591934204102, 0.007676928043365478, 0.007705599784851074, 0.0076912641525268555, 0.007678976058959961, 0.007665664196014404, 0.007641088008880615, 0.007677951812744141, 0.007651328086853027, 0.007704576015472412, 0.007705599784851074, 0.007712768077850342, 0.007723008155822754, 0.007697408199310303, 0.007685120105743408, 0.007697408199310303, 0.007738368034362793, 0.007599103927612305, 0.007572480201721191, 0.0075939841270446775, 0.00759500789642334, 0.007543807983398438, 0.007571455955505371, 0.007548927783966064, 0.007615488052368164, 0.007558176040649414, 0.0075447998046875, 0.007518208026885987, 0.008243200302124023, 0.007885824203491211, 0.007724031925201416, 0.007699456214904785, 0.007686143875122071, 0.007675903797149658, 0.007723008155822754, 0.007670783996582031, 0.007636991977691651, 0.007703551769256592, 0.007671807765960693, 0.007681024074554443, 0.0077608962059021, 0.0076605439186096195, 0.007634943962097168, 0.007617536067962646, 0.007719999790191651, 0.0076522879600524904, 0.007711743831634522, 0.007671807765960693, 0.0076912641525268555, 0.007686143875122071, 0.007896063804626464, 0.008217599868774414, 0.007928832054138184, 0.008215583801269532, 0.008083423614501954, 0.007780352115631104, 0.007712768077850342, 0.007734272003173828, 0.007771135807037354, 0.016397312164306642, 0.007726079940795898, 0.007714816093444824, 0.0077844481468200685, 0.007677951812744141, 0.007669760227203369, 0.007629824161529541, 0.007642111778259277, 0.007662591934204102, 0.007708672046661377, 0.0076687359809875484, 0.007710720062255859, 0.00765235185623169, 0.00765337610244751, 0.007685120105743408, 0.007689216136932373, 0.0076871681213378906, 0.0076687359809875484, 0.007513088226318359, 0.00760319995880127, 0.007512063980102539, 0.007638016223907471, 0.007721983909606934, 0.0076984319686889645, 0.007775231838226319, 0.007686143875122071, 0.007701504230499268, 0.007663616180419922, 0.007675968170166016, 0.007621568202972412, 0.007632895946502686, 0.007664639949798584, 0.007646207809448242, 0.007672832012176513, 0.007661568164825439, 0.007671840190887451, 0.007666687965393066, 0.007698400020599365, 0.007676928043365478, 0.007662591934204102, 0.007673855781555176, 0.007702591896057129, 0.007655360221862793, 0.0076871681213378906, 0.00764518404006958, 0.007671840190887451, 0.007691232204437256, 0.007700479984283447, 0.007697408199310303, 0.007723008155822754, 0.007701504230499268, 0.007707647800445557, 0.007729152202606201, 0.007676928043365478, 0.007772160053253174, 0.007649280071258545, 0.0076605439186096195, 0.007618559837341309, 0.007655424118041992, 0.007716864109039307, 0.007682047843933106, 0.007669760227203369, 0.007650335788726806, 0.016283615112304688, 0.007674943923950195, 0.007661503791809082, 0.007671807765960693, 0.007667712211608887, 0.007738368034362793, 0.007628799915313721, 0.007650303840637207, 0.007654399871826172, 0.007664639949798584, 0.007674880027770996, 0.007669760227203369, 0.007654399871826172, 0.0076943359375, 0.007699456214904785, 0.007693312168121338, 0.007756800174713135, 0.007675903797149658, 0.007675903797149658, 0.007661568164825439, 0.007662591934204102, 0.007663616180419922, 0.007685120105743408, 0.007696383953094482, 0.0076687359809875484, 0.007700479984283447, 0.007696383953094482, 0.007648287773132324, 0.007666687965393066, 0.007688159942626953, 0.007673855781555176, 0.00765337610244751, 0.0076605439186096195, 0.007641088008880615, 0.007664639949798584, 0.007659520149230957, 0.0076605439186096195, 0.007673855781555176, 0.007623680114746094, 0.00765337610244751, 0.007664639949798584, 0.00764518404006958, 0.0077547521591186525, 0.007669760227203369, 0.007762944221496582, 0.007649280071258545, 0.007681056022644043, 0.007692255973815918, 0.0077281279563903805, 0.0076943359375, 0.007726079940795898, 0.007672832012176513, 0.0076605439186096195, 0.007708672046661377, 0.007573503971099854, 0.007522304058074952, 0.00753868818283081, 0.007554048061370849, 0.007547904014587403, 0.007527423858642578, 0.007566336154937744, 0.007544832229614258, 0.007546879768371582, 0.016550912857055664, 0.007790592193603516, 0.008324095726013184, 0.007972864151000977, 0.007778304100036621, 0.007731200218200684, 0.007709695816040039, 0.007667712211608887, 0.007663616180419922, 0.0076871681213378906, 0.007675968170166016, 0.007672768115997314, 0.007716864109039307, 0.007697408199310303, 0.007701504230499268, 0.0076912641525268555, 0.007684095859527588, 0.007732223987579345, 0.0076308479309082035, 0.007643136024475097, 0.007649280071258545, 0.007674880027770996, 0.007624703884124756, 0.007667712211608887, 0.007677951812744141, 0.007624703884124756, 0.007683072090148926, 0.007701504230499268, 0.007667712211608887, 0.007664639949798584, 0.007677951812744141, 0.00765235185623169, 0.007651328086853027, 0.007658527851104736, 0.007616479873657226, 0.007672832012176513, 0.007639039993286132, 0.007648255825042725, 0.007730175971984863, 0.00764518404006958, 0.007659584045410156, 0.007690176010131836, 0.0076687359809875484, 0.007747583866119385, 0.007681024074554443, 0.007682047843933106, 0.00760319995880127, 0.00764518404006958, 0.00764518404006958, 0.007661568164825439, 0.007667712211608887, 0.007654399871826172, 0.0076984319686889645, 0.007693312168121338, 0.007664639949798584, 0.007640063762664795, 0.007636991977691651, 0.007641088008880615, 0.0076605439186096195, 0.007710720062255859, 0.007663616180419922, 0.007651328086853027, 0.0076687359809875484, 0.01621401596069336, 0.007667712211608887, 0.00764518404006958, 0.007679999828338623, 0.007654399871826172, 0.007701504230499268, 0.00765235185623169, 0.007671807765960693, 0.007676928043365478, 0.007639039993286132, 0.007688191890716553, 0.00765337610244751, 0.007658495903015137, 0.007681024074554443, 0.007683072090148926, 0.0076687359809875484, 0.007700479984283447, 0.00764518404006958, 0.007771135807037354, 0.007686143875122071, 0.007669760227203369, 0.00753868818283081, 0.007576576232910156, 0.007585792064666748, 0.007557119846343994, 0.00753766393661499, 0.007608320236206055, 0.007524352073669433, 0.007278592109680176, 0.007284736156463623, 0.007310336112976074, 0.007329792022705078, 0.007324672222137451, 0.007779327869415284, 0.007463935852050781, 0.007377920150756836, 0.007362559795379638, 0.0074301438331604, 0.00740556812286377, 0.0074301438331604, 0.007398399829864502, 0.007364607810974121, 0.007392255783081054, 0.007345151901245117, 0.007398399829864502, 0.007434239864349365, 0.007396351814270019, 0.007379968166351319, 0.007365632057189942, 0.007364607810974121, 0.007402495861053467, 0.007412735939025879, 0.007408639907836914, 0.00738099193572998, 0.007361536026000977, 0.007361536026000977, 0.007480319976806641, 0.007412735939025879, 0.007378943920135498, 0.007364607810974121, 0.007361536026000977, 0.007365632057189942, 0.007361536026000977, 0.0156559362411499, 0.007377920150756836, 0.007386112213134765, 0.007384064197540283, 0.00738099193572998, 0.007365632057189942, 0.00733900785446167, 0.0073472318649291995, 0.007423967838287354, 0.007371808052062988, 0.007373792171478271, 0.00738099193572998, 0.00734822416305542, 0.007372799873352051, 0.007411712169647216, 0.007392255783081054, 0.007402495861053467, 0.007345183849334717, 0.007376863956451416, 0.007349247932434082, 0.007371776103973389, 0.007395328044891358, 0.007371776103973389, 0.007372799873352051, 0.007326720237731933, 0.007321599960327148, 0.007365632057189942, 0.007445504188537597, 0.007366655826568603, 0.007371776103973389, 0.0074301438331604, 0.007391232013702393, 0.007358463764190673, 0.007360511779785156, 0.007398399829864502, 0.007402495861053467, 0.007356416225433349, 0.007336959838867187, 0.007646207809448242, 0.007674880027770996, 0.007667712211608887, 0.007679999828338623, 0.007647232055664062, 0.007686143875122071, 0.007655424118041992, 0.007648255825042725, 0.00765337610244751, 0.007667712211608887, 0.00765337610244751, 0.007700479984283447, 0.007701504230499268, 0.007689216136932373, 0.007701504230499268, 0.007667712211608887, 0.007701504230499268, 0.007697408199310303, 0.0076943359375, 0.007688191890716553, 0.007675903797149658, 0.007731200218200684, 0.007679999828338623, 0.007655424118041992, 0.007692351818084717, 0.01561900806427002, 0.007393280029296875, 0.007370751857757568, 0.007378943920135498, 0.007379968166351319, 0.007387135982513428, 0.007362559795379638, 0.007363584041595459, 0.007407616138458252, 0.00742195177078247, 0.007388160228729248, 0.007401472091674805, 0.007394303798675537, 0.007366655826568603, 0.007392255783081054, 0.007396351814270019, 0.007401535987854004, 0.007330751895904541, 0.007383039951324463, 0.007351295948028564, 0.007345151901245117, 0.007377920150756836, 0.007448575973510742, 0.007396351814270019, 0.007396351814270019, 0.007361536026000977, 0.007325695991516113, 0.0073697280883789065, 0.007375872135162354, 0.007426047801971435, 0.007325695991516113, 0.00728985595703125, 0.007270400047302246, 0.007261184215545655, 0.007361536026000977, 0.007296000003814697, 0.007258111953735351, 0.007267327785491944, 0.007395328044891358, 0.007804927825927735, 0.0076574721336364745, 0.00764518404006958, 0.007676928043365478, 0.007739391803741455, 0.007699456214904785, 0.007674880027770996, 0.007642111778259277, 0.007688191890716553, 0.0076574721336364745, 0.007546879768371582, 0.007524352073669433, 0.0075939841270446775, 0.0076902399063110355, 0.007580671787261963, 0.007497727870941162, 0.007346176147460938, 0.007362559795379638, 0.007375872135162354, 0.0073431038856506346, 0.007374847888946533, 0.007352320194244385, 0.0073697280883789065, 0.007363647937774658, 0.01567638397216797, 0.007364575862884522, 0.007354368209838867, 0.00739737606048584, 0.007385087966918945, 0.007407616138458252, 0.00738099193572998, 0.007359519958496093, 0.007383008003234864, 0.007362559795379638, 0.007404543876647949, 0.007392255783081054, 0.007372799873352051, 0.007394303798675537, 0.007402495861053467, 0.007374847888946533, 0.0073994240760803225, 0.007374847888946533, 0.007438399791717529, 0.007394271850585938, 0.007375840187072754, 0.0074741759300231934, 0.007402495861053467, 0.007414783954620361, 0.007362559795379638, 0.0073820161819458, 0.007402495861053467, 0.007396351814270019, 0.007412735939025879, 0.007388160228729248, 0.007396351814270019, 0.0074065918922424315, 0.00739737606048584, 0.007263232231140137, 0.007335968017578125, 0.00739734411239624, 0.0075038719177246095, 0.0073994240760803225, 0.0073697280883789065, 0.007396351814270019, 0.007400447845458984, 0.007393280029296875, 0.007400447845458984, 0.007388160228729248, 0.0073994240760803225, 0.007412735939025879, 0.007416831970214844, 0.007412735939025879, 0.007352352142333984, 0.007434207916259766, 0.007423999786376953, 0.007434239864349365, 0.007411712169647216, 0.007364640235900879, 0.007370719909667969, 0.0074301438331604, 0.007374847888946533, 0.0073994240760803225, 0.007393280029296875, 0.00743936014175415, 0.007366655826568603, 0.007361536026000977, 0.00738918399810791, 0.01574092769622803, 0.007400447845458984, 0.007386112213134765, 0.007407616138458252, 0.0073820161819458, 0.007384064197540283, 0.0074035201072692874, 0.0074403839111328125, 0.0074106879234313965, 0.007392255783081054, 0.00738918399810791, 0.0073994240760803225, 0.0073994240760803225, 0.007390207767486572, 0.007418879985809326, 0.007426047801971435, 0.007351295948028564, 0.007263232231140137, 0.0072960638999938964, 0.0072979841232299805, 0.007377920150756836, 0.007384064197540283, 0.00740556812286377, 0.007387135982513428, 0.007378943920135498, 0.007363584041595459, 0.007459839820861816, 0.00738099193572998, 0.007366655826568603, 0.007377920150756836, 0.007383039951324463, 0.007371776103973389, 0.007378943920135498, 0.007357439994812012, 0.007388160228729248, 0.0074106879234313965, 0.007364607810974121, 0.007372799873352051, 0.00739737606048584, 0.00740556812286377, 0.00738918399810791, 0.007435264110565186, 0.007419904232025146, 0.007354368209838867, 0.007371776103973389, 0.0074414081573486324, 0.007385087966918945, 0.007356416225433349, 0.007312384128570557, 0.007264256000518799, 0.0072837119102478025, 0.007322624206542969, 0.0073471999168395995, 0.007331840038299561, 0.007378943920135498, 0.007390207767486572, 0.007377920150756836, 0.007387135982513428, 0.007411712169647216, 0.007385087966918945, 0.007393280029296875, 0.007394303798675537, 0.007395328044891358, 0.015649791717529296, 0.007411712169647216, 0.007391263961791992, 0.00740553617477417, 0.007412735939025879, 0.007411712169647216, 0.0073820161819458, 0.007357439994812012, 0.007420928001403809, 0.007412735939025879, 0.007396351814270019, 0.007371776103973389, 0.007401472091674805, 0.007377920150756836, 0.007415808200836181, 0.00740556812286377, 0.007315455913543701, 0.007308288097381592, 0.007303167819976806, 0.007323679924011231, 0.007309279918670654, 0.00733081579208374, 0.007311359882354736, 0.007400447845458984, 0.008755200386047364, 0.008035327911376953, 0.007821375846862794, 0.00771782398223877, 0.007775231838226319, 0.007707647800445557, 0.007746560096740723, 0.0076995201110839845, 0.007703487873077393, 0.007682047843933106, 0.007733248233795166, 0.007883776187896728, 0.007564352035522461, 0.007551936149597168, 0.007683135986328125, 0.007680960178375244, 0.007654399871826172, 0.00753766393661499, 0.007549952030181885, 0.0075335679054260255, 0.007548927783966064, 0.007525375843048096, 0.007574528217315674, 0.007600128173828125, 0.007554048061370849, 0.007589888095855713, 0.007572480201721191, 0.007592959880828858, 0.007561215877532959, 0.007558144092559814, 0.007566336154937744, 0.007528448104858398, 0.007329792022705078, 0.007431168079376221, 0.007398399829864502, 0.007428095817565918, 0.00738918399810791, 0.007349247932434082, 0.007400447845458984, 0.015781888008117674, 0.007443488121032715, 0.007380959987640381, 0.007375872135162354, 0.007364607810974121, 0.007367680072784424, 0.007391263961791992, 0.007384064197540283, 0.00740451192855835, 0.007401472091674805, 0.007392255783081054, 0.007364607810974121, 0.007366655826568603, 0.007394303798675537, 0.007425024032592774, 0.007407616138458252, 0.007394303798675537, 0.007346176147460938, 0.007378943920135498, 0.007374847888946533, 0.007375872135162354, 0.007365632057189942, 0.007372799873352051, 0.007393280029296875, 0.007372799873352051, 0.007358496189117432, 0.007380959987640381, 0.007319551944732666, 0.007417856216430664, 0.007304192066192627, 0.007320576190948487, 0.007308288097381592, 0.0073175039291381834, 0.00728985595703125, 0.007305215835571289, 0.007374847888946533, 0.007365632057189942, 0.007395328044891358, 0.007388160228729248, 0.007371776103973389, 0.007379968166351319, 0.007392255783081054, 0.007365632057189942, 0.007355391979217529, 0.007404543876647949, 0.007415808200836181, 0.007385087966918945, 0.007371776103973389, 0.007375936031341553, 0.007378880023956299, 0.007391232013702393, 0.0074035201072692874, 0.007415808200836181, 0.007390207767486572, 0.007393280029296875, 0.007333951950073242, 0.0073614721298217774, 0.0074065918922424315, 0.0074035201072692874, 0.007378943920135498, 0.007374911785125732, 0.007392191886901855, 0.007392255783081054, 0.015719424247741698, 0.007379968166351319, 0.00738918399810791, 0.007395328044891358, 0.007365632057189942, 0.0073697280883789065, 0.007378943920135498, 0.007354368209838867, 0.0074106879234313965, 0.007387135982513428, 0.007385087966918945, 0.007385087966918945, 0.0073471999168395995, 0.007364607810974121, 0.007412735939025879, 0.007341087818145752, 0.007336927890777588, 0.007327744007110596, 0.007299071788787842, 0.007301119804382325, 0.007413760185241699, 0.0073175039291381834, 0.007294976234436036, 0.007336959838867187, 0.007268383979797364, 0.007279583930969238, 0.007331840038299561, 0.007334911823272705, 0.007318528175354004, 0.0073062400817871095, 0.007285759925842285, 0.00727347183227539, 0.007318528175354004, 0.007302144050598145, 0.00742195177078247, 0.007435264110565186, 0.007355391979217529, 0.007367680072784424, 0.007369791984558106, 0.007402431964874267, 0.007560192108154297, 0.007434239864349365, 0.007358463764190673, 0.0073768959045410155, 0.007364607810974121, 0.007368703842163086, 0.007390207767486572, 0.007412735939025879, 0.007370751857757568, 0.007393280029296875, 0.007383039951324463, 0.007404543876647949, 0.007404543876647949, 0.007373824119567871, 0.007407616138458252, 0.00742195177078247, 0.007388160228729248, 0.0073820161819458, 0.007372799873352051, 0.00743936014175415, 0.007371776103973389, 0.007361536026000977, 0.007387135982513428, 0.015711263656616212, 0.007366623878479004, 0.007387135982513428, 0.007398399829864502, 0.0074065918922424315, 0.007390207767486572, 0.007378943920135498, 0.00738918399810791, 0.007393280029296875, 0.007409664154052734, 0.007394303798675537, 0.00739737606048584, 0.007400447845458984, 0.007364607810974121, 0.007390207767486572, 0.007418879985809326, 0.0074414081573486324, 0.007411712169647216, 0.007400447845458984, 0.007377920150756836, 0.007398399829864502, 0.007390272140502929, 0.0074280319213867185, 0.007404543876647949, 0.007390207767486572, 0.007371776103973389, 0.007383039951324463, 0.007350272178649903, 0.007373824119567871, 0.007416831970214844, 0.007515135765075683, 0.007432191848754883, 0.007402495861053467, 0.007525440216064453, 0.00740447998046875, 0.007375872135162354, 0.007377920150756836, 0.0073994240760803225, 0.007387135982513428, 0.007408639907836914, 0.007351295948028564, 0.007384064197540283, 0.0074035201072692874, 0.007388160228729248, 0.0074291200637817386, 0.007391232013702393, 0.007385087966918945, 0.007531551837921142, 0.007364575862884522, 0.0073994240760803225, 0.0076267518997192385, 0.007418879985809326, 0.007401472091674805, 0.007378943920135498, 0.007378943920135498, 0.007366655826568603, 0.007370751857757568, 0.0073768959045410155, 0.007390207767486572, 0.0073697280883789065, 0.00734822416305542, 0.007363584041595459, 0.007373888015747071, 0.01573574447631836, 0.007395328044891358, 0.0073994240760803225, 0.007390207767486572, 0.007352320194244385, 0.007368703842163086, 0.007385087966918945, 0.007396351814270019, 0.007409664154052734, 0.007372799873352051, 0.00739737606048584, 0.00738099193572998, 0.007366655826568603, 0.007390207767486572, 0.007420928001403809, 0.00740556812286377, 0.007412735939025879, 0.007353343963623047, 0.007373824119567871, 0.007384064197540283, 0.007404543876647949, 0.007434239864349365, 0.007366655826568603, 0.007373824119567871, 0.007270400047302246, 0.007271455764770508, 0.007312352180480957, 0.007277567863464355, 0.007293951988220215, 0.007299071788787842, 0.007250944137573242, 0.007284736156463623, 0.007324672222137451, 0.007338047981262207, 0.007430079936981201, 0.00738918399810791, 0.007358463764190673, 0.007390207767486572, 0.007359519958496093, 0.007379936218261719, 0.007390207767486572, 0.007388160228729248, 0.007409664154052734, 0.007365632057189942, 0.007387135982513428, 0.007549952030181885, 0.00742195177078247, 0.007391232013702393, 0.007366655826568603, 0.007412735939025879, 0.007390207767486572, 0.007388160228729248, 0.007420928001403809, 0.007378943920135498, 0.007383039951324463, 0.007412735939025879, 0.007353343963623047, 0.0074332160949707035, 0.007392255783081054, 0.00744652795791626, 0.007373824119567871, 0.0073851518630981446, 0.007392191886901855, 0.015643648147583008, 0.007444479942321777, 0.007362559795379638, 0.007434239864349365, 0.007387135982513428, 0.007407616138458252, 0.007363584041595459, 0.007378943920135498, 0.007417856216430664, 0.007408639907836914, 0.007395328044891358, 0.007408639907836914, 0.007406623840332032, 0.007435232162475586, 0.007392255783081054, 0.007408639907836914, 0.00738918399810791, 0.007392255783081054, 0.00739737606048584, 0.007349247932434082, 0.007363584041595459, 0.0073820161819458, 0.0074301438331604, 0.007390207767486572, 0.007391232013702393, 0.007384064197540283, 0.00738099193572998, 0.007354368209838867, 0.007386112213134765, 0.007391232013702393, 0.007364607810974121, 0.007356416225433349, 0.007345151901245117, 0.007331840038299561, 0.007379968166351319, 0.007378943920135498, 0.007362559795379638, 0.0073769278526306156, 0.007376863956451416, 0.007371776103973389, 0.007478271961212158, 0.007414783954620361, 0.007415808200836181, 0.007413760185241699, 0.007477248191833496, 0.007494656085968018, 0.007546879768371582, 0.007393280029296875, 0.0074670081138610836, 0.007398399829864502, 0.007418879985809326, 0.007432191848754883, 0.007437312126159668, 0.007404543876647949, 0.0073768959045410155, 0.0073768959045410155, 0.00738918399810791, 0.007393280029296875, 0.007390207767486572, 0.007355391979217529, 0.007346176147460938, 0.00739737606048584, 0.007373824119567871, 0.015738880157470703, 0.0074117441177368164, 0.007378911972045899, 0.007386112213134765, 0.0073523840904235836, 0.007378880023956299, 0.00739737606048584, 0.007362559795379638, 0.007384064197540283, 0.007366655826568603, 0.007387135982513428, 0.007365632057189942, 0.007395328044891358, 0.007372799873352051, 0.007437376022338868, 0.007419839859008789, 0.007411712169647216, 0.007391232013702393, 0.007398399829864502, 0.007400447845458984, 0.007418879985809326, 0.007420928001403809, 0.0073820161819458, 0.0073768959045410155, 0.007360544204711914, 0.007381984233856201, 0.007418879985809326, 0.007398399829864502, 0.007414783954620361, 0.007370783805847168, 0.007406559944152832, 0.007364607810974121, 0.007387135982513428, 0.007358463764190673, 0.0074106879234313965, 0.007400447845458984, 0.007365632057189942, 0.0073994240760803225, 0.007568384170532226, 0.007576576232910156, 0.0075642881393432615, 0.007554048061370849, 0.00753766393661499, 0.00733081579208374, 0.0073134078979492185, 0.007333888053894043, 0.007400447845458984, 0.00740556812286377, 0.007408639907836914, 0.007590911865234375, 0.007489535808563232, 0.007449600219726562, 0.007478271961212158, 0.0076687359809875484, 0.007718912124633789, 0.007711743831634522, 0.007670783996582031, 0.007672832012176513, 0.007721983909606934, 0.007824384212493896, 0.007670783996582031, 0.0076574721336364745, 0.007717887878417969]",tokens/s,130.38016167454896,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,2,2,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/2/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669324e2-498f0735401b7e9259008e2d;5c2b6abc-e53e-458f-b25a-d05c76709e5a) Repository Not Found for url: https://huggingface.co/2/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: 2 is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/xglm-564M,facebook/xglm-564M,cuda,0,42,,,True,,,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",xglm,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-14B,Qwen/Qwen-14B,cuda,0,42,,,True,,,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",qwen,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,huggyllama/llama-7b,huggyllama/llama-7b,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.1,,,,MB,1231.99488,14529.593344,0.0,13883.14624,13705.186304,s,10,0.9352619552612306,0.09352619552612304,0.00022871080860555472,0.09344286346435547,0.09363615264892577,0.09390740509033203,0.09412440704345702,"[0.09417865753173828, 0.09348995208740235, 0.09342777252197265, 0.09343241882324219, 0.09332745361328125, 0.09357587432861328, 0.09344054412841797, 0.09356108856201172, 0.0933830108642578, 0.09344518280029297]",tokens/s,2737.2010436209393,kWh,1.1073021734614982e-06,6.066724194408091e-07,6.123927536728964e-06,7.837902129631271e-06,tokens/kWh,32661801.049057417,MB,1231.99488,14529.593344,0.0,13883.14624,13856.22528,s,10,22.602925048828123,2.260292504882812,0.0016421842502448833,2.2602330322265622,2.262653857421875,2.2627298828124998,2.262790703125,"[2.262636962890625, 2.260993408203125, 2.259475341796875, 2.262805908203125, 2.25965234375, 2.260813720703125, 2.25700244140625, 2.26097216796875, 2.259251220703125, 2.259321533203125]",tokens/s,27.872498742487455,kWh,2.662618296230569e-05,1.4590572437131258e-05,0.00011656403172467102,0.000157780787124108,tokens/kWh,399288.15889633726,,s,629,23.407387664794918,0.037213652885206554,0.01117367162514878,0.03586969757080078,0.03598151626586914,0.03602227020263672,0.12987019348144532,"[0.03587481689453125, 0.03576627349853516, 0.03585228729248047, 0.03581849670410156, 0.0357498893737793, 0.03581849670410156, 0.035762176513671876, 0.035743743896484374, 0.035783679962158206, 0.0357386245727539, 0.035781631469726564, 0.03577036666870117, 0.035784702301025394, 0.03578572845458984, 0.035899425506591795, 0.03585123062133789, 0.03577446365356445, 0.03582259368896484, 0.035789825439453124, 0.03584204864501953, 0.03582156753540039, 0.035833854675292966, 0.03583795166015625, 0.035879936218261715, 0.035883007049560545, 0.03590348815917969, 0.0358205451965332, 0.03582361602783203, 0.035883007049560545, 0.0359106559753418, 0.035899391174316404, 0.03589324951171875, 0.03590348815917969, 0.03587891387939453, 0.03590963363647461, 0.03588710403442383, 0.03585228729248047, 0.035980289459228515, 0.036040702819824216, 0.035931137084960936, 0.035884033203125, 0.035966976165771485, 0.035934207916259765, 0.03590041732788086, 0.035952640533447267, 0.035939327239990236, 0.0359444465637207, 0.035920894622802735, 0.03590553665161133, 0.035939327239990236, 0.03618815994262695, 0.03666124725341797, 0.03601919937133789, 0.03595980834960937, 0.035912704467773435, 0.03601513671875, 0.035977184295654295, 0.03596806335449219, 0.03598432159423828, 0.03598643112182617, 0.03602022552490235, 0.03594137573242188, 0.13005926513671875, 0.03581849670410156, 0.035757057189941405, 0.035760128021240234, 0.03573657608032227, 0.035811328887939455, 0.03576732635498047, 0.035827678680419923, 0.035817470550537106, 0.0357130241394043, 0.035784702301025394, 0.03576422500610352, 0.03578060913085938, 0.03583488082885742, 0.03577241516113281, 0.0358021125793457, 0.03583692932128906, 0.03580416107177734, 0.035776512145996094, 0.035781631469726564, 0.03579904174804688, 0.03579084777832031, 0.03590963363647461, 0.03587276840209961, 0.03577036666870117, 0.03585228729248047, 0.035850238800048825, 0.03590655899047852, 0.03586969757080078, 0.03581951904296875, 0.03585433578491211, 0.03581849670410156, 0.035901439666748046, 0.03594649505615234, 0.03597619247436523, 0.03590860748291016, 0.03594649505615234, 0.035939327239990236, 0.035883007049560545, 0.03589017486572266, 0.035883007049560545, 0.03590041732788086, 0.03593625640869141, 0.03600998306274414, 0.03591372680664062, 0.03590963363647461, 0.03597619247436523, 0.035920894622802735, 0.035917823791503906, 0.03590860748291016, 0.036001792907714845, 0.03596492767333984, 0.03599871826171875, 0.03599462509155273, 0.03595673751831055, 0.03598233413696289, 0.035972095489501955, 0.03600588989257812, 0.03593625640869141, 0.0359813117980957, 0.035996673583984375, 0.03594035339355469, 0.03602227020263672, 0.12988108825683595, 0.03569356918334961, 0.0357130241394043, 0.03576627349853516, 0.035732479095458985, 0.035762176513671876, 0.03582668685913086, 0.035710975646972655, 0.03577446365356445, 0.03574272155761719, 0.03573350524902344, 0.035748863220214845, 0.035937278747558594, 0.03594137573242188, 0.03578675079345703, 0.03581951904296875, 0.03575091171264649, 0.03585740661621094, 0.03577446365356445, 0.03579596710205078, 0.03577139282226562, 0.03580825424194336, 0.035798015594482424, 0.03580825424194336, 0.0358021125793457, 0.03578572845458984, 0.03580928039550781, 0.03580416107177734, 0.03581139373779297, 0.035843006134033205, 0.03587686538696289, 0.03585638427734375, 0.03585638427734375, 0.035871742248535156, 0.03582566452026367, 0.03583488082885742, 0.03584819030761719, 0.03589734268188476, 0.03605196762084961, 0.035912704467773435, 0.03583795166015625, 0.035884033203125, 0.035929088592529294, 0.03585638427734375, 0.0359106559753418, 0.03588710403442383, 0.035996673583984375, 0.03597107315063477, 0.03593625640869141, 0.03591167831420899, 0.035934207916259765, 0.035937278747558594, 0.03591167831420899, 0.035914752960205076, 0.03590655899047852, 0.03597312164306641, 0.0359813117980957, 0.03593011093139648, 0.0359444465637207, 0.035879936218261715, 0.0359106559753418, 0.035937278747558594, 0.03603148651123047, 0.1299251251220703, 0.03576627349853516, 0.03577446365356445, 0.03571200180053711, 0.03578060913085938, 0.03576934432983398, 0.035765247344970705, 0.03574476623535156, 0.035798015594482424, 0.03576115036010742, 0.03580416107177734, 0.035814399719238284, 0.03584307098388672, 0.03583078384399414, 0.03584716796875, 0.03587891387939453, 0.03583180618286133, 0.03575603103637695, 0.035810302734375, 0.03585331344604492, 0.03585126495361328, 0.03585433578491211, 0.03589017486572266, 0.035833854675292966, 0.03591167831420899, 0.03587071990966797, 0.03587686538696289, 0.03586969757080078, 0.035862529754638675, 0.0359813117980957, 0.036311038970947264, 0.0359813117980957, 0.03597926330566406, 0.035862529754638675, 0.03593011093139648, 0.03589734268188476, 0.03594956970214844, 0.03589836883544922, 0.03593222427368164, 0.03586553573608398, 0.03596799850463867, 0.03596799850463867, 0.03598950576782227, 0.035914752960205076, 0.03593116760253906, 0.03594646453857422, 0.03594342422485351, 0.035929088592529294, 0.03592601776123047, 0.035939327239990236, 0.03597721481323242, 0.036122623443603515, 0.03602739334106445, 0.03592499160766602, 0.03603251266479492, 0.035985408782958986, 0.03609395217895508, 0.0361338882446289, 0.03610009765625, 0.03594956970214844, 0.03597619247436523, 0.035961856842041014, 0.036034561157226565, 0.12981248474121093, 0.03573555374145508, 0.03570278549194336, 0.035707904815673826, 0.035730430603027344, 0.03568435287475586, 0.0357386245727539, 0.03571916961669922, 0.035776512145996094, 0.035745792388916016, 0.03579084777832031, 0.035800064086914066, 0.03574476623535156, 0.035757057189941405, 0.035768318176269534, 0.03578265762329102, 0.03573964691162109, 0.03580825424194336, 0.03576422500610352, 0.035798015594482424, 0.03578777694702148, 0.03577241516113281, 0.035784702301025394, 0.035746814727783204, 0.0357918701171875, 0.03579296112060547, 0.03579692840576172, 0.035827713012695314, 0.0358205451965332, 0.035814399719238284, 0.03584102249145508, 0.03578675079345703, 0.035827713012695314, 0.035810302734375, 0.03584204864501953, 0.03584921646118164, 0.035883007049560545, 0.035844097137451174, 0.03590867233276367, 0.035937217712402346, 0.03592704010009766, 0.03597107315063477, 0.03595980834960937, 0.03587788772583008, 0.03597824096679687, 0.03594649505615234, 0.036036609649658206, 0.035912704467773435, 0.03599257659912109, 0.0359444465637207, 0.035952640533447267, 0.035955711364746096, 0.035945472717285154, 0.03595775985717774, 0.03601408004760742, 0.036038654327392575, 0.03600588989257812, 0.03594854354858398, 0.03600588989257812, 0.035953662872314454, 0.036016128540039063, 0.03598745727539063, 0.03612160110473633, 0.1300807647705078, 0.035796993255615236, 0.03581542587280274, 0.03583283233642578, 0.03588710403442383, 0.03578777694702148, 0.03577753448486328, 0.03576319885253906, 0.035733535766601564, 0.03579395294189453, 0.03573958587646484, 0.03583283233642578, 0.03576319885253906, 0.035813377380371096, 0.03584000015258789, 0.03578879928588867, 0.035796993255615236, 0.035833854675292966, 0.0358021125793457, 0.03577139282226562, 0.03581542587280274, 0.03578879928588867, 0.03586355209350586, 0.0358656005859375, 0.03588198471069336, 0.035883007049560545, 0.03584511947631836, 0.03588710403442383, 0.035862529754638675, 0.03594137573242188, 0.03584000015258789, 0.03591372680664062, 0.03590044784545898, 0.035929054260253904, 0.03590655899047852, 0.035885055541992186, 0.03593318557739258, 0.03588198471069336, 0.03590041732788086, 0.03588915252685547, 0.035947521209716796, 0.03589836883544922, 0.03595775985717774, 0.03586355209350586, 0.03586457443237305, 0.0359444465637207, 0.035983360290527344, 0.03593523025512695, 0.03593523025512695, 0.035896320343017575, 0.035937278747558594, 0.0359628791809082, 0.03595471954345703, 0.03591984176635742, 0.035932159423828124, 0.035931137084960936, 0.03598643112182617, 0.035922943115234376, 0.035963905334472655, 0.0359741439819336, 0.03592601776123047, 0.0359628791809082, 0.035972095489501955, 0.12984217834472656, 0.035730430603027344, 0.03571916961669922, 0.035643390655517575, 0.035681278228759765, 0.0357437744140625, 0.0357130241394043, 0.035682273864746095, 0.035699710845947266, 0.0356864013671875, 0.03569356918334961, 0.03571712112426758, 0.03577036666870117, 0.035707904815673826, 0.035757057189941405, 0.03575398254394531, 0.035748863220214845, 0.03570073699951172, 0.035714046478271484, 0.03571814346313477, 0.03573452758789063, 0.03575296020507813, 0.03577139282226562, 0.03576224136352539, 0.035769279479980466, 0.0357918701171875, 0.03576115036010742, 0.03575296020507813, 0.03582361602783203, 0.036034561157226565, 0.03601206588745117, 0.03588502502441406, 0.03585945510864258, 0.035829761505126956, 0.03585331344604492, 0.03585331344604492, 0.03584511947631836, 0.035827713012695314, 0.035817470550537106, 0.035810302734375, 0.03584102249145508, 0.035846145629882815, 0.03589120101928711, 0.0358287353515625, 0.03586457443237305, 0.03587276840209961, 0.03589836883544922, 0.035885055541992186, 0.03585945510864258, 0.03586355209350586, 0.03589734268188476, 0.0359024658203125, 0.035914752960205076, 0.03587481689453125, 0.03590348815917969, 0.03589529418945313, 0.03589529418945313, 0.035884033203125, 0.0359024658203125, 0.03589734268188476, 0.03592601776123047, 0.035939327239990236, 0.0359813117980957, 0.1301822052001953, 0.03573241424560547, 0.03572431945800781, 0.03571708679199219, 0.03575398254394531, 0.03574272155761719, 0.03577862548828125, 0.03575494384765625, 0.0358021125793457, 0.035773441314697264, 0.03584204864501953, 0.035773441314697264, 0.03577756881713867, 0.035756000518798826, 0.03583488082885742, 0.03583590316772461, 0.035885055541992186, 0.03577446365356445, 0.03577446365356445, 0.03580416107177734, 0.035789825439453124, 0.03580723190307617, 0.03582668685913086, 0.03578675079345703, 0.035811328887939455, 0.03584204864501953, 0.03584307098388672, 0.03586355209350586, 0.03581644821166992, 0.03586355209350586, 0.03590041732788086, 0.03584000015258789, 0.03590860748291016, 0.03591372680664062, 0.03592601776123047, 0.035922943115234376, 0.03592704010009766, 0.03587686538696289, 0.03592704010009766, 0.03591680145263672, 0.03596799850463867, 0.035912704467773435, 0.03590348815917969, 0.03592499160766602, 0.03592806243896484, 0.035929088592529294, 0.036168704986572264, 0.036071422576904294, 0.03598438262939453, 0.0359813117980957, 0.03596799850463867, 0.03594956970214844, 0.03595980834960937, 0.035942401885986325, 0.03592192077636719, 0.03594854354858398, 0.03602227020263672, 0.03594649505615234, 0.03595775985717774, 0.03598643112182617, 0.03599462509155273, 0.035945472717285154, 0.035958782196044925, 0.1298903045654297, 0.03576627349853516, 0.03575296020507813, 0.03571718215942383, 0.03572934341430664, 0.03575807952880859, 0.03574476623535156, 0.03575603103637695, 0.03591167831420899, 0.03595673751831055, 0.03572531127929687, 0.03573555374145508, 0.03573555374145508, 0.035768318176269534, 0.035811328887939455, 0.03581644821166992, 0.035760128021240234, 0.03583180618286133, 0.035778560638427735, 0.03574476623535156, 0.03577753448486328, 0.0357918701171875, 0.03578572845458984, 0.03580825424194336, 0.035800064086914066, 0.03583795166015625, 0.03583488082885742, 0.03581951904296875, 0.03585331344604492, 0.035860481262207033, 0.03594649505615234, 0.035888126373291016, 0.03586969757080078, 0.03580518341064453, 0.03584511947631836, 0.035844097137451174, 0.03584921646118164, 0.03583692932128906, 0.03585331344604492, 0.03584102249145508, 0.035860481262207033, 0.03586764907836914, 0.03590758514404297, 0.0358809585571289, 0.03589427185058594, 0.035896320343017575, 0.03593011093139648, 0.03587276840209961, 0.03593011093139648, 0.035917823791503906, 0.035896320343017575, 0.03593830490112305, 0.035952640533447267, 0.03591884613037109, 0.03595468902587891, 0.03595673751831055, 0.03598233413696289, 0.035922943115234376, 0.035955711364746096, 0.03594649505615234, 0.03592499160766602, 0.035945472717285154, 0.03596083068847656, 0.1298841552734375, 0.0357283821105957, 0.03572633743286133, 0.035659774780273434, 0.035705856323242184, 0.03574172973632812, 0.03574985504150391, 0.03571814346313477, 0.035746814727783204, 0.03572326278686523, 0.03574784088134766, 0.03573657608032227, 0.03578265762329102, 0.03574169540405273, 0.03585126495361328, 0.03577241516113281, 0.035768318176269534, 0.03576220703125, 0.035764190673828126, 0.0357386245727539, 0.035888126373291016, 0.036119552612304685, 0.035901439666748046, 0.03579596710205078, 0.0358389778137207, 0.035844097137451174, 0.03584307098388672, 0.03581235122680664, 0.03581542587280274, 0.035810302734375, 0.03581542587280274, 0.0358287353515625, 0.03588608169555664, 0.03581235122680664, 0.03589529418945313, 0.03585843276977539, 0.0358922233581543, 0.03584511947631836, 0.03585638427734375, 0.035862529754638675, 0.03587686538696289, 0.035901439666748046, 0.03592396926879883, 0.03584921646118164, 0.03587276840209961, 0.035901439666748046, 0.03594956970214844, 0.03593318557739258, 0.03591167831420899, 0.035892288208007814, 0.03592595291137695, 0.03603558349609375, 0.03605401611328125, 0.035972095489501955, 0.036004863739013675, 0.03595468902587891, 0.03591987228393555, 0.03588710403442383, 0.03591884613037109, 0.03591987228393555, 0.03595161437988281, 0.03599769592285156, 0.0359659538269043]",tokens/s,26.871858107687338,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-180B,tiiuae/falcon-180B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 403 Client Error: Forbidden for url: https://huggingface.co/tiiuae/falcon-180B/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-66932719-362393661d8dd1ca0352b206;b0a9e331-de24-449e-9cc6-6726d39596a2) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/tiiuae/falcon-180B/resolve/main/config.json. If you are trying to create or update content,make sure you have a token with the `write` role. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1826, in _raise_on_head_call_error raise LocalEntryNotFoundError( huggingface_hub.utils._errors.LocalEntryNotFoundError: An error happened while trying to locate the file on the Hub and we cannot find the requested files in the local cache. Please check your connection and try again or make sure your Internet connection is on. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 445, in cached_file raise EnvironmentError( OSError: We couldn't connect to 'https://huggingface.co' to load this file, couldn't find it in the cached files and it looks like tiiuae/falcon-180B is not the path to a directory containing a file named config.json. Checkout your internet connection or see how to run the library in offline mode at 'https://huggingface.co/docs/transformers/installation#offline-mode'. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/gpt-neo-1.3B,EleutherAI/gpt-neo-1.3B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1212.882944,3760.717824,0.0,3114.27072,2817.469952,s,12,0.2584707832336426,0.021539231936136882,0.0002784461869800534,0.021448639869689944,0.021528425407409665,0.021945079708099363,0.022351076488494874,"[0.02245257568359375, 0.021529855728149413, 0.021418399810791015, 0.02141900825500488, 0.021458080291748047, 0.02142201614379883, 0.021486400604248047, 0.021439199447631837, 0.02140889549255371, 0.02141596794128418, 0.021504831314086915, 0.021515552520751952]",tokens/s,11885.289167182545,kWh,2.5583993053998213e-07,1.4018601528472834e-07,1.255946170701312e-06,1.6519721165260226e-06,tokens/kWh,154966296.0040448,MB,1213.210624,3760.717824,0.0,3114.27072,2884.408832,s,12,10.525667846679688,0.8771389872233074,0.016420040062672146,0.8723400573730469,0.8880776062011719,0.904820425415039,0.9206031964111328,"[0.9245488891601562, 0.8745811767578126, 0.8697117919921875, 0.86923681640625, 0.8620210571289062, 0.8762533569335937, 0.8818305053710938, 0.863391845703125, 0.8626490478515625, 0.888678955078125, 0.8826654663085938, 0.8700989379882812]",tokens/s,71.82442112102933,kWh,1.0449062463410376e-05,5.725704700073816e-06,3.299637289779798e-05,4.917114006128217e-05,tokens/kWh,1281239.3595406343,,s,755,10.746529797554018,0.014233814301396048,0.0026334416179496054,0.013725695610046386,0.014288486289978027,0.01489530897140503,0.03558335289001465,"[0.015141887664794922, 0.014989312171936036, 0.014914560317993163, 0.014832639694213867, 0.014922752380371093, 0.014581760406494141, 0.01457151985168457, 0.014869503974914551, 0.015184896469116211, 0.016276479721069336, 0.015005696296691894, 0.015013888359069824, 0.014782464027404785, 0.014862367630004883, 0.014994400024414063, 0.015145983695983887, 0.015032383918762207, 0.01489094352722168, 0.015101951599121094, 0.01498316764831543, 0.014750720024108887, 0.015342592239379883, 0.014898176193237305, 0.014951423645019531, 0.01489408016204834, 0.01499443244934082, 0.015243264198303222, 0.015650815963745117, 0.015285247802734376, 0.014784511566162109, 0.014821375846862793, 0.014651391983032227, 0.014508031845092773, 0.01658880043029785, 0.017382400512695313, 0.01635327911376953, 0.014817279815673828, 0.014822400093078614, 0.014493696212768555, 0.014246912002563476, 0.014188544273376465, 0.01406771183013916, 0.013826047897338867, 0.01387827205657959, 0.013728768348693847, 0.013726719856262207, 0.013670399665832519, 0.013606911659240722, 0.014008319854736329, 0.014111743927001954, 0.014133248329162598, 0.014220288276672363, 0.014150688171386719, 0.014099424362182617, 0.013708288192749024, 0.01376153564453125, 0.013961215972900391, 0.013711359977722168, 0.013723648071289063, 0.013710335731506347, 0.013749247550964355, 0.013708288192749024, 0.03569561767578125, 0.013699071884155273, 0.01366220760345459, 0.013730815887451172, 0.013666303634643554, 0.01367961597442627, 0.013674495697021484, 0.013755392074584961, 0.01377791976928711, 0.013805567741394043, 0.01368166446685791, 0.013757439613342285, 0.013632512092590332, 0.013641728401184081, 0.014040063858032227, 0.01417420768737793, 0.014142463684082032, 0.014094335556030273, 0.014210047721862793, 0.014172160148620605, 0.01406873607635498, 0.014112768173217773, 0.014109696388244629, 0.014123007774353028, 0.014137344360351562, 0.014196736335754395, 0.014128128051757812, 0.014155776023864745, 0.014188544273376465, 0.014134271621704102, 0.014089216232299804, 0.014065664291381836, 0.014090239524841308, 0.0141844482421875, 0.01415782356262207, 0.014169152259826661, 0.01412499237060547, 0.01409331226348877, 0.014234623908996581, 0.014088191986083985, 0.013751296043395997, 0.013622271537780761, 0.013650943756103515, 0.013644800186157227, 0.01367142391204834, 0.013686783790588379, 0.013770751953125, 0.013694975852966309, 0.013662240028381348, 0.013637599945068359, 0.01368064022064209, 0.01376153564453125, 0.013728768348693847, 0.013667360305786132, 0.013665247917175292, 0.013670399665832519, 0.013648896217346192, 0.013661184310913087, 0.013647871971130371, 0.013703167915344238, 0.01377894401550293, 0.01366528034210205, 0.013729791641235351, 0.03570175933837891, 0.013714431762695312, 0.013791232109069825, 0.013775872230529786, 0.013699071884155273, 0.013599743843078613, 0.013621248245239258, 0.013619199752807617, 0.013571071624755859, 0.01364684772491455, 0.013570048332214356, 0.013591551780700683, 0.013650943756103515, 0.013628416061401367, 0.013621248245239258, 0.013585408210754395, 0.01364684772491455, 0.013606911659240722, 0.013716480255126954, 0.013649920463562011, 0.01366220760345459, 0.01365503978729248, 0.01368064022064209, 0.013727744102478028, 0.013642751693725585, 0.01366323184967041, 0.013668352127075196, 0.013651968002319336, 0.013657088279724122, 0.013635583877563476, 0.013689855575561523, 0.013709312438964843, 0.013600768089294434, 0.015032320022583008, 0.014403583526611329, 0.01419059181213379, 0.014392319679260255, 0.014271488189697265, 0.01427353572845459, 0.014139391899108887, 0.014127103805541993, 0.01419878387451172, 0.013700096130371094, 0.013667327880859375, 0.013656064033508301, 0.013594623565673827, 0.013623295783996582, 0.013666303634643554, 0.013648896217346192, 0.013644800186157227, 0.013636608123779297, 0.013622271537780761, 0.013721599578857421, 0.013615103721618652, 0.013641728401184081, 0.013643775939941406, 0.013669407844543457, 0.01371337604522705, 0.01415987205505371, 0.014181376457214356, 0.014366720199584961, 0.014116864204406738, 0.014172160148620605, 0.03561676788330078, 0.013675552368164062, 0.01362838363647461, 0.01358950424194336, 0.013742079734802246, 0.01367961597442627, 0.01377791976928711, 0.013728768348693847, 0.014028800010681153, 0.014214143753051758, 0.014103551864624024, 0.014160896301269531, 0.014270463943481445, 0.01415167999267578, 0.014129152297973633, 0.014170111656188965, 0.014051327705383301, 0.014089216232299804, 0.014072832107543945, 0.014114815711975098, 0.01420902442932129, 0.014090239524841308, 0.01415782356262207, 0.014133248329162598, 0.014193663597106934, 0.01375334358215332, 0.01367244815826416, 0.01360588836669922, 0.013690879821777344, 0.013721599578857421, 0.013657088279724122, 0.013780991554260253, 0.013645824432373046, 0.013658111572265624, 0.013641728401184081, 0.01359977626800537, 0.013726688385009766, 0.013657088279724122, 0.013682687759399414, 0.013614080429077148, 0.01369600009918213, 0.013647871971130371, 0.013511679649353027, 0.013499391555786134, 0.013599743843078613, 0.013641728401184081, 0.013705216407775878, 0.013640704154968262, 0.013642784118652343, 0.013652959823608398, 0.013631487846374512, 0.013616191864013671, 0.013641695976257324, 0.01362224006652832, 0.013676544189453126, 0.013698047637939453, 0.013702143669128418, 0.013676544189453126, 0.01375436782836914, 0.013601792335510255, 0.013647904396057129, 0.0136212158203125, 0.0136878080368042, 0.03556966400146484, 0.013666303634643554, 0.01377996826171875, 0.013626367568969726, 0.013684736251831055, 0.013620223999023438, 0.01358950424194336, 0.013627391815185547, 0.013647871971130371, 0.013622271537780761, 0.013743103981018067, 0.013584383964538574, 0.013596672058105469, 0.013613056182861329, 0.013645824432373046, 0.01367142391204834, 0.01386086368560791, 0.013732864379882812, 0.013669376373291015, 0.013646880149841308, 0.013659104347229004, 0.01367961597442627, 0.01360588836669922, 0.013590527534484862, 0.013643775939941406, 0.01369600009918213, 0.013646880149841308, 0.01371235179901123, 0.013693984031677247, 0.013668319702148438, 0.01366323184967041, 0.013702143669128418, 0.013675519943237305, 0.013614080429077148, 0.01376255989074707, 0.013717503547668456, 0.013640704154968262, 0.013688863754272461, 0.013659104347229004, 0.013824000358581542, 0.013648896217346192, 0.01368883228302002, 0.013689855575561523, 0.013651968002319336, 0.013698047637939453, 0.013592576026916504, 0.013670399665832519, 0.013592576026916504, 0.01367142391204834, 0.013642751693725585, 0.013636608123779297, 0.01377996826171875, 0.013789183616638183, 0.013644800186157227, 0.013722687721252442, 0.013691840171813964, 0.01368883228302002, 0.013774847984313965, 0.013694975852966309, 0.013687871932983398, 0.013731776237487793, 0.013745152473449707, 0.013633536338806153, 0.03561369705200195, 0.013724672317504882, 0.01375334358215332, 0.013650943756103515, 0.01368064022064209, 0.013674495697021484, 0.013675519943237305, 0.013664256095886231, 0.013614080429077148, 0.013616127967834473, 0.01367347240447998, 0.01377791976928711, 0.013664287567138672, 0.013667296409606934, 0.01362230396270752, 0.013802463531494141, 0.013675519943237305, 0.013667391777038575, 0.013951935768127442, 0.014205951690673829, 0.01368166446685791, 0.013644800186157227, 0.013705216407775878, 0.01366220760345459, 0.013842432022094727, 0.014475263595581055, 0.014227456092834472, 0.014050304412841797, 0.014050304412841797, 0.014128128051757812, 0.01426636791229248, 0.014105600357055664, 0.01417420768737793, 0.014413887977600098, 0.0142806396484375, 0.01428889560699463, 0.014123007774353028, 0.014123007774353028, 0.01417523193359375, 0.014158847808837891, 0.0141845121383667, 0.014145471572875977, 0.014064640045166015, 0.014136320114135742, 0.014078975677490235, 0.01417728042602539, 0.014263296127319336, 0.014130175590515137, 0.014100480079650878, 0.01408409595489502, 0.014127103805541993, 0.01367347240447998, 0.013684736251831055, 0.01366323184967041, 0.013763584136962891, 0.013631487846374512, 0.013689855575561523, 0.013660160064697266, 0.013649920463562011, 0.01377996826171875, 0.013611007690429687, 0.013618176460266113, 0.013832192420959472, 0.035775489807128906, 0.013590527534484862, 0.013600768089294434, 0.014013440132141113, 0.014204992294311523, 0.014073792457580567, 0.014221311569213867, 0.014126079559326172, 0.014171135902404786, 0.014234623908996581, 0.01425715160369873, 0.014222335815429688, 0.014213120460510254, 0.014103551864624024, 0.014113792419433594, 0.014147583961486816, 0.014219264030456542, 0.014211071968078613, 0.014287872314453125, 0.014163999557495118, 0.014477279663085938, 0.013888511657714844, 0.014135295867919923, 0.015129599571228027, 0.014427136421203614, 0.014311424255371094, 0.014217215538024902, 0.014150655746459961, 0.014260224342346191, 0.014196736335754395, 0.014136320114135742, 0.014379008293151856, 0.014109696388244629, 0.01415782356262207, 0.014148608207702636, 0.014120960235595703, 0.014178303718566895, 0.014199808120727539, 0.014183423995971679, 0.013792256355285644, 0.013719552040100098, 0.013796352386474609, 0.013708288192749024, 0.013682687759399414, 0.013692928314208984, 0.013609984397888183, 0.013701120376586913, 0.013642751693725585, 0.01370419216156006, 0.013705216407775878, 0.013728768348693847, 0.013648896217346192, 0.01369600009918213, 0.013634559631347656, 0.013731840133666993, 0.013686783790588379, 0.01367961597442627, 0.013661184310913087, 0.013715456008911133, 0.013645824432373046, 0.01364684772491455, 0.013712384223937989, 0.013640704154968262, 0.035535873413085936, 0.01380352020263672, 0.014193663597106934, 0.013634559631347656, 0.013725695610046386, 0.013652992248535157, 0.013664256095886231, 0.01374617576599121, 0.013693951606750488, 0.013674495697021484, 0.013640704154968262, 0.013667327880859375, 0.01365401554107666, 0.01358131217956543, 0.013622271537780761, 0.013608960151672364, 0.013625344276428223, 0.013627391815185547, 0.013637663841247559, 0.013680607795715331, 0.013718527793884277, 0.013695039749145509, 0.013697983741760254, 0.013642751693725585, 0.013706239700317382, 0.013730815887451172, 0.013684736251831055, 0.013771776199340821, 0.013656064033508301, 0.013765631675720215, 0.013677568435668945, 0.013697024345397948, 0.013703167915344238, 0.013737983703613281, 0.013706239700317382, 0.0136878080368042, 0.013682687759399414, 0.013642751693725585, 0.013789183616638183, 0.01367961597442627, 0.013685759544372558, 0.01366528034210205, 0.013646880149841308, 0.013707232475280762, 0.013650943756103515, 0.01366323184967041, 0.013701120376586913, 0.01374617576599121, 0.013694975852966309, 0.013660160064697266, 0.0136878080368042, 0.013729791641235351, 0.013722623825073242, 0.013725695610046386, 0.01375641632080078, 0.013674495697021484, 0.013697024345397948, 0.013755392074584961, 0.013667327880859375, 0.01365503978729248, 0.01364684772491455, 0.013804544448852539, 0.013792256355285644, 0.03559942245483398, 0.01368671989440918, 0.013741056442260742, 0.013644800186157227, 0.013693951606750488, 0.01407692813873291, 0.013912063598632812, 0.013843456268310546, 0.013577216148376465, 0.013553664207458497, 0.013604864120483399, 0.013611007690429687, 0.013524991989135742, 0.013515775680541992, 0.013575167655944824, 0.013599743843078613, 0.01378713607788086, 0.01366323184967041, 0.013644800186157227, 0.013659135818481445, 0.013632512092590332, 0.013625344276428223, 0.013637696266174316, 0.01360377597808838, 0.01368064022064209, 0.013600768089294434, 0.013606911659240722, 0.013602815628051757, 0.013639679908752441, 0.013590527534484862, 0.013620223999023438, 0.013711359977722168, 0.013667327880859375, 0.013611007690429687, 0.013700096130371094, 0.013570048332214356, 0.013677568435668945, 0.013621248245239258, 0.01366528034210205, 0.013586432456970214, 0.013658111572265624, 0.013606911659240722, 0.01363257598876953, 0.013646783828735352, 0.013649920463562011, 0.013716480255126954, 0.013652992248535157, 0.013691904067993164, 0.013740032196044923, 0.01367347240447998, 0.013650943756103515, 0.013736960411071777, 0.013612031936645508, 0.01360588836669922, 0.013643775939941406, 0.013608960151672364, 0.013649920463562011, 0.013571071624755859, 0.013836288452148437, 0.013829119682312012, 0.01428377628326416, 0.014087167739868164, 0.014131199836730958, 0.035535873413085936, 0.013677568435668945, 0.013693951606750488, 0.013629440307617188, 0.013748224258422852, 0.013649920463562011, 0.014090239524841308, 0.014124032020568847, 0.014162943840026856, 0.01426636791229248, 0.014128128051757812, 0.014125056266784668, 0.014231552124023437, 0.014072832107543945, 0.014214143753051758, 0.01415167999267578, 0.014127103805541993, 0.014109696388244629, 0.014282784461975097, 0.014240768432617188, 0.014172127723693848, 0.014109696388244629, 0.01415071964263916, 0.014183423995971679, 0.014162879943847656, 0.014513152122497559, 0.01429196834564209, 0.014201888084411621, 0.014140383720397949, 0.013674495697021484, 0.013670399665832519, 0.01402064037322998, 0.014122976303100585, 0.014111743927001954, 0.014120960235595703, 0.014108672142028808, 0.014098431587219238, 0.014130175590515137, 0.014134271621704102, 0.014141440391540527, 0.014216192245483398, 0.014193663597106934, 0.01417728042602539, 0.01416806411743164, 0.014137344360351562, 0.014256128311157227, 0.014111776351928711, 0.014293984413146972, 0.014134271621704102, 0.014171135902404786, 0.014152704238891601, 0.014078975677490235, 0.014073856353759765, 0.014131199836730958, 0.014104576110839843, 0.01417625617980957, 0.01420083236694336, 0.014096384048461913, 0.014134271621704102, 0.014172160148620605, 0.014148608207702636, 0.014130175590515137, 0.014134271621704102, 0.03568742370605469, 0.013752320289611816, 0.013650943756103515, 0.013725695610046386, 0.013649920463562011, 0.013677632331848145, 0.014422975540161132, 0.015025152206420898, 0.01472000026702881, 0.014326784133911133, 0.014320639610290528, 0.014269439697265626, 0.014352383613586426, 0.014224384307861328, 0.01415782356262207, 0.014102527618408203, 0.014118911743164063, 0.014212096214294433, 0.014163968086242675, 0.01415167999267578, 0.01409126377105713, 0.014160896301269531, 0.014083135604858398, 0.014171072006225586, 0.013800448417663574, 0.013749247550964355, 0.013637632369995116, 0.014172160148620605, 0.014121983528137207, 0.014153727531433105, 0.014170111656188965, 0.01408614444732666, 0.014138367652893067, 0.014110719680786133, 0.014460927963256836, 0.014189567565917969, 0.014127103805541993, 0.014164992332458496, 0.014167039871215821, 0.013682687759399414, 0.013724672317504882, 0.01368064022064209, 0.013661184310913087, 0.013669376373291015, 0.013664256095886231, 0.013692928314208984, 0.013725695610046386, 0.013697024345397948, 0.013908991813659668, 0.013750271797180176, 0.013926400184631347, 0.01417728042602539, 0.01428275203704834, 0.014131199836730958, 0.014129152297973633, 0.014153727531433105, 0.013750271797180176, 0.01368064022064209, 0.013763584136962891, 0.01368883228302002, 0.013830143928527832, 0.013643808364868164, 0.01371235179901123, 0.035705856323242184, 0.013660160064697266, 0.013716480255126954, 0.013765695571899415, 0.013714367866516114, 0.013736960411071777, 0.013709312438964843, 0.013668352127075196, 0.013707263946533203, 0.013705216407775878, 0.01376153564453125, 0.013666303634643554, 0.013706239700317382, 0.013651968002319336, 0.01388646411895752, 0.013697024345397948, 0.013745152473449707, 0.014042112350463867, 0.01460531234741211, 0.015430656433105469, 0.014432255744934081, 0.014278656005859374, 0.014204928398132324, 0.014324735641479493, 0.014221311569213867, 0.014196736335754395, 0.014199808120727539, 0.013817888259887695, 0.013739999771118163, 0.013835264205932616, 0.01369600009918213, 0.013685759544372558, 0.013733887672424316, 0.013622271537780761, 0.013697024345397948, 0.013700096130371094, 0.013659135818481445, 0.013737983703613281, 0.013637632369995116, 0.013701120376586913, 0.013643775939941406, 0.013712384223937989, 0.013637632369995116, 0.01376972770690918, 0.013708288192749024, 0.013619199752807617, 0.01363865566253662, 0.01368883228302002, 0.013641728401184081, 0.013674559593200684, 0.013691840171813964, 0.013660160064697266, 0.013609984397888183, 0.01365503978729248, 0.013669376373291015, 0.013669376373291015, 0.013708288192749024, 0.013633567810058594, 0.01365500831604004, 0.013661184310913087, 0.013628416061401367, 0.013616127967834473, 0.013666303634643554]",tokens/s,70.25523719962543,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,mistralai/Mistral-7B-v0.1,mistralai/Mistral-7B-v0.1,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.1,,,,MB,1270.10816,15987.113984,0.0,15340.66688,15114.865664,s,10,0.8908641891479493,0.08908641891479493,0.002248653393122016,0.08837084579467774,0.08936671295166015,0.09259126205444335,0.09517090133666993,"[0.09581581115722657, 0.08845929718017578, 0.08830262756347657, 0.08810169219970704, 0.088650146484375, 0.0884567642211914, 0.08838969421386719, 0.08835199737548828, 0.08812274932861328, 0.08821340942382813]",tokens/s,2873.614217727693,kWh,1.0468745888041418e-06,5.736383472775847e-07,6.696745770371453e-06,8.31725870645318e-06,tokens/kWh,30779372.030519523,MB,1270.10816,15987.113984,0.0,15340.66688,15134.114304,s,10,31.351924072265625,3.1351924072265627,0.01343642581392157,3.135545288085938,3.1540085205078126,3.1547807495117186,3.1553985327148437,"[3.155552978515625, 3.132075439453125, 3.1538369140625, 3.13901513671875, 3.146458984375, 3.139015869140625, 3.125260986328125, 3.11593408203125, 3.116068603515625, 3.128705078125]",tokens/s,20.09446050417388,kWh,3.680663284386857e-05,2.0171711897392094e-05,0.00013959245459422792,0.00019657079933548858,tokens/kWh,320495.21196928906,,s,629,32.095909866332995,0.05102688373025915,0.010553122864185334,0.04955648040771484,0.05033717651367187,0.050655847167968754,0.13825425720214846,"[0.05150822448730469, 0.05064396667480469, 0.05129523086547851, 0.05078326416015625, 0.05084156799316406, 0.05203763198852539, 0.051097599029541016, 0.05099622344970703, 0.05120716857910156, 0.05065523147583008, 0.050206718444824217, 0.05030400085449219, 0.04976435089111328, 0.049514495849609375, 0.049509376525878904, 0.051076095581054685, 0.05063884735107422, 0.05023641586303711, 0.05009100723266602, 0.050151424407958986, 0.050012161254882816, 0.05011663818359375, 0.04939878463745117, 0.04968854522705078, 0.04967833709716797, 0.049827838897705076, 0.05010227203369141, 0.049858558654785154, 0.05008486557006836, 0.05005926513671875, 0.04980633544921875, 0.04946432113647461, 0.04954111862182617, 0.049481727600097655, 0.04949913787841797, 0.04947558212280274, 0.04950630569458008, 0.049498111724853515, 0.049516544342041016, 0.04955340957641602, 0.04962611389160156, 0.049509376525878904, 0.04947558212280274, 0.04943462371826172, 0.04942540740966797, 0.04960870361328125, 0.04969472122192383, 0.050229248046875, 0.05003366470336914, 0.05021798324584961, 0.050135040283203126, 0.050260990142822266, 0.050267135620117184, 0.05017910385131836, 0.05020975875854492, 0.050331649780273435, 0.05035212707519531, 0.05030092620849609, 0.04970700836181641, 0.04964556884765625, 0.04923494338989258, 0.04957183837890625, 0.13839155578613282, 0.04947251129150391, 0.04955648040771484, 0.04964044952392578, 0.050106433868408205, 0.050107326507568356, 0.05006748962402344, 0.049955806732177734, 0.04954214477539062, 0.04988313674926758, 0.05001728057861328, 0.04995993423461914, 0.05014220809936523, 0.05013401412963867, 0.049616897583007816, 0.04947763061523437, 0.04960768127441406, 0.05022719955444336, 0.04988415908813477, 0.04981248092651367, 0.050051071166992187, 0.049926143646240234, 0.05006335830688476, 0.049353729248046874, 0.04936294555664063, 0.049729534149169925, 0.049554431915283206, 0.04951039886474609, 0.04948787307739258, 0.049462272644042966, 0.04944588851928711, 0.04953804779052735, 0.04948179244995117, 0.049499073028564454, 0.04950527954101563, 0.04995993423461914, 0.05023027038574219, 0.05015039825439453, 0.04978278350830078, 0.050282497406005856, 0.05033369445800781, 0.049585151672363284, 0.0496629753112793, 0.04955750274658203, 0.049522689819335934, 0.04967833709716797, 0.04951039886474609, 0.04962713623046875, 0.04952166366577149, 0.04959231948852539, 0.04946124649047851, 0.04948582458496094, 0.049503231048583986, 0.049539070129394534, 0.04958924865722656, 0.04955852890014648, 0.049509376525878904, 0.049473567962646486, 0.04956361770629883, 0.049535999298095705, 0.04949401473999023, 0.0496445426940918, 0.04965785598754883, 0.13929983520507813, 0.05099520111083984, 0.05012582397460937, 0.05029580688476563, 0.05021084976196289, 0.04981449508666992, 0.05031942367553711, 0.05046572875976563, 0.05136383819580078, 0.050411518096923826, 0.050383872985839843, 0.050495487213134765, 0.0498524169921875, 0.049560577392578124, 0.04931071853637695, 0.04940390396118164, 0.04941823959350586, 0.04949708938598633, 0.04954111862182617, 0.049980415344238284, 0.05037875366210937, 0.05026508712768555, 0.05039718246459961, 0.05020876693725586, 0.05030297470092773, 0.0502026252746582, 0.050229248046875, 0.050078720092773435, 0.05004390335083008, 0.04976025772094726, 0.05017804718017578, 0.05026924896240234, 0.05031315231323242, 0.05025484848022461, 0.05010227203369141, 0.05037875366210937, 0.05018316650390625, 0.05012070465087891, 0.049895423889160156, 0.05019340896606445, 0.04981350326538086, 0.049498111724853515, 0.04962303924560547, 0.04949606323242187, 0.04951347351074219, 0.049584129333496096, 0.04954111862182617, 0.04955750274658203, 0.049563648223876954, 0.0496517105102539, 0.050270206451416014, 0.05033062362670898, 0.05029683303833008, 0.0502026252746582, 0.0500316162109375, 0.050307071685791016, 0.05026201629638672, 0.04968038558959961, 0.05005721664428711, 0.050129920959472656, 0.04984115219116211, 0.05033267211914062, 0.050070526123046875, 0.13817240905761718, 0.05058867263793945, 0.051151905059814456, 0.05071356964111328, 0.05061529541015625, 0.05039820861816406, 0.05039923095703125, 0.05047808074951172, 0.05027635192871094, 0.05030297470092773, 0.050301952362060545, 0.050146305084228515, 0.05001728057861328, 0.05039718246459961, 0.05006950378417969, 0.05029785537719727, 0.050298881530761716, 0.04991078567504883, 0.050282497406005856, 0.0502026252746582, 0.0501288948059082, 0.04972544097900391, 0.0512542724609375, 0.0502108154296875, 0.04959539031982422, 0.049509376525878904, 0.04960665512084961, 0.04942950439453125, 0.04935168075561523, 0.049452030181884765, 0.049486846923828126, 0.04944076919555664, 0.04976537704467773, 0.04984524917602539, 0.04969574356079102, 0.04945619201660156, 0.04941817474365234, 0.04943360137939453, 0.04935475158691406, 0.04957183837890625, 0.04941823959350586, 0.049462272644042966, 0.049500160217285157, 0.04948889541625977, 0.049519615173339845, 0.049509376525878904, 0.04964966583251953, 0.04968243026733398, 0.04969267272949219, 0.04940902328491211, 0.04943462371826172, 0.04943462371826172, 0.04942233657836914, 0.049460224151611325, 0.049363967895507815, 0.04937318420410156, 0.04945407867431641, 0.04951859283447266, 0.04944793701171875, 0.049460224151611325, 0.04945510482788086, 0.04940288162231445, 0.049838081359863284, 0.13938482666015625, 0.05005721664428711, 0.050495487213134765, 0.050092033386230465, 0.04980838394165039, 0.05002342224121094, 0.05002035140991211, 0.050165760040283204, 0.04978278350830078, 0.04995072174072265, 0.05005926513671875, 0.05003776168823242, 0.049748992919921874, 0.049593345642089844, 0.04950531387329102, 0.04947657775878906, 0.049430526733398435, 0.05007257461547852, 0.05000396728515625, 0.050071617126464844, 0.04999980926513672, 0.04993228912353516, 0.05018931198120117, 0.04961996841430664, 0.04951551818847656, 0.04973875045776367, 0.04973161697387695, 0.04977967834472656, 0.05037875366210937, 0.05026201629638672, 0.05004185485839844, 0.050200576782226565, 0.05030297470092773, 0.049873920440673826, 0.05007462310791016, 0.04976128005981445, 0.05007462310791016, 0.049683456420898435, 0.04982271957397461, 0.0502927360534668, 0.049718273162841796, 0.04939059066772461, 0.04944076919555664, 0.04933017730712891, 0.0497520637512207, 0.04955033493041992, 0.05018624114990235, 0.049748992919921874, 0.05006950378417969, 0.05020774459838867, 0.04993228912353516, 0.049979393005371096, 0.050407424926757816, 0.049964031219482424, 0.04987596893310547, 0.04954828643798828, 0.05025689697265625, 0.05018828964233398, 0.04990259170532227, 0.04986777496337891, 0.05033574295043945, 0.049865726470947266, 0.05046169662475586, 0.13854208374023438, 0.05013708877563477, 0.05030809783935547, 0.05033267211914062, 0.050239486694335936, 0.05022719955444336, 0.05025791931152344, 0.05054054260253906, 0.050285568237304686, 0.050388992309570314, 0.05001113510131836, 0.050293758392333986, 0.05029171371459961, 0.05048012924194336, 0.05126860809326172, 0.05061427307128906, 0.04989440155029297, 0.05028044891357422, 0.04963840103149414, 0.05015654373168945, 0.04982988739013672, 0.049616897583007816, 0.049783809661865235, 0.04948582458496094, 0.05065625762939453, 0.05040435028076172, 0.05024051284790039, 0.050310142517089845, 0.05030809783935547, 0.05027328109741211, 0.05018316650390625, 0.050342910766601565, 0.050165760040283204, 0.0504535026550293, 0.05025894546508789, 0.04985139083862305, 0.049384513854980466, 0.0494376335144043, 0.0493568000793457, 0.04944179153442383, 0.049446910858154294, 0.04942748641967774, 0.049387489318847656, 0.04940083312988281, 0.04912025451660156, 0.04925337600708008, 0.049223678588867184, 0.049276927947998046, 0.049165313720703124, 0.049355777740478515, 0.04916428756713867, 0.049255424499511716, 0.049159168243408206, 0.04920217514038086, 0.049099777221679686, 0.04922880172729492, 0.049261566162109374, 0.049363967895507815, 0.04929024124145508, 0.04944076919555664, 0.0493568000793457, 0.04930460739135742, 0.04947452926635742, 0.13843763732910155, 0.049470462799072266, 0.04944076919555664, 0.049476608276367184, 0.049452030181884765, 0.04938751983642578, 0.04941312026977539, 0.04939980697631836, 0.05066649627685547, 0.05098086547851562, 0.05060095977783203, 0.050267135620117184, 0.05009100723266602, 0.049555454254150394, 0.04936703872680664, 0.04934860610961914, 0.04946432113647461, 0.04939980697631836, 0.049463294982910154, 0.049554431915283206, 0.04955750274658203, 0.04935168075561523, 0.0494284782409668, 0.04946636962890625, 0.04936601638793945, 0.049503231048583986, 0.049317886352539066, 0.049459201812744144, 0.04959027099609375, 0.050514942169189454, 0.04999174499511719, 0.049602497100830076, 0.04951551818847656, 0.049549312591552735, 0.04944486236572266, 0.049397792816162106, 0.049336288452148436, 0.049463294982910154, 0.049481727600097655, 0.04997836685180664, 0.05016371154785156, 0.049331199645996096, 0.04944998550415039, 0.04951859283447266, 0.04943360137939453, 0.049517566680908204, 0.049484798431396484, 0.04974796676635742, 0.04963328170776367, 0.04942745590209961, 0.04949401473999023, 0.049438751220703125, 0.04965270233154297, 0.04947251129150391, 0.04961177444458008, 0.04948787307739258, 0.04948992156982422, 0.049426433563232425, 0.04952166366577149, 0.049484798431396484, 0.04940697479248047, 0.049462272644042966, 0.049463294982910154, 0.13833319091796875, 0.049393665313720705, 0.04953702545166016, 0.049364990234375, 0.04948787307739258, 0.049535999298095705, 0.049337345123291014, 0.04940492630004883, 0.04935475158691406, 0.049446910858154294, 0.049465343475341796, 0.04941823959350586, 0.049301502227783206, 0.04938956832885742, 0.049307647705078124, 0.04935065460205078, 0.049430526733398435, 0.04942745590209961, 0.049324031829833984, 0.04948992156982422, 0.04935782241821289, 0.049377281188964846, 0.04944486236572266, 0.049430526733398435, 0.04940492630004883, 0.04945305633544922, 0.04943462371826172, 0.04951551818847656, 0.04935475158691406, 0.04977664184570312, 0.04969369506835938, 0.049484798431396484, 0.04934963226318359, 0.04940806579589844, 0.049469375610351564, 0.05001523208618164, 0.049430526733398435, 0.04937420654296875, 0.04934348678588867, 0.04939263916015625, 0.04958617782592773, 0.04946944046020508, 0.049391616821289064, 0.04938137435913086, 0.049375232696533204, 0.049344512939453126, 0.04950527954101563, 0.049585151672363284, 0.049463294982910154, 0.04938547134399414, 0.04948992156982422, 0.04953497695922852, 0.049648670196533205, 0.04945711898803711, 0.04945305633544922, 0.04946739196777344, 0.04941516876220703, 0.04949401473999023, 0.04941823959350586, 0.04942335891723633, 0.04950732803344727, 0.04944076919555664, 0.049481727600097655, 0.13801370239257812, 0.049347583770751956, 0.04943769454956055, 0.04936908721923828, 0.04947455978393555, 0.04968038558959961, 0.0495206413269043, 0.04937420654296875, 0.04921446228027344, 0.04945612716674805, 0.04941619110107422, 0.04937011337280273, 0.04930252838134765, 0.049442817687988284, 0.04929024124145508, 0.049400863647460935, 0.049390560150146486, 0.049442817687988284, 0.049364990234375, 0.04983603286743164, 0.04946944046020508, 0.05073920059204102, 0.05070848083496094, 0.04989132690429687, 0.049511425018310545, 0.04998758316040039, 0.0494919662475586, 0.049397758483886715, 0.049452030181884765, 0.04941209411621094, 0.04955852890014648, 0.04943155288696289, 0.04941209411621094, 0.049552383422851565, 0.049271808624267575, 0.04940083312988281, 0.04940492630004883, 0.04987395095825195, 0.04926153564453125, 0.04942134475708008, 0.04945612716674805, 0.049363937377929684, 0.049228832244873046, 0.04926358413696289, 0.049160190582275394, 0.04935987091064453, 0.049206272125244144, 0.049239040374755856, 0.049203201293945314, 0.04919705581665039, 0.049102848052978515, 0.049135616302490234, 0.04922470474243164, 0.049353729248046874, 0.04939468765258789, 0.04946124649047851, 0.04940083312988281, 0.0494284782409668, 0.04942335891723633, 0.0494571533203125, 0.049468414306640625, 0.049468414306640625, 0.04942950439453125, 0.13828608703613282, 0.04951551818847656, 0.04944588851928711, 0.04945407867431641, 0.04940902328491211, 0.04939263916015625, 0.04941519927978515, 0.04929840087890625, 0.049401920318603514, 0.04938643264770508, 0.04948992156982422, 0.04942233657836914, 0.0493568000793457, 0.04933324813842774, 0.04933017730712891, 0.04942335891723633, 0.05143040084838867, 0.050685951232910156, 0.05024870300292969, 0.04951555252075195, 0.04936291122436524, 0.049509376525878904, 0.049495040893554686, 0.04945407867431641, 0.0494284782409668, 0.049500160217285157, 0.04971212768554688, 0.050369537353515625, 0.04944588851928711, 0.04949401473999023, 0.04957593536376953, 0.04948992156982422, 0.04967119979858398, 0.04950729751586914, 0.049436702728271484, 0.0494601936340332, 0.05009100723266602, 0.050715648651123046, 0.04955750274658203, 0.049914878845214845, 0.049552383422851565, 0.0499947509765625, 0.049535999298095705, 0.049833984375, 0.049620990753173826, 0.04941516876220703, 0.04950732803344727, 0.04948889541625977, 0.049704959869384766, 0.04957593536376953, 0.049479679107666014, 0.049522689819335934, 0.04950835037231445, 0.04954009628295898, 0.050326526641845705, 0.049444961547851565, 0.049495967864990234, 0.04948793411254883, 0.04950214385986328, 0.05026406478881836, 0.049793022155761715, 0.05005926513671875, 0.049838081359863284]",tokens/s,19.597512661879367,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,google/recurrentgemma-7b,google/recurrentgemma-7b,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/google/recurrentgemma-7b/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6694704f-2baf8c5f7c21869a7a6e8b6b;5daef34a-af20-4316-b5f7-49f6f89eecf3) Repository Not Found for url: https://huggingface.co/google/recurrentgemma-7b/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: google/recurrentgemma-7b is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,mistralai/Mixtral-8x7B-v0.1,mistralai/Mixtral-8x7B-v0.1,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ self.load_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 250, in load_model_with_no_weights self.pretrained_model = self.automodel_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3710, in from_pretrained model = cls(config, *model_args, **model_kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/mixtral/modeling_mixtral.py"", line 1323, in __init__ self.model = MixtralModel(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/mixtral/modeling_mixtral.py"", line 1087, in __init__ [MixtralDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/mixtral/modeling_mixtral.py"", line 1087, in [MixtralDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/mixtral/modeling_mixtral.py"", line 877, in __init__ self.block_sparse_moe = MixtralSparseMoeBlock(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/mixtral/modeling_mixtral.py"", line 824, in __init__ self.experts = nn.ModuleList([MixtralBlockSparseTop2MLP(config) for _ in range(self.num_experts)]) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/mixtral/modeling_mixtral.py"", line 824, in self.experts = nn.ModuleList([MixtralBlockSparseTop2MLP(config) for _ in range(self.num_experts)]) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/mixtral/modeling_mixtral.py"", line 791, in __init__ self.w2 = nn.Linear(self.ffn_dim, self.hidden_dim, bias=False) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/linear.py"", line 98, in __init__ self.weight = Parameter(torch.empty((out_features, in_features), **factory_kwargs)) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_device.py"", line 78, in __torch_function__ return func(*args, **kwargs) torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 112.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,t,t,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/t/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669473b5-7f93ef0e47e0c47c21f285f3;03ef5cd4-75af-4c13-8d4a-d3a8905c7b6f) Repository Not Found for url: https://huggingface.co/t/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: t is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-72B,Qwen/Qwen-72B,cuda,0,42,,,True,,,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",qwen,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-30b,facebook/opt-30b,cuda,0,42,,,True,,,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",opt,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-MoE-A2.7B,Qwen/Qwen1.5-MoE-A2.7B,cuda,0,42,,,True,,,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",qwen2_moe,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-1.3b,EleutherAI/pythia-1.3b,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1273.880576,3852.992512,0.0,3206.545408,3105.82784,s,10,0.22610268783569334,0.022610268783569337,0.0003043157777445479,0.022503392219543457,0.022695283699035645,0.023103321933746337,0.023429752521514892,"[0.02351136016845703, 0.022567775726318358, 0.022604608535766603, 0.022471967697143554, 0.02250752067565918, 0.022471071243286133, 0.022437023162841796, 0.022499263763427733, 0.022476512908935545, 0.02255558395385742]",tokens/s,11322.289109010182,kWh,2.67829261879742e-07,1.4675720045043093e-07,1.2030231342494404e-06,1.6176095965796133e-06,tokens/kWh,158258210.4738401,MB,1273.880576,3852.992512,0.0,3206.545408,3189.19168,s,10,10.90676721191406,1.0906767211914064,0.031225445779376405,1.0901775512695313,1.1228399291992188,1.1271606872558593,1.1306172937011718,"[1.1314814453125, 1.1189879150390625, 1.1182877197265626, 1.117508056640625, 1.1218797607421875, 1.0628470458984376, 1.06013671875, 1.0576103515625, 1.055681884765625, 1.0623463134765625]",tokens/s,57.76230369268506,kWh,1.248988201618795e-05,6.843971163420882e-06,3.5835700941150545e-05,5.516955412075939e-05,tokens/kWh,1141934.1882318049,,s,629,11.090413570404062,0.01763181807695398,0.002740089693097124,0.017496063232421876,0.01792020492553711,0.018042905044555663,0.039768857116699224,"[0.017466367721557616, 0.018114559173583983, 0.018130943298339842, 0.01821388816833496, 0.017960960388183594, 0.017921024322509766, 0.017947647094726564, 0.01789952087402344, 0.017757183074951173, 0.017733631134033204, 0.01826918411254883, 0.017854463577270507, 0.017765375137329103, 0.018148351669311523, 0.017920000076293945, 0.01775103950500488, 0.01787494468688965, 0.01783296012878418, 0.01784217643737793, 0.01778483200073242, 0.01794867134094238, 0.01788313674926758, 0.017745920181274414, 0.017741823196411134, 0.017960960388183594, 0.017898496627807618, 0.01773465538024902, 0.01809721565246582, 0.017894336700439453, 0.017936384201049805, 0.01816268730163574, 0.01802444839477539, 0.018181119918823242, 0.01823641586303711, 0.01800499153137207, 0.017970176696777345, 0.01785753631591797, 0.01773161506652832, 0.01773257637023926, 0.01782374382019043, 0.017904640197753906, 0.01801523208618164, 0.017771520614624024, 0.01805516815185547, 0.017889280319213868, 0.01785958480834961, 0.018250751495361327, 0.018108448028564452, 0.017756128311157228, 0.01787596893310547, 0.017762304306030274, 0.017770496368408203, 0.017821695327758787, 0.017872896194458008, 0.017944576263427735, 0.01802137565612793, 0.01820979118347168, 0.01887846374511719, 0.01822412872314453, 0.0178288631439209, 0.01777663993835449, 0.01786476707458496, 0.039884735107421875, 0.016793600082397463, 0.016763904571533202, 0.016755712509155272, 0.017069055557250978, 0.017730560302734375, 0.017739776611328126, 0.017765375137329103, 0.01774284744262695, 0.017736703872680663, 0.01776639938354492, 0.01781760025024414, 0.017744895935058593, 0.017698816299438477, 0.01780735969543457, 0.01769267272949219, 0.0177838077545166, 0.01768454360961914, 0.017737728118896484, 0.01772640037536621, 0.01784012794494629, 0.01783296012878418, 0.017693695068359376, 0.017786880493164063, 0.017667072296142578, 0.017745920181274414, 0.017744895935058593, 0.01776639938354492, 0.017777664184570312, 0.01776639938354492, 0.01778278350830078, 0.01775923156738281, 0.01777561569213867, 0.017789951324462892, 0.017779712677001954, 0.01776742362976074, 0.01782681655883789, 0.017829887390136717, 0.0178606071472168, 0.0178155517578125, 0.017803264617919923, 0.017862655639648437, 0.01785958480834961, 0.01784217643737793, 0.017946687698364258, 0.017865663528442384, 0.01780940818786621, 0.017900543212890627, 0.01784832000732422, 0.018039871215820312, 0.017830848693847656, 0.017778688430786133, 0.0184453125, 0.01901568031311035, 0.018107391357421874, 0.017942527770996093, 0.017804288864135744, 0.017833984375, 0.017772544860839845, 0.0178606071472168, 0.017768447875976562, 0.017780736923217775, 0.017821695327758787, 0.03989299011230469, 0.017688575744628905, 0.017720319747924804, 0.017738752365112305, 0.017715200424194336, 0.01779814338684082, 0.017622016906738282, 0.01779302406311035, 0.017725439071655275, 0.017999872207641602, 0.01779199981689453, 0.01780121612548828, 0.017941503524780272, 0.017957887649536132, 0.01776639938354492, 0.017789951324462892, 0.017803264617919923, 0.017773567199707033, 0.01785036849975586, 0.01777663993835449, 0.01763942337036133, 0.01774393653869629, 0.01773356819152832, 0.017704959869384765, 0.01770086479187012, 0.017743871688842772, 0.017913856506347657, 0.017752063751220702, 0.017836032867431642, 0.01780019187927246, 0.01787596893310547, 0.017915903091430666, 0.017729536056518554, 0.017819648742675782, 0.017758207321166994, 0.01806540870666504, 0.017719295501708983, 0.017896448135375977, 0.017596416473388672, 0.017488895416259767, 0.017496063232421876, 0.01759129524230957, 0.017819648742675782, 0.017992704391479493, 0.017768447875976562, 0.017698816299438477, 0.017746944427490235, 0.017732608795166017, 0.017715200424194336, 0.017770496368408203, 0.017735679626464843, 0.017773567199707033, 0.017863679885864257, 0.017510400772094727, 0.01742438316345215, 0.017501184463500977, 0.017483776092529296, 0.017735679626464843, 0.01789952087402344, 0.0176629753112793, 0.017881088256835938, 0.017917951583862304, 0.017812480926513673, 0.039839744567871094, 0.016785408020019533, 0.016777215957641603, 0.016749568939208984, 0.01696873664855957, 0.01807049560546875, 0.017889280319213868, 0.017779712677001954, 0.017724416732788087, 0.017715200424194336, 0.017777664184570312, 0.01781760025024414, 0.017763328552246094, 0.017902591705322265, 0.017829887390136717, 0.017699840545654297, 0.017752063751220702, 0.017687551498413084, 0.01776742362976074, 0.01781862449645996, 0.017731584548950196, 0.017753087997436523, 0.017999872207641602, 0.01774284744262695, 0.01770086479187012, 0.017740800857543947, 0.017785856246948242, 0.017755136489868165, 0.01780019187927246, 0.017721343994140625, 0.01789030456542969, 0.01781760025024414, 0.01779916763305664, 0.017745920181274414, 0.017777664184570312, 0.01777459144592285, 0.01781350326538086, 0.017770496368408203, 0.017836032867431642, 0.01800499153137207, 0.017757183074951173, 0.01776950454711914, 0.017792991638183594, 0.01781657600402832, 0.017844224929809572, 0.01780735969543457, 0.017812480926513673, 0.01807673645019531, 0.01772640037536621, 0.017780736923217775, 0.01789446449279785, 0.017663936614990234, 0.017945600509643556, 0.017923072814941408, 0.01785651206970215, 0.01764352035522461, 0.017904640197753906, 0.017718271255493166, 0.017789951324462892, 0.017757183074951173, 0.01779097557067871, 0.018036735534667968, 0.017819648742675782, 0.04085452651977539, 0.017826847076416016, 0.017741792678833006, 0.0178155517578125, 0.017718271255493166, 0.017940511703491212, 0.017727455139160155, 0.017779712677001954, 0.017723392486572266, 0.01784524726867676, 0.017955839157104494, 0.01773673629760742, 0.017733600616455077, 0.017861631393432616, 0.017730560302734375, 0.017888256072998047, 0.01784934425354004, 0.017887231826782226, 0.017836032867431642, 0.01777561569213867, 0.017702911376953127, 0.01749504089355469, 0.017521696090698244, 0.017771488189697267, 0.017819648742675782, 0.017758207321166994, 0.017675264358520508, 0.017546239852905272, 0.017547264099121093, 0.017768447875976562, 0.017771520614624024, 0.017918975830078124, 0.01783193588256836, 0.01779302406311035, 0.018044927597045898, 0.017740800857543947, 0.01779814338684082, 0.017821695327758787, 0.017514495849609374, 0.017582143783569336, 0.01771001625061035, 0.01784524726867676, 0.017880064010620117, 0.018009151458740234, 0.01780940818786621, 0.01795577621459961, 0.017804288864135744, 0.017821695327758787, 0.01783500862121582, 0.01779097557067871, 0.017797119140625, 0.01798041534423828, 0.017929216384887696, 0.018099264144897462, 0.017889215469360353, 0.017804288864135744, 0.01796505546569824, 0.017839103698730468, 0.01788313674926758, 0.017862655639648437, 0.01774185562133789, 0.017760223388671875, 0.017771520614624024, 0.039783489227294924, 0.016850879669189453, 0.016776191711425782, 0.016773120880126953, 0.017558528900146485, 0.01782681655883789, 0.017107967376708985, 0.017753087997436523, 0.017699840545654297, 0.017762304306030274, 0.017876991271972655, 0.016827392578125, 0.01680998420715332, 0.01684787178039551, 0.01679769515991211, 0.016760831832885743, 0.01681203269958496, 0.01678950309753418, 0.01680998420715332, 0.01680998420715332, 0.016784383773803712, 0.01677107238769531, 0.01678643226623535, 0.01677004814147949, 0.01681305694580078, 0.016959487915039064, 0.016897024154663084, 0.01680998420715332, 0.01678745651245117, 0.01680179214477539, 0.016718847274780273, 0.01660006332397461, 0.016630783081054687, 0.016703487396240235, 0.01681407928466797, 0.01685196876525879, 0.0167956485748291, 0.01679769515991211, 0.01677516746520996, 0.016698368072509767, 0.016590848922729492, 0.0166430721282959, 0.016652288436889647, 0.016694271087646484, 0.01664614486694336, 0.016673791885375978, 0.016683008193969725, 0.016744447708129884, 0.016753664016723634, 0.016662527084350585, 0.016730112075805666, 0.01662873649597168, 0.016816160202026368, 0.016820192337036133, 0.01682329559326172, 0.01682841682434082, 0.01680281639099121, 0.0168222713470459, 0.016740352630615234, 0.0166430721282959, 0.01678335952758789, 0.016785408020019533, 0.01681203269958496, 0.039726112365722654, 0.016792543411254884, 0.016727039337158203, 0.0167587833404541, 0.0167956485748291, 0.016753664016723634, 0.01680384063720703, 0.01679974365234375, 0.01677926445007324, 0.01685196876525879, 0.01676595115661621, 0.01681100845336914, 0.01679871940612793, 0.01684377670288086, 0.01680384063720703, 0.016877567291259766, 0.016745471954345705, 0.01680384063720703, 0.016777215957641603, 0.016887807846069337, 0.01681305694580078, 0.01677516746520996, 0.01681612777709961, 0.016741376876831054, 0.01677414321899414, 0.01679052734375, 0.016767999649047852, 0.0167445125579834, 0.016777151107788085, 0.01661235237121582, 0.01662566375732422, 0.016678911209106445, 0.016694271087646484, 0.01676595115661621, 0.016825344085693358, 0.01680793571472168, 0.01681715202331543, 0.01679462432861328, 0.016759807586669923, 0.017071104049682616, 0.01713667106628418, 0.01688675117492676, 0.01680793571472168, 0.016829439163208008, 0.016763904571533202, 0.01677004814147949, 0.016829439163208008, 0.016891904830932617, 0.016755712509155272, 0.01683660888671875, 0.01706188774108887, 0.01720832061767578, 0.016982015609741212, 0.016907264709472656, 0.016923648834228516, 0.016955392837524414, 0.016870399475097657, 0.01683046340942383, 0.016785408020019533, 0.01679155158996582, 0.016772096633911132, 0.01684480094909668, 0.01679974365234375, 0.03984076690673828, 0.01683660888671875, 0.016760831832885743, 0.016788480758666992, 0.016874496459960937, 0.016756736755371093, 0.016751615524291993, 0.016784383773803712, 0.016777215957641603, 0.01675775909423828, 0.016752639770507814, 0.016749568939208984, 0.016781312942504883, 0.01678643226623535, 0.016759807586669923, 0.016773120880126953, 0.01681305694580078, 0.01677516746520996, 0.01681715202331543, 0.016782400131225585, 0.016764863967895508, 0.01677824020385742, 0.0168089599609375, 0.016777215957641603, 0.016738304138183592, 0.016841728210449217, 0.01683865547180176, 0.01677107238769531, 0.01676288032531738, 0.01678335952758789, 0.016759807586669923, 0.016740352630615234, 0.01677414321899414, 0.016799808502197266, 0.016801727294921874, 0.016827392578125, 0.01685196876525879, 0.01682841682434082, 0.01679155158996582, 0.01683967971801758, 0.01678233528137207, 0.01680998420715332, 0.0168089599609375, 0.01681715202331543, 0.01676595115661621, 0.01675468826293945, 0.01684377670288086, 0.01681920051574707, 0.01677004814147949, 0.01678950309753418, 0.016736255645751954, 0.016710655212402344, 0.016780288696289062, 0.016730112075805666, 0.016735231399536133, 0.016617504119873047, 0.016613344192504882, 0.01681715202331543, 0.016788480758666992, 0.01676288032531738, 0.016741439819335936, 0.01675667190551758, 0.01679769515991211, 0.039731231689453125, 0.016744415283203126, 0.016673791885375978, 0.016702463150024414, 0.016752639770507814, 0.016697343826293946, 0.016676864624023437, 0.01654374313354492, 0.01657344055175781, 0.01662259292602539, 0.016718847274780273, 0.01680179214477539, 0.01678950309753418, 0.01680998420715332, 0.016718847274780273, 0.016773120880126953, 0.01721855926513672, 0.016749568939208984, 0.016728063583374024, 0.01675468826293945, 0.016736255645751954, 0.01675775909423828, 0.01678950309753418, 0.016745471954345705, 0.016746496200561522, 0.016781312942504883, 0.016750591278076172, 0.01681510353088379, 0.01676288032531738, 0.01675468826293945, 0.016745471954345705, 0.01680384063720703, 0.016631807327270508, 0.01660518455505371, 0.016673791885375978, 0.01658780860900879, 0.01662870407104492, 0.01676288032531738, 0.016717824935913086, 0.01675468826293945, 0.016649215698242188, 0.01658678436279297, 0.01661231994628906, 0.01666975975036621, 0.016779199600219726, 0.01677824020385742, 0.01680486488342285, 0.016769023895263673, 0.016780288696289062, 0.016793600082397463, 0.016750591278076172, 0.016764928817749023, 0.01681612777709961, 0.016781312942504883, 0.01677107238769531, 0.01681305694580078, 0.01679155158996582, 0.016947200775146484, 0.01701785659790039, 0.01677516746520996, 0.01680588722229004, 0.01682022476196289, 0.016761856079101564, 0.03988582229614258, 0.01681612777709961, 0.016760831832885743, 0.01678643226623535, 0.016775232315063476, 0.01672902488708496, 0.01677107238769531, 0.016784383773803712, 0.01677516746520996, 0.016825344085693358, 0.016780288696289062, 0.01683456039428711, 0.01677107238769531, 0.01681203269958496, 0.01678950309753418, 0.016729087829589845, 0.01679667282104492, 0.01678950309753418, 0.01680076789855957, 0.016685056686401366, 0.016631872177124023, 0.016600000381469728, 0.01662566375732422, 0.01678643226623535, 0.016718847274780273, 0.01681715202331543, 0.01680998420715332, 0.016793600082397463, 0.01676595115661621, 0.01677516746520996, 0.016729087829589845, 0.016892927169799805, 0.016745471954345705, 0.016792575836181642, 0.01681407928466797, 0.01680179214477539, 0.01679155158996582, 0.016784383773803712, 0.01680281639099121, 0.01678335952758789, 0.016911359786987306, 0.016740352630615234, 0.016846847534179688, 0.016947200775146484, 0.01683046340942383, 0.01686016082763672, 0.016776191711425782, 0.01681920051574707, 0.01676697540283203, 0.01680588722229004, 0.01678950309753418, 0.01700044822692871, 0.01720524787902832, 0.016745471954345705, 0.01678335952758789, 0.01675574493408203, 0.016754655838012694, 0.0167587833404541, 0.017737728118896484, 0.017736703872680663, 0.017539072036743163, 0.017468416213989257, 0.017398784637451172]",tokens/s,56.71564870029314,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-1.4b,EleutherAI/pythia-1.4b,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1301.925888,3852.992512,0.0,3206.545408,3105.82784,s,10,0.22998556900024414,0.022998556900024412,0.0013529496380716877,0.022550992012023926,0.023032716751098633,0.02504474258422851,0.026654363250732425,"[0.0270567684173584, 0.02258086395263672, 0.022585599899291993, 0.022520320892333984, 0.02251375961303711, 0.022565088272094726, 0.022544511795043944, 0.022516672134399413, 0.022553632736206055, 0.022548351287841797]",tokens/s,11131.13318861012,kWh,2.679527560509432e-07,1.4682554384078958e-07,1.2156201460907105e-06,1.6303984459824433e-06,tokens/kWh,157016832.68333825,MB,1301.925888,3852.992512,0.0,3206.545408,3189.19168,s,10,10.880635986328127,1.0880635986328124,0.011995065648385062,1.0871802978515626,1.10194228515625,1.102078076171875,1.1021867089843749,"[1.101912109375, 1.097273681640625, 1.0840592041015624, 1.0657208251953125, 1.077572998046875, 1.0848209228515624, 1.0760526123046874, 1.0895396728515625, 1.1014700927734375, 1.1022138671875]",tokens/s,57.901027181831616,kWh,1.2661301302283622e-05,6.9378927495859155e-06,3.6167298649108625e-05,5.5766492700978175e-05,tokens/kWh,1129710.6371348856,,s,629,11.064073196411131,0.01758994148872994,0.0027467619663067533,0.017339391708374022,0.017653350830078126,0.01776842918395996,0.04001446121215821,"[0.01734758377075195, 0.016983039855957033, 0.017040384292602538, 0.016946176528930663, 0.01700864028930664, 0.01697689628601074, 0.017137664794921875, 0.017107967376708985, 0.0170199031829834, 0.016940031051635742, 0.01703321647644043, 0.017565696716308594, 0.01767731285095215, 0.017745920181274414, 0.017785856246948242, 0.017787904739379884, 0.01775923156738281, 0.01780633544921875, 0.017753087997436523, 0.017839103698730468, 0.017753087997436523, 0.01781862449645996, 0.017719295501708983, 0.017802240371704102, 0.01779814338684082, 0.01770086479187012, 0.01779199981689453, 0.017804288864135744, 0.01740083122253418, 0.01726464080810547, 0.017140735626220704, 0.01722368049621582, 0.017089536666870117, 0.01722777557373047, 0.017722368240356445, 0.01764659118652344, 0.017292287826538084, 0.017296384811401368, 0.017542144775390626, 0.01742131233215332, 0.01765888023376465, 0.017657855987548828, 0.017573888778686524, 0.01759231948852539, 0.017556480407714844, 0.017829887390136717, 0.01756876754760742, 0.01764556884765625, 0.017567743301391603, 0.017617919921875, 0.017584127426147463, 0.017305599212646485, 0.017211423873901368, 0.017415136337280274, 0.017277984619140624, 0.017285087585449218, 0.01738956832885742, 0.01721343994140625, 0.017141759872436522, 0.017300479888916014, 0.01741312026977539, 0.017560575485229494, 0.040629249572753906, 0.016951295852661134, 0.016931840896606445, 0.0169564151763916, 0.01700966453552246, 0.016904191970825197, 0.016966655731201173, 0.01719808006286621, 0.017755136489868165, 0.017351743698120117, 0.017246143341064453, 0.017370111465454103, 0.017494016647338868, 0.017551359176635743, 0.017479679107666016, 0.01742540740966797, 0.017736703872680663, 0.017497087478637697, 0.017476608276367187, 0.017526784896850587, 0.017527807235717775, 0.017348608016967772, 0.017228799819946287, 0.017177631378173828, 0.017170400619506837, 0.017712127685546874, 0.0175861759185791, 0.017527807235717775, 0.01759129524230957, 0.01745510482788086, 0.017510400772094727, 0.01760358428955078, 0.017511423110961915, 0.01723391914367676, 0.017693695068359376, 0.0176312313079834, 0.017687551498413084, 0.017481727600097655, 0.01754729652404785, 0.016971744537353516, 0.016990207672119142, 0.016898048400878905, 0.01716531181335449, 0.01754521560668945, 0.017575935363769533, 0.017108991622924806, 0.016907264709472656, 0.017003551483154297, 0.017286144256591796, 0.01768342399597168, 0.017510400772094727, 0.017702911376953127, 0.017503231048583985, 0.017596416473388672, 0.017704959869384765, 0.0176312313079834, 0.017501184463500977, 0.017496063232421876, 0.01757798385620117, 0.017552383422851564, 0.017844224929809572, 0.017604608535766602, 0.01782476806640625, 0.04033331298828125, 0.016849920272827147, 0.016887807846069337, 0.016888832092285155, 0.01756159973144531, 0.017501184463500977, 0.017509376525878906, 0.017475584030151366, 0.017505279541015627, 0.017500160217285156, 0.017597440719604493, 0.017531936645507812, 0.017553375244140624, 0.017735679626464843, 0.017588224411010742, 0.017558528900146485, 0.018251775741577148, 0.018109439849853515, 0.0174653434753418, 0.017566719055175782, 0.017534975051879884, 0.017497087478637697, 0.017559551239013673, 0.017546239852905272, 0.017534975051879884, 0.017485824584960938, 0.017508352279663086, 0.017580032348632812, 0.01763737678527832, 0.017507328033447265, 0.017245183944702147, 0.017282047271728516, 0.017468416213989257, 0.017512447357177736, 0.017318912506103516, 0.01722777557373047, 0.01724928092956543, 0.017262592315673828, 0.01721036720275879, 0.016780288696289062, 0.016821247100830078, 0.016726015090942382, 0.016722944259643553, 0.016733184814453125, 0.016720895767211915, 0.016743423461914063, 0.01679462432861328, 0.016703487396240235, 0.016779296875, 0.016794591903686523, 0.016756736755371093, 0.016735231399536133, 0.016862207412719727, 0.01678643226623535, 0.01677414321899414, 0.016854015350341797, 0.016749568939208984, 0.0168222713470459, 0.016781312942504883, 0.01686425590515137, 0.016748544692993163, 0.016825344085693358, 0.016719871520996094, 0.0400230712890625, 0.016929759979248046, 0.016959487915039064, 0.016919551849365236, 0.016895999908447267, 0.016882688522338866, 0.016937984466552734, 0.01695337677001953, 0.01689187240600586, 0.016870399475097657, 0.016906240463256835, 0.016910335540771485, 0.01697177505493164, 0.016897024154663084, 0.017565696716308594, 0.017464319229125978, 0.016917503356933594, 0.016893951416015626, 0.016949247360229493, 0.016942079544067384, 0.01696051216125488, 0.01685196876525879, 0.017519615173339845, 0.017084415435791016, 0.016949247360229493, 0.01702604866027832, 0.01701375961303711, 0.016910335540771485, 0.0169902400970459, 0.016977888107299804, 0.016949247360229493, 0.016959487915039064, 0.016738304138183592, 0.016716800689697265, 0.01683456039428711, 0.01674959945678711, 0.016807903289794923, 0.016686080932617187, 0.01680998420715332, 0.016776191711425782, 0.01679974365234375, 0.016736255645751954, 0.016743423461914063, 0.01677824020385742, 0.01681407928466797, 0.01680486488342285, 0.016846847534179688, 0.016852991104125976, 0.016776191711425782, 0.016785408020019533, 0.01678950309753418, 0.016743423461914063, 0.016825344085693358, 0.016755712509155272, 0.01677107238769531, 0.016865280151367186, 0.016942079544067384, 0.016921600341796874, 0.016932863235473633, 0.016906272888183593, 0.01694102478027344, 0.016930816650390625, 0.01701375961303711, 0.039992321014404295, 0.01698406410217285, 0.016941055297851563, 0.016906240463256835, 0.016713727951049806, 0.01679769515991211, 0.016895999908447267, 0.01677107238769531, 0.016721920013427736, 0.016963647842407226, 0.01746732711791992, 0.017526784896850587, 0.01757798385620117, 0.017496063232421876, 0.017506303787231444, 0.017488895416259767, 0.017339391708374022, 0.017117183685302736, 0.016744447708129884, 0.016772096633911132, 0.016942079544067384, 0.0169564151763916, 0.01696767997741699, 0.016910367965698243, 0.016962528228759766, 0.016986112594604492, 0.016977920532226562, 0.016973823547363282, 0.017067007064819336, 0.016944128036499022, 0.01698099136352539, 0.016995328903198242, 0.016982015609741212, 0.016987136840820313, 0.016948223114013672, 0.016932863235473633, 0.016923648834228516, 0.016953344345092772, 0.016882688522338866, 0.016934911727905275, 0.016950271606445313, 0.016946176528930663, 0.01696051216125488, 0.016977920532226562, 0.016910335540771485, 0.01696767997741699, 0.016970752716064453, 0.016926719665527345, 0.016940031051635742, 0.01698099136352539, 0.016943103790283204, 0.01698307228088379, 0.017061855316162108, 0.017507328033447265, 0.017943552017211914, 0.017699840545654297, 0.017369087219238282, 0.017625087738037108, 0.01764454460144043, 0.01743974494934082, 0.01745305633544922, 0.017491968154907226, 0.017476608276367187, 0.04009574508666992, 0.016906240463256835, 0.016885759353637696, 0.016867328643798828, 0.016936960220336913, 0.016877567291259766, 0.016959487915039064, 0.016878591537475587, 0.016737279891967775, 0.016671743392944336, 0.016769023895263673, 0.01681203269958496, 0.016894975662231446, 0.016876544952392578, 0.016898048400878905, 0.016854015350341797, 0.016932863235473633, 0.016877567291259766, 0.016897024154663084, 0.016871423721313478, 0.016899072647094726, 0.016874528884887694, 0.016895967483520506, 0.016890880584716796, 0.016871423721313478, 0.016939008712768554, 0.0169932804107666, 0.016903167724609376, 0.016907264709472656, 0.016899072647094726, 0.016941055297851563, 0.016886783599853517, 0.016924671173095703, 0.016920576095581053, 0.018545663833618165, 0.017936384201049805, 0.01773980712890625, 0.017669088363647462, 0.017748992919921876, 0.0175994873046875, 0.017624063491821287, 0.017459199905395507, 0.017542144775390626, 0.01759846305847168, 0.017498111724853514, 0.017541120529174805, 0.017535999298095704, 0.017558528900146485, 0.017549312591552735, 0.017503231048583985, 0.017539072036743163, 0.01754521560668945, 0.01759231948852539, 0.0175861759185791, 0.01753091239929199, 0.017523679733276367, 0.017309696197509765, 0.01727801513671875, 0.017613759994506838, 0.01759334373474121, 0.017460224151611328, 0.017449983596801756, 0.017510400772094727, 0.039964672088623046, 0.016955392837524414, 0.016893951416015626, 0.016919551849365236, 0.016905248641967772, 0.01705571174621582, 0.016939008712768554, 0.016942079544067384, 0.016917503356933594, 0.016932863235473633, 0.016925695419311524, 0.01694419288635254, 0.01691334342956543, 0.016987136840820313, 0.016929792404174804, 0.01695232009887695, 0.016982015609741212, 0.016966655731201173, 0.016915456771850586, 0.01679871940612793, 0.016709632873535156, 0.01678233528137207, 0.016882688522338866, 0.016756736755371093, 0.016726015090942382, 0.016914432525634765, 0.016891904830932617, 0.016902143478393555, 0.016953344345092772, 0.016924671173095703, 0.01697279930114746, 0.016963584899902344, 0.016948223114013672, 0.01700454330444336, 0.01702092742919922, 0.01724415969848633, 0.017531904220581054, 0.01749504089355469, 0.017443840026855468, 0.017456127166748048, 0.017590272903442384, 0.017494016647338868, 0.017531936645507812, 0.017558496475219728, 0.017511423110961915, 0.01757695960998535, 0.017551359176635743, 0.01746329689025879, 0.017498111724853514, 0.017526784896850587, 0.017526784896850587, 0.01748684883117676, 0.016913408279418944, 0.016886783599853517, 0.016893983840942383, 0.017012704849243165, 0.016934911727905275, 0.016918527603149415, 0.016966655731201173, 0.016871423721313478, 0.016865280151367186, 0.016917503356933594, 0.016866304397583007, 0.04018380737304687, 0.016926719665527345, 0.016930816650390625, 0.016957439422607423, 0.01702809524536133, 0.016906240463256835, 0.016919551849365236, 0.016888832092285155, 0.016951295852661134, 0.016922624588012695, 0.016943103790283204, 0.016910335540771485, 0.016917503356933594, 0.016970752716064453, 0.016930816650390625, 0.016866304397583007, 0.01701478385925293, 0.016944128036499022, 0.016970752716064453, 0.016935935974121095, 0.016919551849365236, 0.016923648834228516, 0.01764352035522461, 0.017551359176635743, 0.017484800338745117, 0.017563648223876953, 0.017559551239013673, 0.017624063491821287, 0.017508352279663086, 0.017567743301391603, 0.017460224151611328, 0.017490943908691405, 0.017492992401123047, 0.017529855728149413, 0.01760665512084961, 0.017583103179931642, 0.017500160217285156, 0.017497087478637697, 0.017550336837768556, 0.017558528900146485, 0.017504255294799806, 0.017507360458374022, 0.017541088104248048, 0.01756979179382324, 0.017523712158203125, 0.017509376525878906, 0.01758515167236328, 0.017555456161499023, 0.017552383422851564, 0.017510400772094727, 0.01743052864074707, 0.016948223114013672, 0.01683967971801758, 0.016759807586669923, 0.0167587833404541, 0.016793600082397463, 0.017063936233520507, 0.018144287109375, 0.017774560928344726, 0.017711103439331053, 0.0176363525390625, 0.01759129524230957, 0.01762816047668457, 0.04059033584594727, 0.017511423110961915, 0.017474559783935546, 0.017522687911987304, 0.017473567962646486, 0.017499103546142578, 0.017508352279663086, 0.017514495849609374, 0.017525760650634766, 0.017285120010375975, 0.01720729637145996, 0.017238016128540038, 0.01719193649291992, 0.017524736404418945, 0.017475584030151366, 0.017532928466796875, 0.017543167114257813, 0.017520639419555666, 0.017537023544311522, 0.017254432678222655, 0.017564640045166016, 0.01759846305847168, 0.017549312591552735, 0.017510400772094727, 0.017506368637084962, 0.01755743980407715, 0.017467391967773437, 0.0175861759185791, 0.017492992401123047, 0.01760051155090332, 0.017563648223876953, 0.017550336837768556, 0.017550336837768556, 0.017480703353881837, 0.017572864532470703, 0.017540096282958984, 0.017520639419555666, 0.017512447357177736, 0.017473535537719728, 0.017535999298095704, 0.017571840286254883, 0.01718681526184082, 0.016926719665527345, 0.016931840896606445, 0.016977920532226562, 0.016929792404174804, 0.01696460723876953, 0.016917503356933594, 0.016926719665527345, 0.01696870422363281, 0.017667072296142578, 0.01760051155090332, 0.018037792205810546, 0.018824159622192384, 0.017888256072998047, 0.017950719833374023, 0.017572864532470703, 0.01759334373474121, 0.017588224411010742, 0.01760358428955078, 0.017571840286254883, 0.0176312313079834, 0.017572864532470703, 0.04003123092651367, 0.01747865676879883, 0.01759334373474121, 0.017540096282958984, 0.017482751846313475, 0.01751865577697754, 0.01752876853942871, 0.017649663925170898, 0.01760358428955078, 0.017509376525878906, 0.01756159973144531, 0.017655807495117186, 0.01759846305847168, 0.01762303924560547, 0.017612800598144532, 0.017452032089233398, 0.01759334373474121, 0.017528831481933595, 0.017663999557495116, 0.017508352279663086, 0.017589248657226563, 0.017479679107666016, 0.017467391967773437, 0.017511423110961915, 0.01755340766906738, 0.017554431915283202, 0.017563648223876953, 0.017464319229125978, 0.017468416213989257, 0.017488895416259767, 0.017529855728149413, 0.01759129524230957, 0.017584127426147463, 0.017515520095825195, 0.017525760650634766, 0.01754422378540039, 0.017546207427978515, 0.017548288345336914, 0.01726361656188965, 0.017287168502807617, 0.017236991882324217, 0.017304576873779298, 0.017570816040039062, 0.017665023803710937, 0.017581056594848633, 0.017475584030151366, 0.01759129524230957, 0.017665023803710937, 0.017689599990844726, 0.01742540740966797, 0.01706598472595215, 0.016927743911743166, 0.017002527236938476, 0.017564640045166016, 0.01765273666381836, 0.017512447357177736, 0.01722777557373047, 0.017208351135253906, 0.017300447463989257, 0.017699840545654297, 0.01756159973144531, 0.017531904220581054, 0.017522687911987304]",tokens/s,56.85067233684151,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,x,x,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/x/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6693247b-2f8b9f1a72538b780ff0242f;36bc13a6-62e7-48d6-8fae-c9cd1a5171f7) Repository Not Found for url: https://huggingface.co/x/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: x is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,01-ai/Yi-34B,01-ai/Yi-34B,cuda,0,42,,,True,,,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",llama,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/xglm-4.5B,facebook/xglm-4.5B,cuda,0,42,,,True,,,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",xglm,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/xglm-7.5B,facebook/xglm-7.5B,cuda,0,42,,,True,,,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",xglm,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-14B,Qwen/Qwen1.5-14B,cuda,0,42,,,True,,,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",qwen2,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,meta-llama/Llama-2-70b-hf,meta-llama/Llama-2-70b-hf,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ self.load_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 250, in load_model_with_no_weights self.pretrained_model = self.automodel_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3710, in from_pretrained model = cls(config, *model_args, **model_kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 1101, in __init__ self.model = LlamaModel(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 884, in __init__ [LlamaDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 884, in [LlamaDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 679, in __init__ self.mlp = LlamaMLP(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 192, in __init__ self.gate_proj = nn.Linear(self.hidden_size, self.intermediate_size, bias=config.mlp_bias) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/linear.py"", line 98, in __init__ self.weight = Parameter(torch.empty((out_features, in_features), **factory_kwargs)) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_device.py"", line 78, in __torch_function__ return func(*args, **kwargs) torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 448.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Deci/DeciCoder-1b,Deci/DeciCoder-1b,cuda,0,42,,,True,,,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",llama,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,meta-llama/Meta-Llama-3-70B,meta-llama/Meta-Llama-3-70B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ self.load_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 250, in load_model_with_no_weights self.pretrained_model = self.automodel_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3710, in from_pretrained model = cls(config, *model_args, **model_kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 1101, in __init__ self.model = LlamaModel(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 884, in __init__ [LlamaDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 884, in [LlamaDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 679, in __init__ self.mlp = LlamaMLP(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 193, in __init__ self.up_proj = nn.Linear(self.hidden_size, self.intermediate_size, bias=config.mlp_bias) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/linear.py"", line 98, in __init__ self.weight = Parameter(torch.empty((out_features, in_features), **factory_kwargs)) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_device.py"", line 78, in __torch_function__ return func(*args, **kwargs) torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 448.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,r,r,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/r/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669473f2-4ba3570e1c0301a66d936d4d;81272ae6-f6a0-494f-8cbf-babca6f900ef) Repository Not Found for url: https://huggingface.co/r/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: r is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,togethercomputer/RedPajama-INCITE-Base-7B-v0.1,togethercomputer/RedPajama-INCITE-Base-7B-v0.1,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1282.695168,15192.293376,0.0,14545.846272,14287.956992,s,10,0.8550193710327147,0.08550193710327147,0.00024395574847383124,0.08541804885864257,0.08555268630981445,0.08589143257141113,0.08616242958068847,"[0.0862301788330078, 0.08539382171630859, 0.08544351959228516, 0.08539584350585938, 0.08543373107910156, 0.08540460968017578, 0.08540415954589843, 0.08541136169433594, 0.08542473602294921, 0.08547740936279297]",tokens/s,2994.0842122769277,kWh,1.0120247758691468e-06,5.545426958773393e-07,5.59528700234186e-06,7.161854474088345e-06,tokens/kWh,35744931.83660885,MB,1282.695168,15192.293376,0.0,14545.846272,14491.309056,s,10,21.402462646484377,2.1402462646484377,0.003115315439825715,2.1391610107421872,2.1417872802734372,2.1454981323242186,2.1484668139648435,"[2.149208984375, 2.140309814453125, 2.1391005859375, 2.139102294921875, 2.139119873046875, 2.140962646484375, 2.139553466796875, 2.137626220703125, 2.1392021484375, 2.138276611328125]",tokens/s,29.435864947227717,kWh,2.520578727550656e-05,1.3811139764388477e-05,0.00012119098109326038,0.0001602079081331554,tokens/kWh,393239.01506558654,,s,629,22.136534984588632,0.03519321937136506,0.010185011836285417,0.03395888137817383,0.03408015213012695,0.03426424865722656,0.11969617736816406,"[0.03463372802734375, 0.03450579071044922, 0.03455897521972656, 0.03454969787597656, 0.03457024002075195, 0.0345456657409668, 0.034566143035888675, 0.034561023712158204, 0.034553886413574215, 0.034541534423828124, 0.034446334838867186, 0.03438694381713867, 0.03439206314086914, 0.03438489532470703, 0.03419340896606445, 0.0339128303527832, 0.03389440155029297, 0.033888256072998044, 0.033857536315917966, 0.0339128303527832, 0.03383193588256836, 0.03392409515380859, 0.03393024063110352, 0.03393535995483398, 0.033903617858886716, 0.033896446228027344, 0.03391696166992188, 0.03393737411499023, 0.0339159049987793, 0.03394559860229492, 0.03393535995483398, 0.03395174407958984, 0.0339159049987793, 0.033942527770996093, 0.03398451232910156, 0.03398355102539063, 0.03399571228027344, 0.03401523208618164, 0.03398559951782226, 0.033966014862060544, 0.03400601577758789, 0.03396198272705078, 0.03401113510131836, 0.03395276641845703, 0.03401523208618164, 0.033982463836669925, 0.03402137756347656, 0.03400806427001953, 0.03404800033569336, 0.034032638549804685, 0.034010112762451174, 0.03402751922607422, 0.03403571319580078, 0.03404390335083008, 0.034034687042236327, 0.034050048828125, 0.03400400161743164, 0.034036705017089844, 0.03405311965942383, 0.034045951843261715, 0.034062335968017575, 0.03407257461547852, 0.11971177673339843, 0.03383599853515625, 0.033800193786621094, 0.03379302215576172, 0.03387187194824219, 0.03379609680175781, 0.03383193588256836, 0.03384627151489258, 0.03383500671386719, 0.03385961532592773, 0.03385955047607422, 0.033808383941650394, 0.03384832000732422, 0.03386880111694336, 0.033898494720458985, 0.03386470413208008, 0.03391897583007813, 0.03382681655883789, 0.03395481491088867, 0.034181121826171876, 0.03449651336669922, 0.034044929504394535, 0.033957889556884766, 0.03387289428710937, 0.03387801742553711, 0.0338974723815918, 0.03392921447753906, 0.033982463836669925, 0.03428966522216797, 0.03405311965942383, 0.03396710586547851, 0.03388620758056641, 0.033977344512939454, 0.03391795349121094, 0.0339947509765625, 0.03390771102905273, 0.03398758316040039, 0.033941505432128906, 0.034032638549804685, 0.03393843078613281, 0.03403059387207031, 0.03394153594970703, 0.033995742797851565, 0.03397017669677734, 0.03403059387207031, 0.03396915054321289, 0.034098175048828124, 0.03402751922607422, 0.03406335830688476, 0.03400089645385742, 0.03404288101196289, 0.03399782562255859, 0.03404390335083008, 0.03398758316040039, 0.03404185485839844, 0.03400908660888672, 0.03407360076904297, 0.0340398063659668, 0.034049057006835935, 0.03402134323120117, 0.03404185485839844, 0.03404390335083008, 0.034051071166992186, 0.11977219390869141, 0.03384009552001953, 0.033854465484619144, 0.03384627151489258, 0.03381862258911133, 0.03383603286743164, 0.03381248092651367, 0.03383193588256836, 0.033865726470947266, 0.03405311965942383, 0.03423436737060547, 0.03390566253662109, 0.03386675262451172, 0.03386163330078125, 0.033844223022460936, 0.03385036849975586, 0.03388415908813477, 0.03386265563964844, 0.03383910369873047, 0.033911808013916016, 0.033865726470947266, 0.03389952087402344, 0.03384524917602539, 0.03390566253662109, 0.033876991271972655, 0.03392921447753906, 0.03389440155029297, 0.03395283126831055, 0.033874881744384765, 0.03395590209960937, 0.0339219856262207, 0.033944576263427735, 0.033942527770996093, 0.03400294494628906, 0.033947647094726564, 0.033977344512939454, 0.0339128303527832, 0.033963008880615236, 0.033926143646240234, 0.03396505737304688, 0.033963008880615236, 0.03402751922607422, 0.03396716690063477, 0.03401824188232422, 0.033979393005371096, 0.03401932907104492, 0.033980415344238284, 0.03401830291748047, 0.03401113510131836, 0.03401318359375, 0.033990657806396485, 0.034050048828125, 0.034038848876953125, 0.034017215728759764, 0.03401728057861328, 0.034051071166992186, 0.03400703811645508, 0.034062335968017575, 0.034051071166992186, 0.03406438446044922, 0.03404288101196289, 0.034049022674560545, 0.034045951843261715, 0.11964313507080078, 0.0337786865234375, 0.03378995132446289, 0.03378278350830078, 0.033797119140625, 0.03380940628051758, 0.03383708953857422, 0.03377967834472656, 0.03383814239501953, 0.03379808044433594, 0.03383295822143555, 0.033788928985595705, 0.03388518524169922, 0.03381760025024414, 0.03384012985229492, 0.033873920440673826, 0.033838081359863284, 0.03383603286743164, 0.03386265563964844, 0.03386368179321289, 0.03385651016235352, 0.033898494720458985, 0.03386470413208008, 0.03385651016235352, 0.03387596893310547, 0.033893375396728515, 0.033888256072998044, 0.03393535995483398, 0.03390771102905273, 0.033931262969970705, 0.03392716979980469, 0.033911808013916016, 0.03395993423461914, 0.03405619049072266, 0.03424051284790039, 0.033972225189208984, 0.03396198272705078, 0.033979393005371096, 0.033960960388183595, 0.03394259262084961, 0.03399983978271484, 0.03395888137817383, 0.034012161254882815, 0.03394355010986328, 0.03401113510131836, 0.03398761749267578, 0.034055137634277345, 0.03401830291748047, 0.03406643295288086, 0.034028545379638675, 0.03401932907104492, 0.03402137756347656, 0.03402239990234375, 0.034012161254882815, 0.03404083251953125, 0.03404083251953125, 0.03415859222412109, 0.0343900146484375, 0.034200576782226565, 0.03408793640136719, 0.03405926513671875, 0.03412486267089844, 0.03415750503540039, 0.11974655914306641, 0.03382271957397461, 0.033772544860839845, 0.03379814529418945, 0.03376947021484375, 0.03382476806640625, 0.03378585433959961, 0.03380940628051758, 0.03381350326538086, 0.03381043243408203, 0.03387289428710937, 0.03382886505126953, 0.03397836685180664, 0.03384729766845703, 0.033870849609375, 0.03383295822143555, 0.033842174530029294, 0.0338524169921875, 0.033882110595703126, 0.0339046401977539, 0.0339128303527832, 0.03392409515380859, 0.033833984375, 0.033923072814941405, 0.03387289428710937, 0.03393740844726562, 0.03387596893310547, 0.03395174407958984, 0.03393535995483398, 0.0339947509765625, 0.03395276641845703, 0.03398451232910156, 0.03391692733764649, 0.03399270248413086, 0.033957889556884766, 0.03398963165283203, 0.03395379257202148, 0.033983486175537106, 0.033966079711914066, 0.034045951843261715, 0.03395481491088867, 0.03402547073364258, 0.03392409515380859, 0.0340316162109375, 0.033974273681640625, 0.0340684814453125, 0.03396915054321289, 0.03406028747558594, 0.03402444839477539, 0.03404083251953125, 0.034034687042236327, 0.03416166305541992, 0.03400089645385742, 0.03406028747558594, 0.03402444839477539, 0.03412275314331055, 0.0340582389831543, 0.03417702484130859, 0.03404083251953125, 0.034065406799316404, 0.034045951843261715, 0.03407462310791016, 0.03405926513671875, 0.11972096252441407, 0.03380121612548828, 0.03379814529418945, 0.03381452941894531, 0.03378278350830078, 0.03384524917602539, 0.033833984375, 0.03381862258911133, 0.03385036849975586, 0.03383295822143555, 0.03383603286743164, 0.03383603286743164, 0.033896446228027344, 0.03383295822143555, 0.03385651016235352, 0.03382275390625, 0.03389539337158203, 0.03386368179321289, 0.03394976043701172, 0.03384313583374023, 0.033931262969970705, 0.0339128303527832, 0.03391897583007813, 0.03386982345581055, 0.03393228912353516, 0.03386985778808594, 0.033906654357910154, 0.033925121307373046, 0.033944576263427735, 0.033957889556884766, 0.033941505432128906, 0.033960960388183595, 0.03396201705932617, 0.03397219085693359, 0.033960960388183595, 0.03394867324829102, 0.033972225189208984, 0.0339947509765625, 0.033974273681640625, 0.033960960388183595, 0.03397119903564453, 0.033984542846679684, 0.03398550415039062, 0.0339947509765625, 0.033993728637695314, 0.033990657806396485, 0.034336769104003906, 0.034293758392333985, 0.03417190551757812, 0.03405311965942383, 0.034098175048828124, 0.03425075149536133, 0.03430198287963867, 0.03435830307006836, 0.03427324676513672, 0.03405926513671875, 0.03404899215698242, 0.03400908660888672, 0.03405926513671875, 0.03408281707763672, 0.03423027038574219, 0.03424563217163086, 0.034081790924072264, 0.11971481323242188, 0.03381350326538086, 0.033787967681884766, 0.03381958389282227, 0.03381043243408203, 0.03381964874267578, 0.03382377624511719, 0.033834014892578125, 0.03381856155395508, 0.03383910369873047, 0.03387494277954101, 0.033854465484619144, 0.033858558654785154, 0.03385958480834961, 0.033873920440673826, 0.03390771102905273, 0.03384729766845703, 0.0339128303527832, 0.033879039764404296, 0.03390982437133789, 0.033877952575683594, 0.03392204666137695, 0.033849342346191406, 0.03393638229370117, 0.03388313674926758, 0.033926143646240234, 0.03388108825683594, 0.03392204666137695, 0.0339128303527832, 0.03396915054321289, 0.03395891189575195, 0.03410227203369141, 0.03394047927856445, 0.03398860931396484, 0.03393231964111328, 0.03401520156860351, 0.03398758316040039, 0.034083839416503905, 0.03394867324829102, 0.03399782562255859, 0.033949695587158206, 0.034116607666015625, 0.0341739501953125, 0.03415961456298828, 0.03400191879272461, 0.034032638549804685, 0.03398451232910156, 0.03401728057861328, 0.03399679946899414, 0.03402137756347656, 0.03400908660888672, 0.03403878402709961, 0.03398963165283203, 0.03405516815185547, 0.03400601577758789, 0.03409305572509766, 0.03402547073364258, 0.03407462310791016, 0.034044929504394535, 0.03409305572509766, 0.03405721664428711, 0.03410739135742188, 0.03404288101196289, 0.11968511962890625, 0.03376639938354492, 0.03375820922851563, 0.03376435089111328, 0.03377151870727539, 0.033788928985595705, 0.03381964874267578, 0.03378688049316406, 0.033857536315917966, 0.03380223846435547, 0.03402649688720703, 0.03380633544921875, 0.033888256072998044, 0.03380428695678711, 0.033841152191162106, 0.033841152191162106, 0.033880062103271484, 0.033810497283935544, 0.03386470413208008, 0.03385951995849609, 0.033838081359863284, 0.03386163330078125, 0.03387705612182617, 0.033886142730712894, 0.03389132690429687, 0.033931262969970705, 0.033904670715332035, 0.03396502304077149, 0.033895423889160156, 0.033941505432128906, 0.033933311462402346, 0.03391795349121094, 0.03392921447753906, 0.033928192138671875, 0.033949695587158206, 0.033939456939697264, 0.03393535995483398, 0.03395686340332031, 0.03394662475585938, 0.03396710586547851, 0.03395174407958984, 0.03399270248413086, 0.03398860931396484, 0.033976318359375, 0.03395686340332031, 0.03398553466796875, 0.033993728637695314, 0.03399782562255859, 0.03401728057861328, 0.03401318359375, 0.03401932907104492, 0.03402035140991211, 0.03400089645385742, 0.0340398063659668, 0.0340049934387207, 0.03404800033569336, 0.03401830291748047, 0.034051071166992186, 0.03402342224121094, 0.03403059387207031, 0.034030654907226565, 0.034065345764160156, 0.03404083251953125, 0.11970047760009765, 0.03380121612548828, 0.03379814529418945, 0.033811454772949216, 0.0338166389465332, 0.03381856155395508, 0.03383295822143555, 0.03383910369873047, 0.033840190887451174, 0.033833919525146486, 0.03382681655883789, 0.033825790405273434, 0.033844223022460936, 0.033880062103271484, 0.033870849609375, 0.03387494277954101, 0.033849342346191406, 0.03390771102905273, 0.03381862258911133, 0.033882110595703126, 0.03386265563964844, 0.03390054321289063, 0.03388415908813477, 0.03399884796142578, 0.03418828964233398, 0.034105342864990236, 0.0338974723815918, 0.03394662475585938, 0.033911808013916016, 0.034010112762451174, 0.033944576263427735, 0.034032638549804685, 0.03392409515380859, 0.03396710586547851, 0.03394559860229492, 0.033977344512939454, 0.03394047927856445, 0.033966079711914066, 0.03395891189575195, 0.033972225189208984, 0.033947647094726564, 0.033983486175537106, 0.03394867324829102, 0.03398963165283203, 0.0339865608215332, 0.03403366470336914, 0.0339681282043457, 0.03401830291748047, 0.03401523208618164, 0.03405209732055664, 0.03398758316040039, 0.03401420974731445, 0.034032638549804685, 0.034044929504394535, 0.03404800033569336, 0.03405516815185547, 0.03401932907104492, 0.03407564926147461, 0.03405414581298828, 0.03407462310791016, 0.03404800033569336, 0.0341473274230957, 0.034070526123046875, 0.11979366302490234, 0.03381862258911133, 0.03383091354370117, 0.03378073501586914, 0.03376844787597656, 0.03377459335327149, 0.033838081359863284, 0.03380121612548828, 0.03382988739013672, 0.03380326461791992, 0.03387494277954101, 0.033833984375, 0.03388620758056641, 0.03382988739013672, 0.03389440155029297, 0.0339159049987793, 0.03387494277954101, 0.03382988739013672, 0.03384320068359375, 0.03385347366333008, 0.03385347366333008, 0.033886142730712894, 0.0339159049987793, 0.03388927841186523, 0.033882110595703126, 0.03395686340332031, 0.03390566253662109, 0.03390566253662109, 0.03389235305786133, 0.033893375396728515, 0.033898494720458985, 0.033906688690185545, 0.03391078567504883, 0.03395379257202148, 0.033931262969970705, 0.03395891189575195, 0.03398863983154297, 0.03401110458374024, 0.033960960388183595, 0.03400601577758789, 0.033976318359375, 0.0340049934387207, 0.03396198272705078, 0.03399270248413086, 0.033963008880615236, 0.033977344512939454, 0.03399168014526367, 0.03402239990234375, 0.03397836685180664, 0.03398553466796875, 0.033979393005371096, 0.03400089645385742, 0.034016254425048825, 0.03406438446044922, 0.03407974243164062, 0.03404083251953125, 0.034012161254882815, 0.03405414581298828, 0.03403878402709961, 0.03402956771850586, 0.03407155227661133, 0.03430297470092773, 0.0340766716003418]",tokens/s,28.414564449129347,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-72B,Qwen/Qwen1.5-72B,cuda,0,42,,,True,,,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",qwen2,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,internlm/internlm-20b,internlm/internlm-20b,cuda,0,42,,,True,,,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",internlm,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-110B,Qwen/Qwen1.5-110B,cuda,0,42,,,True,,,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",qwen2,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Salesforce/codegen-16B-nl,Salesforce/codegen-16B-nl,cuda,0,42,,,True,,,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",codegen,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,m,m,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/m/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669472f0-36e202fb7a9b46df3c949191;8f2a5909-b11d-4c1d-9ad4-de6828372d32) Repository Not Found for url: https://huggingface.co/m/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: m is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,openai-community/gpt2-large,openai-community/gpt2-large,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1252.02432,2663.907328,0.0,2017.460224,1692.280832,s,10,0.2064598731994629,0.02064598731994629,0.0005540037501787667,0.02062006378173828,0.021011696624755858,0.02152368869781494,0.021933282356262206,"[0.022035680770874023, 0.020726720809936525, 0.020512096405029295, 0.020646879196166992, 0.02059401512145996, 0.0206461124420166, 0.02054812812805176, 0.01995315170288086, 0.019899168014526368, 0.02089792060852051]",tokens/s,12399.503885807191,kWh,2.3735125400774293e-07,1.30057457986313e-07,8.802355323895943e-07,1.2476442443836503e-06,tokens/kWh,205186695.768766,MB,1252.352,2663.907328,0.0,2017.460224,1740.079616,s,10,12.773597900390625,1.2773597900390625,0.022071669222506918,1.284090393066406,1.2936079833984375,1.3011394409179686,1.3071646069335936,"[1.3086708984375, 1.28357763671875, 1.2846031494140624, 1.291934326171875, 1.2860447998046876, 1.2825037841796876, 1.2351209716796876, 1.2369967041015626, 1.2783282470703126, 1.2858173828125]",tokens/s,49.320481583402135,kWh,1.459130997917898e-05,7.9937436635783e-06,3.176198502580928e-05,5.434703866856655e-05,tokens/kWh,1159216.7953106556,,s,629,12.936046607971182,0.020566051840971687,0.0025236247108550132,0.020340736389160157,0.020603699493408203,0.02112983055114746,0.04058865051269533,"[0.02168934440612793, 0.021246976852416992, 0.021126144409179686, 0.021160959243774414, 0.021172224044799806, 0.02147737693786621, 0.021358591079711914, 0.022424575805664062, 0.021336063385009766, 0.021005311965942384, 0.021204992294311522, 0.02127052879333496, 0.02122137641906738, 0.021399551391601563, 0.02148044776916504, 0.021146623611450196, 0.02107187271118164, 0.02125004768371582, 0.02130636787414551, 0.021184511184692383, 0.020970495223999023, 0.02106265640258789, 0.02083839988708496, 0.021404672622680664, 0.02129817581176758, 0.020954111099243163, 0.020926464080810548, 0.020780031204223632, 0.020561983108520508, 0.020260799407958986, 0.020297727584838866, 0.02040012741088867, 0.020324352264404297, 0.020388864517211915, 0.020208639144897463, 0.020373504638671876, 0.02053222465515137, 0.02026393508911133, 0.020298751831054687, 0.020607999801635742, 0.020366336822509764, 0.020324352264404297, 0.020281343460083007, 0.02022502326965332, 0.020296703338623046, 0.02029465675354004, 0.020323328018188477, 0.02039091110229492, 0.020338720321655273, 0.020221920013427735, 0.020338687896728515, 0.020288511276245116, 0.02023321533203125, 0.020303871154785155, 0.020229120254516602, 0.02042572784423828, 0.02045952033996582, 0.020323328018188477, 0.020298751831054687, 0.020379648208618165, 0.020298751831054687, 0.020478975296020507, 0.041780223846435545, 0.02040934371948242, 0.020289535522460937, 0.020179967880249023, 0.020343807220458983, 0.02025881576538086, 0.020166656494140626, 0.020307968139648438, 0.02031820869445801, 0.020303871154785155, 0.020398080825805662, 0.020307968139648438, 0.020332544326782227, 0.020382720947265624, 0.0204769287109375, 0.02062745666503906, 0.02046668815612793, 0.020384767532348632, 0.020319232940673827, 0.020338687896728515, 0.02024345588684082, 0.020328447341918944, 0.020315135955810547, 0.02028339195251465, 0.02052403259277344, 0.020511743545532226, 0.020402175903320312, 0.020389888763427736, 0.02028646469116211, 0.02045132827758789, 0.020398080825805662, 0.02032537651062012, 0.020271104812622072, 0.020511743545532226, 0.020347904205322266, 0.020311040878295897, 0.02025164794921875, 0.020296703338623046, 0.020369407653808593, 0.020333568572998048, 0.020239360809326173, 0.020414464950561522, 0.020379648208618165, 0.020512767791748047, 0.020388864517211915, 0.020369407653808593, 0.02030182456970215, 0.020351999282836913, 0.020620288848876952, 0.02066022491455078, 0.02048307228088379, 0.020393983840942383, 0.020264959335327147, 0.020287488937377928, 0.020386816024780274, 0.02020147132873535, 0.020332544326782227, 0.020338687896728515, 0.020410367965698242, 0.02042367935180664, 0.020346879959106445, 0.02026188850402832, 0.020413440704345705, 0.04141567993164062, 0.020381696701049806, 0.020397056579589845, 0.02040115165710449, 0.021132287979125978, 0.020323328018188477, 0.02043187141418457, 0.020329471588134765, 0.020221952438354493, 0.020557823181152343, 0.02062950325012207, 0.020319232940673827, 0.020297727584838866, 0.020299776077270508, 0.020313087463378905, 0.020508672714233397, 0.02044313621520996, 0.02020966339111328, 0.0204083194732666, 0.020361215591430663, 0.020271104812622072, 0.020411392211914063, 0.020388864517211915, 0.02050764846801758, 0.020344831466674804, 0.02026905632019043, 0.020405248641967775, 0.020338687896728515, 0.02017791938781738, 0.020221952438354493, 0.02027008056640625, 0.020238336563110353, 0.020356096267700196, 0.020331552505493164, 0.020701152801513672, 0.020482112884521484, 0.02038265609741211, 0.020304895401000975, 0.02028441619873047, 0.02025164794921875, 0.020328447341918944, 0.02038172721862793, 0.020375520706176757, 0.020281343460083007, 0.020413440704345705, 0.020289535522460937, 0.02043903923034668, 0.02047385597229004, 0.02022502326965332, 0.020388864517211915, 0.020554752349853517, 0.020303871154785155, 0.020322303771972656, 0.020452352523803712, 0.020305919647216796, 0.020396032333374024, 0.020446207046508787, 0.02038374328613281, 0.02066022491455078, 0.020385791778564453, 0.020455423355102538, 0.020265983581542968, 0.0203223991394043, 0.04163779067993164, 0.020274175643920898, 0.020381696701049806, 0.02045132827758789, 0.020298751831054687, 0.02045849609375, 0.020426752090454102, 0.02025267219543457, 0.02038380813598633, 0.02033555221557617, 0.020413440704345705, 0.02059878349304199, 0.02043391990661621, 0.020420608520507814, 0.020462591171264647, 0.02045747184753418, 0.020460544586181642, 0.02040115165710449, 0.02064384078979492, 0.02069196891784668, 0.02043391990661621, 0.020617216110229493, 0.020578336715698243, 0.020899808883666993, 0.020501504898071288, 0.020402175903320312, 0.020324352264404297, 0.02046771240234375, 0.020454399108886717, 0.020387840270996094, 0.020402175903320312, 0.020290559768676757, 0.0203305606842041, 0.02036832046508789, 0.020314144134521483, 0.020506591796875, 0.020562944412231447, 0.020340736389160157, 0.020965375900268556, 0.02066739273071289, 0.020363264083862305, 0.02041753578186035, 0.02050048065185547, 0.02044108772277832, 0.020349952697753908, 0.020338687896728515, 0.0204083194732666, 0.020374528884887694, 0.02042572784423828, 0.020265983581542968, 0.020316192626953125, 0.020360160827636718, 0.020351999282836913, 0.02041651153564453, 0.02043801689147949, 0.021553152084350585, 0.022665216445922853, 0.020616191864013672, 0.020446207046508787, 0.020402175903320312, 0.020602880477905275, 0.02031718444824219, 0.020530176162719727, 0.041611263275146484, 0.020493312835693358, 0.020378623962402344, 0.020380672454833985, 0.020707328796386718, 0.020382720947265624, 0.02022809600830078, 0.02028441619873047, 0.020237312316894532, 0.020525056838989256, 0.020555776596069338, 0.020412416458129884, 0.02044211196899414, 0.020295679092407228, 0.020320255279541014, 0.020374528884887694, 0.02032537651062012, 0.02042163276672363, 0.020332544326782227, 0.020363264083862305, 0.02033875274658203, 0.02030995178222656, 0.020332544326782227, 0.02042265510559082, 0.020406272888183592, 0.020377599716186523, 0.020350976943969725, 0.020356096267700196, 0.020376575469970702, 0.020485120773315428, 0.020327423095703127, 0.02049126434326172, 0.020502527236938475, 0.020365312576293947, 0.02035001564025879, 0.020302783966064452, 0.020322303771972656, 0.02038374328613281, 0.02043903923034668, 0.020304895401000975, 0.020346879959106445, 0.020335615158081053, 0.02026700782775879, 0.020382720947265624, 0.020528127670288086, 0.02050662422180176, 0.020411392211914063, 0.020315135955810547, 0.02025267219543457, 0.020414464950561522, 0.020333568572998048, 0.020414527893066407, 0.020468671798706053, 0.021292032241821288, 0.02066227149963379, 0.02040115165710449, 0.020405248641967775, 0.020246528625488282, 0.0204400634765625, 0.020307968139648438, 0.020489215850830078, 0.020373504638671876, 0.02045747184753418, 0.041545726776123046, 0.020338687896728515, 0.020365312576293947, 0.020374528884887694, 0.020313087463378905, 0.020362239837646484, 0.020290559768676757, 0.020337663650512695, 0.02046976089477539, 0.020412416458129884, 0.020362239837646484, 0.02033875274658203, 0.020492223739624022, 0.020376575469970702, 0.020379648208618165, 0.02039910316467285, 0.020376575469970702, 0.0202608642578125, 0.020331520080566406, 0.02030188751220703, 0.02038060760498047, 0.020322303771972656, 0.020256767272949217, 0.020313087463378905, 0.020295679092407228, 0.020463615417480468, 0.020447231292724608, 0.020333568572998048, 0.02038374328613281, 0.02037555122375488, 0.020282367706298828, 0.02029465675354004, 0.02029363250732422, 0.020371456146240235, 0.02037555122375488, 0.02025574493408203, 0.020323328018188477, 0.02038582420349121, 0.020661216735839844, 0.020368383407592772, 0.020356096267700196, 0.02031001663208008, 0.02025062370300293, 0.02027628707885742, 0.020461503982543944, 0.020322303771972656, 0.020297727584838866, 0.020333568572998048, 0.020340736389160157, 0.020444160461425782, 0.020371456146240235, 0.02028441619873047, 0.02030080032348633, 0.02031718444824219, 0.02027622413635254, 0.020287488937377928, 0.02027622413635254, 0.020289535522460937, 0.020305919647216796, 0.0204083194732666, 0.02042367935180664, 0.020330495834350586, 0.02029779243469238, 0.04078073501586914, 0.019556352615356445, 0.01965260887145996, 0.019647487640380858, 0.01965977668762207, 0.019512319564819337, 0.019544063568115236, 0.019756032943725587, 0.019567615509033204, 0.01960550308227539, 0.01977446365356445, 0.01967001533508301, 0.019680255889892577, 0.019725311279296876, 0.019695615768432616, 0.019696640014648437, 0.019620864868164063, 0.019562496185302734, 0.019513343811035155, 0.019569664001464843, 0.01965363121032715, 0.019636224746704102, 0.01967616081237793, 0.0195830078125, 0.01972425651550293, 0.019689472198486328, 0.019571712493896484, 0.01958502388000488, 0.019668991088867188, 0.019619840621948242, 0.019551231384277345, 0.01946112060546875, 0.019552255630493166, 0.019523584365844726, 0.019540992736816407, 0.019512319564819337, 0.019579904556274414, 0.019586048126220702, 0.019783679962158202, 0.019563583374023436, 0.0195665283203125, 0.01962393569946289, 0.019478527069091797, 0.019588096618652344, 0.019580928802490235, 0.019550207138061524, 0.019580928802490235, 0.01945907211303711, 0.019506175994873046, 0.019509248733520508, 0.019500032424926757, 0.019486719131469727, 0.019507200241088866, 0.01965772819519043, 0.019620864868164063, 0.019546112060546874, 0.019602432250976562, 0.019567615509033204, 0.019583999633789064, 0.01963007926940918, 0.019560447692871095, 0.019567615509033204, 0.019582975387573243, 0.040072223663330075, 0.019545055389404296, 0.019558399200439454, 0.019567615509033204, 0.019528703689575197, 0.019572736740112305, 0.019581951141357423, 0.019581951141357423, 0.019505151748657225, 0.019529727935791014, 0.019579904556274414, 0.019612672805786133, 0.01960960006713867, 0.01960550308227539, 0.019793920516967774, 0.019748863220214845, 0.01963929557800293, 0.01962495994567871, 0.019537919998168944, 0.019478527069091797, 0.019497983932495116, 0.019708927154541016, 0.019573759078979493, 0.01967001533508301, 0.01960140800476074, 0.019571712493896484, 0.019552255630493166, 0.01945702362060547, 0.019553279876708983, 0.01947750473022461, 0.019596319198608398, 0.019571680068969727, 0.01963929557800293, 0.01963315200805664, 0.019589120864868165, 0.019787776947021486, 0.020100095748901366, 0.019734527587890623, 0.01964646339416504, 0.019530752182006835, 0.019555328369140625, 0.01960857582092285, 0.019517440795898438, 0.019582975387573243, 0.01949286460876465, 0.01960038375854492, 0.019567615509033204, 0.019647487640380858, 0.019472415924072266, 0.019618783950805664, 0.019561471939086913, 0.01961676788330078, 0.019664896011352538, 0.019955711364746095, 0.020336639404296874, 0.019862560272216796, 0.019612640380859376, 0.01962598419189453, 0.01969049644470215, 0.019742719650268553, 0.019590143203735352, 0.019730432510375977, 0.019540992736816407, 0.04009471893310547, 0.019528703689575197, 0.01958502388000488, 0.019563520431518554, 0.019508224487304687, 0.019553279876708983, 0.019518463134765626, 0.019549184799194336, 0.02029465675354004, 0.02040729522705078, 0.020312063217163084, 0.02008883285522461, 0.020380672454833985, 0.020505599975585938, 0.020420608520507814, 0.02040729522705078, 0.020444160461425782, 0.02060697555541992, 0.020707328796386718, 0.020419584274291993, 0.020611072540283205, 0.020502527236938475, 0.02039193534851074, 0.020341760635375978, 0.02040012741088867, 0.020411392211914063, 0.020374528884887694, 0.0204769287109375, 0.02045132827758789, 0.02025881576538086, 0.02022809600830078, 0.020373504638671876, 0.020363264083862305, 0.020579328536987306, 0.020339712142944336, 0.02041753578186035, 0.020376575469970702, 0.02029363250732422, 0.020287488937377928, 0.020350976943969725, 0.020298751831054687, 0.020388864517211915, 0.020364288330078126, 0.02022604751586914, 0.02045030403137207, 0.020340736389160157, 0.02042163276672363, 0.02028544044494629, 0.02042163276672363, 0.020340736389160157, 0.020562944412231447, 0.020502527236938475, 0.020406272888183592, 0.020321279525756835, 0.020403200149536133, 0.020254720687866212, 0.020398080825805662, 0.02045030403137207, 0.020369407653808593, 0.020297727584838866, 0.020298751831054687, 0.020311040878295897, 0.02043289566040039, 0.04199628829956055, 0.020382720947265624, 0.02044825553894043, 0.02059878349304199, 0.02042163276672363, 0.020548608779907225, 0.02042572784423828, 0.020378623962402344, 0.020327423095703127, 0.020380672454833985, 0.020517887115478514, 0.020526079177856444, 0.020413440704345705, 0.020254720687866212, 0.02042470359802246, 0.020460544586181642, 0.02044211196899414, 0.020303871154785155, 0.02029363250732422, 0.020406335830688477, 0.020438976287841797, 0.020060192108154296, 0.02008982467651367, 0.020332544326782227, 0.02028339195251465, 0.020221952438354493, 0.020411392211914063, 0.02045747184753418, 0.020428800582885744, 0.020344831466674804, 0.02051481628417969, 0.020594688415527345, 0.02048102378845215, 0.02039091110229492, 0.020360191345214843, 0.020360191345214843, 0.020392959594726562, 0.02028544044494629, 0.02045952033996582, 0.020378623962402344, 0.02024550437927246, 0.020287488937377928, 0.020329471588134765, 0.020299776077270508, 0.020504575729370117, 0.020379648208618165, 0.02066431999206543, 0.02047590446472168, 0.020533248901367186, 0.020512767791748047, 0.020575231552124023, 0.020134912490844727, 0.02065920066833496, 0.02046668815612793, 0.02043187141418457, 0.020404224395751954, 0.02027622413635254, 0.020315135955810547, 0.020385791778564453, 0.020333568572998048, 0.020407327651977537, 0.02034787178039551, 0.020523008346557618]",tokens/s,48.62381986258539,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,huggyllama/llama-65b,huggyllama/llama-65b,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ self.load_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 250, in load_model_with_no_weights self.pretrained_model = self.automodel_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3710, in from_pretrained model = cls(config, *model_args, **model_kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 1101, in __init__ self.model = LlamaModel(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 884, in __init__ [LlamaDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 884, in [LlamaDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 677, in __init__ self.self_attn = LLAMA_ATTENTION_CLASSES[config._attn_implementation](config=config, layer_idx=layer_idx) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 387, in __init__ super().__init__(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 264, in __init__ self.v_proj = nn.Linear(self.hidden_size, self.num_key_value_heads * self.head_dim, bias=config.attention_bias) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/linear.py"", line 98, in __init__ self.weight = Parameter(torch.empty((out_features, in_features), **factory_kwargs)) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_device.py"", line 78, in __torch_function__ return func(*args, **kwargs) torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 128.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,l,l,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/l/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66947463-5b622f330cd55e823d157cd1;8333c871-1397-4b11-8f2f-3dd0d8e1335f) Repository Not Found for url: https://huggingface.co/l/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: l is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,B,B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/B/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66932515-204de0c20a7535c924327efb;57e374e4-5da5-4903-8d63-c989df7b529a) Repository Not Found for url: https://huggingface.co/B/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: B is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,huggyllama/llama-30b,huggyllama/llama-30b,cuda,0,42,,,True,,,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",llama,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,meta-llama/Llama-2-13b-hf,meta-llama/Llama-2-13b-hf,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ self.load_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 250, in load_model_with_no_weights self.pretrained_model = self.automodel_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3710, in from_pretrained model = cls(config, *model_args, **model_kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 1101, in __init__ self.model = LlamaModel(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 884, in __init__ [LlamaDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 884, in [LlamaDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 679, in __init__ self.mlp = LlamaMLP(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 192, in __init__ self.gate_proj = nn.Linear(self.hidden_size, self.intermediate_size, bias=config.mlp_bias) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/linear.py"", line 98, in __init__ self.weight = Parameter(torch.empty((out_features, in_features), **factory_kwargs)) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_device.py"", line 78, in __torch_function__ return func(*args, **kwargs) torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 136.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-7B,,cuda,0,42,,,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,217063f5c507ed7cc255df7e1f64c4333a0b4dfe,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,,MB,1296.539648,17102.798848,0.0,16456.351744,16359.853056,s,10,1.0926036529541017,0.10926036529541014,0.001893364755570167,0.1085610237121582,0.10995124130249023,0.1124175724029541,0.1143906372833252,"[0.11488390350341797, 0.10940316772460937, 0.1085745620727539, 0.10866429138183593, 0.10854630279541015, 0.1085474853515625, 0.10843228912353516, 0.10861603546142579, 0.10850553894042969, 0.1084300765991211]",tokens/s,2343.0271288938675,kWh,1.2849429361342232e-06,7.04090113883823e-07,6.003153957108733e-06,7.99218700712678e-06,tokens/kWh,32031282.5227588,MB,1296.539648,17102.798848,0.0,16456.351744,16416.242176,s,10,29.505367675781255,2.950536767578125,0.007298161312162202,2.95167822265625,2.96021103515625,2.9605828124999998,2.960880234375,"[2.952864990234375, 2.94440087890625, 2.950491455078125, 2.953120361328125, 2.9489072265625, 2.940178466796875, 2.96012841796875, 2.96095458984375, 2.9558095703125, 2.93851171875]",tokens/s,21.352047089286735,kWh,3.488551348987698e-05,1.911889900293555e-05,0.00012785161979329105,0.00018185603228610352,tokens/kWh,346427.88148421585,,s,629,30.434665451049792,0.048385795629649925,0.012948080369770033,0.046676990509033206,0.047583238983154294,0.04774830017089844,0.15544816040039064,"[0.047582206726074217, 0.04730470275878906, 0.04824883270263672, 0.04772249603271484, 0.04878745651245117, 0.04808703994750976, 0.04736000061035156, 0.047241214752197266, 0.047952896118164064, 0.04727807998657227, 0.04649062347412109, 0.04636979293823242, 0.04666470336914062, 0.04662988662719727, 0.04664115142822266, 0.04670873641967774, 0.04672512054443359, 0.046691326141357424, 0.04698214340209961, 0.046519294738769534, 0.04654079818725586, 0.04665856170654297, 0.046635009765625, 0.04673126220703125, 0.04660224151611328, 0.046773246765136715, 0.04655206298828125, 0.046635009765625, 0.04667193603515625, 0.04662572860717774, 0.04678860855102539, 0.046599166870117184, 0.046676990509033206, 0.04663808059692383, 0.046683135986328124, 0.04657766342163086, 0.046693374633789066, 0.04663603210449219, 0.04674560165405273, 0.04665856170654297, 0.04662579345703125, 0.04668415832519531, 0.04674662399291992, 0.04703539276123047, 0.046635009765625, 0.04672819137573242, 0.04648038482666016, 0.04710092926025391, 0.046478336334228515, 0.04648038482666016, 0.04640460968017578, 0.04667494583129883, 0.046835712432861325, 0.0464721908569336, 0.0466165771484375, 0.04666265487670898, 0.046601215362548826, 0.04664524841308594, 0.04665651321411133, 0.046713886260986326, 0.04696368026733398, 0.04662886428833008, 0.15727513122558595, 0.04655718231201172, 0.046685184478759766, 0.046648353576660154, 0.04671894454956055, 0.046706687927246096, 0.04657049560546875, 0.04666470336914062, 0.04660838317871094, 0.046729217529296874, 0.04665651321411133, 0.0466360969543457, 0.046658496856689456, 0.046669822692871094, 0.046614528656005856, 0.046723072052001956, 0.04704358291625976, 0.04674662399291992, 0.047241214752197266, 0.04723814392089844, 0.04694527816772461, 0.047088638305664066, 0.04729446411132812, 0.04644966506958008, 0.04667494583129883, 0.04665958404541016, 0.0466165771484375, 0.04714707183837891, 0.04651411056518555, 0.04662579345703125, 0.04655718231201172, 0.046653438568115234, 0.046568447113037106, 0.04709580612182617, 0.046755840301513675, 0.046535678863525394, 0.04667903900146484, 0.04644147109985351, 0.04665446472167969, 0.04659814453125, 0.04658278274536133, 0.04658380889892578, 0.04668108749389648, 0.04667084884643555, 0.046717952728271485, 0.046685184478759766, 0.04658278274536133, 0.04667391967773438, 0.046680065155029295, 0.04674764633178711, 0.04662988662719727, 0.0466319351196289, 0.04670873641967774, 0.04674662399291992, 0.04654694366455078, 0.04663296127319336, 0.04662783813476563, 0.046737407684326174, 0.04670873641967774, 0.04662374496459961, 0.04715520095825195, 0.04658585739135742, 0.047094783782958984, 0.15548313903808594, 0.046652416229248046, 0.04667289733886719, 0.04660332870483398, 0.04651923370361328, 0.046653438568115234, 0.04668928146362305, 0.046706687927246096, 0.04652544021606445, 0.046633983612060545, 0.04670361709594727, 0.04668620681762695, 0.04676403045654297, 0.04657571029663086, 0.04660316848754883, 0.046712833404541014, 0.04677632141113281, 0.046635009765625, 0.04666777420043945, 0.0466954231262207, 0.04656947326660156, 0.04663808059692383, 0.04665958404541016, 0.04664831924438476, 0.04665139389038086, 0.046630912780761716, 0.046696449279785154, 0.04670054244995117, 0.04667903900146484, 0.0466247673034668, 0.04666470336914062, 0.04714905548095703, 0.04762214279174805, 0.04757503890991211, 0.04760678482055664, 0.04757196807861328, 0.04655104064941406, 0.04684288024902344, 0.046755840301513675, 0.04650188827514649, 0.0472729606628418, 0.04758323287963867, 0.047116287231445314, 0.04761804962158203, 0.047492095947265625, 0.046676990509033206, 0.047413246154785156, 0.04763545608520508, 0.04737843322753906, 0.04647628784179687, 0.046655487060546875, 0.04764879989624023, 0.04658480072021484, 0.0464005126953125, 0.046548992156982424, 0.04667391967773438, 0.046611457824707034, 0.04661862564086914, 0.04665446472167969, 0.046707710266113284, 0.0466431999206543, 0.04665961456298828, 0.046728160858154295, 0.15585382080078125, 0.046706687927246096, 0.0466954231262207, 0.04670361709594727, 0.046717952728271485, 0.04658892822265625, 0.046682113647460936, 0.046685184478759766, 0.04662783813476563, 0.046721023559570314, 0.046693374633789066, 0.04668108749389648, 0.04665961456298828, 0.04667184066772461, 0.046633983612060545, 0.04674662399291992, 0.04663504028320312, 0.04661859130859375, 0.04666470336914062, 0.04666265487670898, 0.046637054443359374, 0.0467322883605957, 0.04677836990356445, 0.04781158447265625, 0.04698624038696289, 0.046614528656005856, 0.04748185729980469, 0.04743987274169922, 0.0471275520324707, 0.04736716842651367, 0.04670566558837891, 0.04651520156860352, 0.04688076782226563, 0.04653670501708984, 0.046868480682373044, 0.04737945556640625, 0.04776243209838867, 0.04832460784912109, 0.047543296813964846, 0.047780864715576174, 0.04768569564819336, 0.0464967041015625, 0.046663681030273435, 0.04671692657470703, 0.0465428466796875, 0.04669235229492188, 0.046626846313476564, 0.04666672134399414, 0.046693374633789066, 0.04731391906738281, 0.046516223907470705, 0.04649062347412109, 0.04653977584838867, 0.04655104064941406, 0.046611457824707034, 0.0466247673034668, 0.04675686264038086, 0.04660326385498047, 0.04664934539794922, 0.04660838317871094, 0.04664934539794922, 0.047282176971435545, 0.047876094818115236, 0.15560089111328124, 0.04665753555297852, 0.04658892822265625, 0.04646297454833984, 0.04610969543457031, 0.04650700759887695, 0.04630323028564453, 0.046453758239746096, 0.046227455139160156, 0.046273536682128906, 0.04635955047607422, 0.046693374633789066, 0.04666572952270508, 0.04664217758178711, 0.04667903900146484, 0.04657356643676758, 0.048691200256347655, 0.04808499145507812, 0.047642623901367184, 0.046473217010498044, 0.046827518463134765, 0.04668928146362305, 0.046736385345458986, 0.046682113647460936, 0.04636979293823242, 0.04649267196655273, 0.04650188827514649, 0.046789630889892575, 0.04672204971313477, 0.046734336853027345, 0.046660606384277346, 0.04664217758178711, 0.046944255828857424, 0.046514175415039063, 0.04661043167114258, 0.046698497772216796, 0.046614528656005856, 0.047119392395019534, 0.04721865463256836, 0.046309375762939455, 0.046453758239746096, 0.046481407165527344, 0.046489601135253904, 0.04689920043945312, 0.04652339172363281, 0.047683582305908204, 0.047164417266845705, 0.046545921325683595, 0.04698624038696289, 0.046429183959960936, 0.04752896118164063, 0.04715212631225586, 0.047666175842285156, 0.04669440078735351, 0.04696473693847656, 0.04652134323120117, 0.046527488708496094, 0.04763238525390625, 0.04729241561889649, 0.04808499145507812, 0.046633983612060545, 0.04718188858032227, 0.04641068649291992, 0.15555072021484376, 0.04655206298828125, 0.04662374496459961, 0.04666265487670898, 0.04666572952270508, 0.04664831924438476, 0.04671897506713867, 0.046584831237792966, 0.046693374633789066, 0.046707710266113284, 0.046653438568115234, 0.04670054244995117, 0.04662783813476563, 0.046665790557861325, 0.046636993408203126, 0.04665856170654297, 0.0465428466796875, 0.04651724624633789, 0.04632371139526367, 0.046494720458984375, 0.04645171356201172, 0.0465428466796875, 0.046584831237792966, 0.046666751861572264, 0.046601215362548826, 0.0466954231262207, 0.04665958404541016, 0.04662374496459961, 0.04667801666259765, 0.04671180725097656, 0.04662169647216797, 0.04667289733886719, 0.046663681030273435, 0.046614528656005856, 0.046653438568115234, 0.046767105102539064, 0.04664115142822266, 0.046707710266113284, 0.046706687927246096, 0.04659814453125, 0.046668800354003906, 0.04672614288330078, 0.04657766342163086, 0.04667801666259765, 0.04664012908935547, 0.04655513763427734, 0.04671590423583984, 0.04671590423583984, 0.04667903900146484, 0.046696449279785154, 0.047449119567871095, 0.04729135894775391, 0.046617599487304685, 0.0465715217590332, 0.04658687973022461, 0.04680704116821289, 0.046698497772216796, 0.0466319351196289, 0.046714881896972656, 0.04661248016357422, 0.046712833404541014, 0.046736385345458986, 0.046595073699951174, 0.1552506866455078, 0.04637081527709961, 0.046486526489257815, 0.046617599487304685, 0.046581760406494144, 0.046545921325683595, 0.04674457550048828, 0.04666777420043945, 0.046714881896972656, 0.04662374496459961, 0.046668800354003906, 0.04666470336914062, 0.0467077751159668, 0.046626750946044924, 0.04668108749389648, 0.046693374633789066, 0.046936065673828124, 0.04641791915893555, 0.04650598526000976, 0.04672204971313477, 0.04665446472167969, 0.04678246307373047, 0.04658182525634766, 0.04666566467285156, 0.046683135986328124, 0.046655487060546875, 0.047250431060791014, 0.04772249603271484, 0.047505409240722656, 0.047683582305908204, 0.04763955307006836, 0.04772556686401367, 0.047505409240722656, 0.04746854400634765, 0.04727705764770508, 0.04767846298217773, 0.04741120147705078, 0.047596542358398435, 0.047375358581542966, 0.04744704055786133, 0.04765388870239258, 0.04801740646362305, 0.04772351837158203, 0.04766003036499023, 0.04641996765136719, 0.04669235229492188, 0.04747468948364258, 0.0478023681640625, 0.047583263397216795, 0.04709379196166992, 0.046405601501464847, 0.047219711303710936, 0.046577632904052736, 0.046354496002197265, 0.04781254577636719, 0.04671180725097656, 0.04717772674560547, 0.046499839782714845, 0.04641996765136719, 0.046475265502929686, 0.04646912002563477, 0.046511104583740234, 0.047297599792480466, 0.155569091796875, 0.046633983612060545, 0.04651827239990235, 0.04679782485961914, 0.04654079818725586, 0.04683161544799805, 0.04654489517211914, 0.04661043167114258, 0.046635009765625, 0.04699852752685547, 0.046755840301513675, 0.04692172622680664, 0.047088638305664066, 0.04734873580932617, 0.048059391021728515, 0.0467138557434082, 0.04768153762817383, 0.04649574279785156, 0.04668415832519531, 0.04657356643676758, 0.04668723297119141, 0.04740095901489258, 0.04648550415039063, 0.047459327697753906, 0.04694937515258789, 0.04746960067749023, 0.04670563125610352, 0.04711116790771484, 0.04659302520751953, 0.046561279296875, 0.04655104064941406, 0.04700364685058594, 0.047713279724121094, 0.04758937454223633, 0.048573440551757815, 0.04681318283081055, 0.04672819137573242, 0.046690303802490236, 0.046497791290283204, 0.047217662811279294, 0.04754739379882812, 0.047631359100341795, 0.04667084884643555, 0.0470129280090332, 0.04737843322753906, 0.047624126434326175, 0.047470592498779295, 0.04712243270874023, 0.04655411148071289, 0.046637054443359374, 0.047661056518554686, 0.04688281631469727, 0.04650598526000976, 0.04779315185546875, 0.04676607894897461, 0.04664627075195313, 0.04662681579589844, 0.04690534210205078, 0.046639102935791016, 0.046516223907470705, 0.046889984130859375, 0.047751167297363284, 0.04753100967407226, 0.15626547241210936, 0.04726067352294922, 0.046535713195800785, 0.047881183624267576, 0.047592449188232425, 0.047077438354492185, 0.046674880981445316, 0.0465428466796875, 0.0468746223449707, 0.04667903900146484, 0.04692377471923828, 0.04800204849243164, 0.04675174331665039, 0.04653055953979492, 0.04704153442382813, 0.04653977584838867, 0.046496768951416016, 0.04745318222045898, 0.047661056518554686, 0.04668415832519531, 0.04676812744140625, 0.04782796859741211, 0.04716339111328125, 0.04757401657104492, 0.04696268844604492, 0.04690124893188476, 0.04645478439331055, 0.046699520111083984, 0.04681625747680664, 0.04649369430541992, 0.04699548721313476, 0.047728607177734375, 0.046565376281738284, 0.046615550994873044, 0.04696473693847656, 0.04654489517211914, 0.04654796981811524, 0.046992385864257816, 0.046773246765136715, 0.046534656524658206, 0.046548992156982424, 0.04774399948120117, 0.04717055892944336, 0.04657254409790039, 0.0474450569152832, 0.046711742401123045, 0.04667596817016602, 0.047234046936035154, 0.046458881378173826, 0.046669822692871094, 0.04778905487060547, 0.0470118408203125, 0.04667596817016602, 0.04657254409790039, 0.04644659042358398, 0.046723072052001956, 0.04658995056152344, 0.04665651321411133, 0.046676990509033206, 0.046698497772216796, 0.04665753555297852, 0.04664217758178711, 0.046630912780761716, 0.15535821533203126, 0.046691326141357424, 0.04637900924682617, 0.046486526489257815, 0.046545921325683595, 0.04657459259033203, 0.04662886428833008, 0.04666265487670898, 0.04666572952270508, 0.046698497772216796, 0.046653438568115234, 0.04659814453125, 0.04668723297119141, 0.04670259094238281, 0.047017982482910156, 0.04664524841308594, 0.046524417877197265, 0.04652851104736328, 0.0466063346862793, 0.046696449279785154, 0.046693374633789066, 0.04659097671508789, 0.046717952728271485, 0.0466165771484375, 0.04657049560546875, 0.04676300811767578, 0.04668723297119141, 0.04664831924438476, 0.046699520111083984, 0.046668800354003906, 0.04665651321411133, 0.046693374633789066, 0.046635009765625, 0.0465530891418457, 0.04670054244995117, 0.04674764633178711, 0.04654796981811524, 0.046706687927246096, 0.04667084884643555, 0.04666777420043945, 0.04692070388793945, 0.046544960021972656, 0.04660627365112305, 0.04667398452758789, 0.04666572952270508, 0.0467496337890625, 0.04657561492919922, 0.046721023559570314, 0.04662783813476563, 0.04666268920898437, 0.046646240234375, 0.046666751861572264, 0.04667084884643555, 0.04650086212158203, 0.04658995056152344, 0.046202880859375, 0.046306304931640625, 0.04681318283081055, 0.04711423873901367, 0.04668108749389648, 0.04639846420288086, 0.04647731018066406, 0.04656742477416992]",tokens/s,20.667222414902657,,,main,False,False,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,M,M,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/M/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669322a9-30c82b134e6b1de73bcdd9cf;f5ad269e-3b15-4360-ac63-2dbd6686f9a4) Repository Not Found for url: https://huggingface.co/M/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: M is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,stabilityai/stablelm-base-alpha-3b,stabilityai/stablelm-base-alpha-3b,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1292.427264,8550.612992,0.0,7904.165888,7764.225024,s,10,0.4270016975402832,0.04270016975402832,0.0002564349820094391,0.04261038398742675,0.04283884506225586,0.04314454307556152,0.043389101486206054,"[0.043450241088867185, 0.04257680130004883, 0.04260335922241211, 0.04263283157348633, 0.04263155364990234, 0.04256819152832031, 0.04277091217041016, 0.04255731201171875, 0.042593086242675784, 0.0426174087524414]",tokens/s,5995.2923249409105,kWh,5.056450261972343e-07,2.7706942103181154e-07,2.7670807986581545e-06,3.5497952458872004e-06,tokens/kWh,72116835.55455828,MB,1292.754944,8550.612992,0.0,7904.165888,7866.521088,s,10,11.46987731933594,1.1469877319335937,0.006756718886698962,1.1452116088867188,1.149155517578125,1.1580077514648437,1.1650895385742186,"[1.1668599853515624, 1.14588037109375, 1.145392578125, 1.1471883544921875, 1.1459871826171875, 1.143307861328125, 1.1442183837890625, 1.14282666015625, 1.1450306396484375, 1.143185302734375]",tokens/s,54.92648111745232,kWh,1.3449507538381353e-05,7.370016608069237e-06,6.197858210894227e-05,8.279810625539286e-05,tokens/kWh,760886.9677003832,,s,629,11.834436584472655,0.018814684554010582,0.0051113973788171445,0.01817087936401367,0.01828678436279297,0.01874044189453125,0.061119956512451175,"[0.018808895111083985, 0.01878112030029297, 0.018876447677612304, 0.018831327438354493, 0.01881088066101074, 0.01883750343322754, 0.018892799377441406, 0.018784255981445314, 0.018769920349121092, 0.018865152359008788, 0.018868223190307617, 0.018769920349121092, 0.018916351318359375, 0.018958335876464845, 0.018961408615112304, 0.019125247955322267, 0.01878118324279785, 0.018892799377441406, 0.01884876823425293, 0.018950143814086915, 0.018900991439819336, 0.018771968841552734, 0.01865011215209961, 0.018696224212646485, 0.01851900863647461, 0.01862553596496582, 0.018909183502197266, 0.018678815841674804, 0.018395103454589844, 0.018354175567626953, 0.018207744598388673, 0.01818726348876953, 0.01818009567260742, 0.018148351669311523, 0.01820159912109375, 0.018155519485473632, 0.018234367370605468, 0.01819545555114746, 0.01823027229309082, 0.01820262336730957, 0.018206720352172853, 0.01823744010925293, 0.01817190361022949, 0.018231296539306642, 0.01820057678222656, 0.01825382423400879, 0.01819545555114746, 0.018265087127685545, 0.018207744598388673, 0.018274303436279296, 0.018255872726440428, 0.018353151321411132, 0.01838387107849121, 0.01834294319152832, 0.0183090877532959, 0.01825279998779297, 0.0182989444732666, 0.018263999938964843, 0.01821696090698242, 0.018275327682495117, 0.018233343124389647, 0.018326528549194337, 0.06113689422607422, 0.018082815170288084, 0.01807974433898926, 0.01807360076904297, 0.018121728897094725, 0.018143232345581056, 0.018118656158447266, 0.018153472900390624, 0.01815449523925781, 0.018177024841308592, 0.018104320526123048, 0.018131967544555663, 0.01817087936401367, 0.018108415603637695, 0.018182144165039063, 0.01810534477233887, 0.018207744598388673, 0.018141183853149414, 0.018192384719848635, 0.018098175048828127, 0.01821900749206543, 0.018139135360717772, 0.018206720352172853, 0.018110496520996094, 0.018215904235839842, 0.018160640716552736, 0.018174016952514648, 0.01815648078918457, 0.018177024841308592, 0.018156543731689453, 0.018176000595092775, 0.01819340705871582, 0.01818009567260742, 0.01819443130493164, 0.018184192657470705, 0.0181790714263916, 0.018215936660766603, 0.018148351669311523, 0.018174976348876954, 0.018197504043579102, 0.018199552536010744, 0.018168832778930662, 0.01824870491027832, 0.018191360473632814, 0.01823539161682129, 0.01820262336730957, 0.01827123260498047, 0.01820159912109375, 0.01822105598449707, 0.01820159912109375, 0.018214912414550782, 0.01824665641784668, 0.01823744010925293, 0.018251775741577148, 0.018214912414550782, 0.018259967803955078, 0.01825279998779297, 0.018223104476928712, 0.01825279998779297, 0.01822208023071289, 0.018284543991088868, 0.018217983245849608, 0.018256895065307616, 0.06113792037963867, 0.018117631912231445, 0.018126848220825196, 0.018101247787475586, 0.018130943298339842, 0.018117631912231445, 0.018159616470336915, 0.018128896713256838, 0.018163711547851562, 0.018092031478881835, 0.018160640716552736, 0.018114559173583983, 0.018165760040283203, 0.018101247787475586, 0.01817804718017578, 0.018085887908935547, 0.01821286392211914, 0.018091007232666014, 0.01822208023071289, 0.018102272033691406, 0.018183168411254884, 0.018125823974609375, 0.01817804718017578, 0.01812384033203125, 0.018181055068969727, 0.018124799728393554, 0.01817804718017578, 0.018140287399291993, 0.018175872802734375, 0.01822211265563965, 0.018164703369140625, 0.01814531135559082, 0.018158559799194337, 0.018159616470336915, 0.018165760040283203, 0.018144256591796876, 0.018197504043579102, 0.018140159606933593, 0.01818726348876953, 0.018150400161743165, 0.01820159912109375, 0.01818726348876953, 0.018207744598388673, 0.018144256591796876, 0.01820364761352539, 0.018191360473632814, 0.01820876884460449, 0.01821900749206543, 0.018214912414550782, 0.018242559432983398, 0.01820364761352539, 0.018265087127685545, 0.018233343124389647, 0.018255872726440428, 0.0182108154296875, 0.01822003173828125, 0.01823027229309082, 0.01820159912109375, 0.01823846435546875, 0.01823027229309082, 0.01825484848022461, 0.018198528289794923, 0.018288639068603514, 0.06123622512817383, 0.01821286392211914, 0.01808896064758301, 0.018147327423095702, 0.01815449523925781, 0.018086912155151368, 0.018121728897094725, 0.018145280838012694, 0.018121728897094725, 0.01813199996948242, 0.018120672225952147, 0.01815449523925781, 0.018116607666015624, 0.018095104217529297, 0.018189311981201172, 0.018152448654174806, 0.01822719955444336, 0.01820057678222656, 0.018164735794067383, 0.01818009567260742, 0.018132991790771484, 0.01822822380065918, 0.018121728897094725, 0.018187295913696288, 0.01813603210449219, 0.01824870491027832, 0.018149375915527344, 0.01826304054260254, 0.018156543731689453, 0.01821286392211914, 0.018353151321411132, 0.01819647979736328, 0.018163711547851562, 0.01820159912109375, 0.018184192657470705, 0.018197504043579102, 0.018239488601684572, 0.01820979118347168, 0.018199552536010744, 0.018280448913574218, 0.01817190361022949, 0.01824358367919922, 0.018160640716552736, 0.01823846435546875, 0.018172927856445312, 0.018310144424438478, 0.018200639724731446, 0.01828755187988281, 0.018322431564331054, 0.018266111373901366, 0.01824563217163086, 0.018258943557739257, 0.018272256851196288, 0.018273279190063475, 0.018225151062011717, 0.018272256851196288, 0.018276351928710938, 0.01828659248352051, 0.01820057678222656, 0.01828556823730469, 0.01823744010925293, 0.018297855377197265, 0.018192384719848635, 0.06128025436401367, 0.018132991790771484, 0.018123775482177733, 0.018151424407958985, 0.01808793640136719, 0.018130943298339842, 0.018103328704833985, 0.018142175674438477, 0.018164735794067383, 0.01815449523925781, 0.01808793640136719, 0.018110464096069336, 0.018084863662719726, 0.018117631912231445, 0.018116607666015624, 0.01813715171813965, 0.018095039367675782, 0.018191360473632814, 0.018095104217529297, 0.01817804718017578, 0.01815449523925781, 0.018165760040283203, 0.018117631912231445, 0.01814630317687988, 0.018127872467041017, 0.018155519485473632, 0.018199552536010744, 0.01816268730163574, 0.01814121627807617, 0.018132959365844727, 0.018198528289794923, 0.01818524742126465, 0.018224096298217772, 0.01822719955444336, 0.01818828773498535, 0.01822412872314453, 0.01821696090698242, 0.018249727249145507, 0.018215936660766603, 0.0181790714263916, 0.01825484848022461, 0.018329599380493163, 0.018494464874267577, 0.018546688079833985, 0.018257919311523436, 0.018190336227416993, 0.01821183967590332, 0.018176000595092775, 0.01817705535888672, 0.018180063247680664, 0.01819545555114746, 0.018174976348876954, 0.018182144165039063, 0.01820364761352539, 0.018191360473632814, 0.018155519485473632, 0.01817804718017578, 0.01828556823730469, 0.018217023849487306, 0.018186176300048828, 0.01823244857788086, 0.018178943634033204, 0.018229248046875, 0.06107648086547852, 0.01804800033569336, 0.018069503784179687, 0.01801420783996582, 0.018095104217529297, 0.018059328079223634, 0.01813395118713379, 0.018098175048828127, 0.01810534477233887, 0.018086912155151368, 0.018076671600341796, 0.018100223541259765, 0.018085887908935547, 0.01808793640136719, 0.01808896064758301, 0.01809715270996094, 0.018110464096069336, 0.018128896713256838, 0.018092031478881835, 0.018121728897094725, 0.018074623107910158, 0.018120704650878908, 0.018122751235961913, 0.01809110450744629, 0.018084768295288087, 0.018094079971313477, 0.018114559173583983, 0.018126848220825196, 0.018155519485473632, 0.01810534477233887, 0.018147327423095702, 0.01821183967590332, 0.01817087936401367, 0.018113536834716795, 0.018163711547851562, 0.018112512588500978, 0.01817804718017578, 0.018139135360717772, 0.018159616470336915, 0.01819545555114746, 0.0182938232421875, 0.018144191741943358, 0.01818009567260742, 0.01816166305541992, 0.018152448654174806, 0.018177087783813477, 0.01816364860534668, 0.01820262336730957, 0.01819443130493164, 0.018190336227416993, 0.01816166305541992, 0.018191423416137695, 0.018180063247680664, 0.018189279556274415, 0.018167808532714845, 0.01821286392211914, 0.018304000854492186, 0.0181790714263916, 0.01820876884460449, 0.01816166305541992, 0.018217983245849608, 0.01822719955444336, 0.018182144165039063, 0.061764606475830076, 0.018349056243896485, 0.018149375915527344, 0.01805619239807129, 0.018100223541259765, 0.018074623107910158, 0.018109439849853515, 0.018053119659423827, 0.018098175048828127, 0.018078720092773438, 0.018117631912231445, 0.018076671600341796, 0.018126848220825196, 0.018085887908935547, 0.018116607666015624, 0.018099199295043944, 0.018081792831420897, 0.018126848220825196, 0.018082815170288084, 0.018150400161743165, 0.018164735794067383, 0.018282495498657226, 0.018099264144897462, 0.01813804817199707, 0.018095104217529297, 0.01816166305541992, 0.018114559173583983, 0.01816268730163574, 0.018120704650878908, 0.01816268730163574, 0.018132991790771484, 0.018156543731689453, 0.01814735984802246, 0.018129888534545897, 0.018151424407958985, 0.018130943298339842, 0.01813817596435547, 0.01813804817199707, 0.018168832778930662, 0.018131967544555663, 0.018153472900390624, 0.0181790714263916, 0.01817804718017578, 0.01814630317687988, 0.018174976348876954, 0.018185216903686522, 0.018165760040283203, 0.01817087936401367, 0.018172927856445312, 0.018181119918823242, 0.018164735794067383, 0.01820467185974121, 0.018156543731689453, 0.018206720352172853, 0.01818623924255371, 0.018199552536010744, 0.018197504043579102, 0.01819241523742676, 0.018196447372436524, 0.018184192657470705, 0.018184192657470705, 0.018198528289794923, 0.018219072341918944, 0.06106515121459961, 0.01808896064758301, 0.018041856765747072, 0.01805619239807129, 0.018058240890502928, 0.01810742378234863, 0.01806742477416992, 0.018112512588500978, 0.018068479537963866, 0.018099199295043944, 0.01805721664428711, 0.01819443130493164, 0.018127872467041017, 0.018132991790771484, 0.018094079971313477, 0.018115583419799804, 0.01808291244506836, 0.018118560791015623, 0.01808793640136719, 0.018094079971313477, 0.018092031478881835, 0.018077856063842775, 0.01811542320251465, 0.01807257652282715, 0.018124799728393554, 0.018113632202148438, 0.018151391983032228, 0.018099136352539062, 0.018168832778930662, 0.018117664337158203, 0.018145376205444336, 0.018119552612304687, 0.01816268730163574, 0.018130943298339842, 0.01819340705871582, 0.018145280838012694, 0.018123775482177733, 0.01813811111450195, 0.018119680404663087, 0.018148351669311523, 0.018127872467041017, 0.018157567977905274, 0.018153472900390624, 0.01820364761352539, 0.018215936660766603, 0.018181119918823242, 0.018182144165039063, 0.01815555191040039, 0.018165727615356446, 0.01817087936401367, 0.01819443130493164, 0.018159616470336915, 0.018191360473632814, 0.018135040283203126, 0.018175071716308593, 0.01816054344177246, 0.018166784286499024, 0.018143295288085937, 0.018219968795776368, 0.018172927856445312, 0.01817190361022949, 0.01820262336730957, 0.018177024841308592, 0.061142017364501956, 0.018111488342285157, 0.018096128463745118, 0.01806438446044922, 0.018086912155151368, 0.01805516815185547, 0.018069503784179687, 0.018119680404663087, 0.018165760040283203, 0.018060287475585936, 0.01806550407409668, 0.018056095123291014, 0.018089984893798827, 0.01805516815185547, 0.018132991790771484, 0.018078720092773438, 0.01812076759338379, 0.01813804817199707, 0.018123775482177733, 0.018076671600341796, 0.018123775482177733, 0.018089984893798827, 0.018142208099365235, 0.018131967544555663, 0.018103296279907227, 0.018132991790771484, 0.018129919052124025, 0.018135040283203126, 0.018177024841308592, 0.018217983245849608, 0.01821183967590332, 0.018234399795532225, 0.018197471618652344, 0.01823744010925293, 0.018225151062011717, 0.018207744598388673, 0.01819340705871582, 0.01822105598449707, 0.01818623924255371, 0.01824563217163086, 0.01824870491027832, 0.01821183967590332, 0.018220064163208007, 0.018357311248779297, 0.01858755111694336, 0.018370559692382812, 0.01824563217163086, 0.018259967803955078, 0.01814630317687988, 0.01818726348876953, 0.01817190361022949, 0.018137088775634767, 0.018168832778930662, 0.018164735794067383, 0.018166784286499024, 0.01817087936401367, 0.01821286392211914, 0.01817923164367676, 0.018199392318725586, 0.018160640716552736, 0.0182108154296875, 0.018191360473632814, 0.018225183486938478, 0.061136863708496095, 0.01805516815185547, 0.018050048828125, 0.01807360076904297, 0.01804902458190918, 0.018084863662719726, 0.01808896064758301, 0.01807360076904297, 0.018084863662719726, 0.01806540870666504, 0.01807974433898926, 0.018068479537963866, 0.018076671600341796, 0.01805414390563965, 0.018373632431030275, 0.018312255859375, 0.01837664031982422, 0.018116607666015624, 0.018137088775634767, 0.01808896064758301, 0.018126848220825196, 0.018118656158447266, 0.01808793640136719, 0.018085887908935547, 0.018122751235961913, 0.018116607666015624, 0.018121728897094725, 0.018118656158447266, 0.018115583419799804, 0.018136064529418947, 0.018100223541259765, 0.018150400161743165, 0.018135040283203126, 0.018143232345581056, 0.018127872467041017, 0.018149375915527344, 0.018136064529418947, 0.018134016036987305, 0.018149375915527344, 0.018145280838012694, 0.01814531135559082, 0.018139104843139648, 0.018140159606933593, 0.018135040283203126, 0.018184192657470705, 0.018151424407958985, 0.018198528289794923, 0.018157567977905274, 0.01818623924255371, 0.018172927856445312, 0.018153472900390624, 0.018164735794067383, 0.018156543731689453, 0.018183168411254884, 0.01814531135559082, 0.018189279556274415, 0.018163711547851562, 0.01820364761352539, 0.01817804718017578, 0.01818623924255371, 0.018181119918823242, 0.018185216903686522, 0.01817087936401367]",tokens/s,53.149974272985496,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,google/recurrentgemma-9b,google/recurrentgemma-9b,cuda,0,42,,,True,,,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",recurrent_gemma,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,mistralai/Mixtral-8x22B-v0.1,mistralai/Mixtral-8x22B-v0.1,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/base.py"", line 43, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 54, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1021, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 403 Client Error: Forbidden for url: https://huggingface.co/mistralai/Mixtral-8x22B-v0.1/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 398, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 321, in hf_raise_for_status raise GatedRepoError(message, response) from e huggingface_hub.utils._errors.GatedRepoError: 403 Client Error. (Request ID: Root=1-6647508f-62c78f273aa930ee0b1e1c1a;a5382531-7132-4f1a-9703-a62d3ed10a9c) Cannot access gated repo for url https://huggingface.co/mistralai/Mixtral-8x22B-v0.1/resolve/main/config.json. Access to model mistralai/Mixtral-8x22B-v0.1 is restricted and you are not in the authorized list. Visit https://huggingface.co/mistralai/Mixtral-8x22B-v0.1 to ask for access. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 87, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/base.py"", line 56, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 37, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 62, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 928, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 631, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 686, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 416, in cached_file raise EnvironmentError( OSError: You are trying to access a gated repo. Make sure to have access to it at https://huggingface.co/mistralai/Mixtral-8x22B-v0.1. 403 Client Error. (Request ID: Root=1-6647508f-62c78f273aa930ee0b1e1c1a;a5382531-7132-4f1a-9703-a62d3ed10a9c) Cannot access gated repo for url https://huggingface.co/mistralai/Mixtral-8x22B-v0.1/resolve/main/config.json. Access to model mistralai/Mixtral-8x22B-v0.1 is restricted and you are not in the authorized list. Visit https://huggingface.co/mistralai/Mixtral-8x22B-v0.1 to ask for access. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,01-ai/Yi-6B,01-ai/Yi-6B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1279.205376,13032.226816,0.0,12385.779712,12315.692032,s,10,0.8440754318237305,0.08440754318237305,0.0002870512613258965,0.08434864044189452,0.08466916046142578,0.08491845397949219,0.08511788879394531,"[0.0851677474975586, 0.08434102630615234, 0.08444358062744141, 0.08430166625976562, 0.0841434555053711, 0.08413075256347656, 0.08435689544677734, 0.08422029113769532, 0.08461376190185547, 0.08435625457763672]",tokens/s,3032.904292059301,kWh,9.987584891773405e-07,5.472732936576896e-07,5.296120016537809e-06,6.842151799372839e-06,tokens/kWh,37415130.13836748,MB,1279.205376,13032.226816,0.0,12385.779712,12315.694592,s,10,20.45511767578125,2.045511767578125,0.015862182844826427,2.050250732421875,2.059363671875,2.064078564453125,2.067850478515625,"[2.06879345703125, 2.041267578125, 2.0355511474609376, 2.048385009765625, 2.0151358642578123, 2.02331005859375, 2.052116455078125, 2.056490966796875, 2.05831591796875, 2.055751220703125]",tokens/s,30.79913838608304,kWh,2.3632088703295543e-05,1.2951004029029809e-05,9.896785228346235e-05,0.0001355509450157877,tokens/kWh,464769.9061977204,,s,629,21.180157951354964,0.033672747140468966,0.010029304652149776,0.03254886245727539,0.03288842391967774,0.03337935256958008,0.1163402426147461,"[0.034119678497314454, 0.034342910766601564, 0.03362713623046875, 0.03365785598754883, 0.03420876693725586, 0.03378073501586914, 0.03395072174072265, 0.03372851181030274, 0.03338137435913086, 0.03354009628295898, 0.033376319885253906, 0.03336185455322266, 0.033532928466796875, 0.033906688690185545, 0.033296382904052735, 0.03293798446655274, 0.03262771224975586, 0.03251200103759765, 0.03267379379272461, 0.032510974884033206, 0.03247923278808594, 0.03244236755371094, 0.03250380706787109, 0.032570369720458986, 0.03281305694580078, 0.032704513549804685, 0.03265126419067383, 0.03261443328857422, 0.03264303970336914, 0.032524288177490236, 0.0324771842956543, 0.03264409637451172, 0.03251200103759765, 0.032742401123046876, 0.032778240203857424, 0.03249868774414062, 0.032492542266845705, 0.03274956893920898, 0.032673854827880856, 0.03260102462768555, 0.03270348739624023, 0.032568321228027344, 0.03222531127929688, 0.032085983276367185, 0.03148595237731933, 0.03255807876586914, 0.03256934356689453, 0.03271987152099609, 0.032435199737548825, 0.03261542510986328, 0.03268307113647461, 0.032457664489746095, 0.03238604736328125, 0.03244240188598633, 0.03250787353515625, 0.0325478401184082, 0.03259392166137695, 0.03248537445068359, 0.032451583862304685, 0.0324659194946289, 0.0320819206237793, 0.03237376022338867, 0.11638066864013671, 0.03243212890625, 0.03244339370727539, 0.03253145599365234, 0.032454654693603514, 0.03263385772705078, 0.03245977783203125, 0.03250483322143555, 0.032464897155761716, 0.032555007934570314, 0.032494590759277346, 0.032574462890625, 0.03250592041015625, 0.032559040069580075, 0.03243110275268555, 0.031546367645263675, 0.03256524658203125, 0.03251507186889648, 0.032205825805664064, 0.03261337661743164, 0.03250688171386719, 0.03261439895629883, 0.03254684829711914, 0.03253859329223633, 0.032517120361328124, 0.03263180923461914, 0.032261119842529294, 0.031422464370727536, 0.032105472564697264, 0.03230822372436523, 0.03161087989807129, 0.03177779197692871, 0.03271475219726563, 0.03252019119262695, 0.03230822372436523, 0.03271475219726563, 0.03237580871582031, 0.03164057540893555, 0.032712703704833986, 0.032562206268310544, 0.03146339225769043, 0.032405502319335935, 0.03252019119262695, 0.03153510475158691, 0.032494590759277346, 0.03252633666992188, 0.032307201385498044, 0.03262259292602539, 0.03266969680786133, 0.03270044708251953, 0.032627681732177734, 0.032625663757324216, 0.032702465057373044, 0.0326195182800293, 0.03276902389526367, 0.03268505477905274, 0.03252633666992188, 0.03261542510986328, 0.031456256866455076, 0.03179417610168457, 0.03253350448608398, 0.033260543823242186, 0.03287859344482422, 0.11655577850341797, 0.03160371208190918, 0.03147881507873535, 0.031556575775146485, 0.031497215270996096, 0.03142560005187988, 0.03144595146179199, 0.031513599395751955, 0.03144601631164551, 0.031442943572998046, 0.031528959274291994, 0.03143987274169922, 0.031531007766723636, 0.03157811164855957, 0.03149926376342774, 0.03149004745483398, 0.031528959274291994, 0.031492095947265625, 0.03354316711425781, 0.03253657531738281, 0.0328458251953125, 0.03251302337646484, 0.03256320190429687, 0.0326297607421875, 0.03245363235473633, 0.03259494400024414, 0.03260518264770508, 0.031548448562622074, 0.032221153259277345, 0.03265945434570312, 0.03280281448364258, 0.03250380706787109, 0.032595966339111326, 0.032494590759277346, 0.03227443313598633, 0.032677886962890625, 0.03266764831542969, 0.03261644744873047, 0.03263897705078125, 0.03267583847045898, 0.03266252899169922, 0.032593952178955075, 0.03313046264648437, 0.033279998779296875, 0.03290009689331055, 0.03249356842041016, 0.032497665405273435, 0.032573440551757815, 0.03257548904418945, 0.03264924621582031, 0.032671710968017575, 0.03245363235473633, 0.03252435302734375, 0.03268703842163086, 0.03255091094970703, 0.032497665405273435, 0.032625663757324216, 0.0324659194946289, 0.0325591049194336, 0.03280384063720703, 0.03268201446533203, 0.03274441528320313, 0.032718849182128903, 0.11735244750976563, 0.03264921569824219, 0.03268403244018555, 0.03250688171386719, 0.03262259292602539, 0.03273011016845703, 0.032702465057373044, 0.03254272079467774, 0.032553985595703126, 0.032648193359375, 0.03259699249267578, 0.03260518264770508, 0.032901119232177735, 0.03474227142333984, 0.032881664276123046, 0.032723968505859374, 0.03266867065429688, 0.03271372985839844, 0.03265945434570312, 0.03256934356689453, 0.032435199737548825, 0.0325928955078125, 0.03245772933959961, 0.032573440551757815, 0.03218431854248047, 0.03228876876831055, 0.03270553588867187, 0.03184025573730469, 0.032271358489990236, 0.03251302337646484, 0.03267583847045898, 0.032574462890625, 0.032246784210205076, 0.03258367919921875, 0.03261955261230469, 0.03242287826538086, 0.032589824676513675, 0.032623615264892575, 0.03265126419067383, 0.032517120361328124, 0.032389118194580076, 0.032449535369873043, 0.03252633666992188, 0.032092159271240234, 0.03213824081420898, 0.03205120086669922, 0.032497665405273435, 0.032555007934570314, 0.03219251251220703, 0.03211980819702148, 0.032322559356689456, 0.031645696640014646, 0.03142451286315918, 0.03239014434814453, 0.032487422943115234, 0.03249663925170899, 0.032454654693603514, 0.03250790405273438, 0.03249158477783203, 0.032527294158935544, 0.032481281280517575, 0.032471038818359374, 0.03256524658203125, 0.11616563415527344, 0.031410175323486327, 0.03148492813110351, 0.03140300750732422, 0.032553985595703126, 0.032399360656738284, 0.03228876876831055, 0.03256729507446289, 0.03314688110351562, 0.03218431854248047, 0.031663103103637694, 0.031403072357177736, 0.03148384094238281, 0.031440895080566404, 0.03141836738586426, 0.032742401123046876, 0.03357183837890625, 0.03276800155639648, 0.03210649490356445, 0.03257855987548828, 0.032543743133544925, 0.03203276824951172, 0.031507455825805664, 0.03143270492553711, 0.03150028800964356, 0.03150028800964356, 0.031704063415527346, 0.032510974884033206, 0.03146854400634766, 0.03165286445617676, 0.03247513580322266, 0.032510974884033206, 0.0325294075012207, 0.03161087989807129, 0.03223654556274414, 0.03160985565185547, 0.03158118438720703, 0.031527936935424807, 0.03157708740234375, 0.03248540878295898, 0.032527328491210934, 0.03170611190795898, 0.03169075202941894, 0.03180953598022461, 0.031532064437866214, 0.03156067276000977, 0.0320819206237793, 0.032543743133544925, 0.03316633605957031, 0.0337448959350586, 0.03260211181640625, 0.0324956169128418, 0.03246899032592773, 0.031457279205322264, 0.031441919326782225, 0.03139891242980957, 0.03148595237731933, 0.03144812774658203, 0.031480768203735354, 0.031474687576293944, 0.03148902320861816, 0.03147980880737305, 0.031488000869750975, 0.11623628997802735, 0.031426559448242186, 0.03149516868591309, 0.03146649551391602, 0.03143680000305176, 0.031425535202026365, 0.03156684875488281, 0.032099327087402346, 0.031562751770019534, 0.03153715133666992, 0.03150540733337402, 0.03149516868591309, 0.032320510864257815, 0.03247622299194336, 0.03263379287719727, 0.031955968856811526, 0.031410207748413084, 0.031388639450073245, 0.031410175323486327, 0.03154944038391113, 0.031526912689208986, 0.03142451286315918, 0.031437824249267575, 0.03162112045288086, 0.03151974487304687, 0.032551937103271485, 0.032709632873535156, 0.03263078308105469, 0.032353279113769534, 0.03265433502197266, 0.03244851303100586, 0.03216281509399414, 0.032217086791992186, 0.03246182250976563, 0.03264921569824219, 0.03229183959960937, 0.03253452682495117, 0.03263078308105469, 0.03260006332397461, 0.03258163070678711, 0.0323061752319336, 0.03246284866333008, 0.032574462890625, 0.032484352111816404, 0.03192831993103027, 0.032276481628417966, 0.03253964614868164, 0.0326297607421875, 0.032522239685058595, 0.032571392059326174, 0.03247513580322266, 0.03198464012145996, 0.03235123062133789, 0.03265945434570312, 0.03228876876831055, 0.032525310516357424, 0.032508926391601564, 0.03197849655151367, 0.03184025573730469, 0.03190480041503906, 0.03253654479980469, 0.0324771842956543, 0.03220787048339844, 0.11679129791259765, 0.032753662109375, 0.032200702667236326, 0.03238812637329101, 0.03259900665283203, 0.03253350448608398, 0.03255807876586914, 0.03256524658203125, 0.03322060775756836, 0.03297689437866211, 0.03253145599365234, 0.032473087310791016, 0.03264409637451172, 0.03250790405273438, 0.03250380706787109, 0.032508926391601564, 0.03251103973388672, 0.03251808166503906, 0.032601089477539064, 0.032522239685058595, 0.03249868774414062, 0.03252019119262695, 0.0324956169128418, 0.03239219284057617, 0.03228876876831055, 0.03254886245727539, 0.03254886245727539, 0.0325478401184082, 0.03259494400024414, 0.03257855987548828, 0.03252326583862305, 0.03263897705078125, 0.03254886245727539, 0.03261238479614258, 0.03257648086547851, 0.03272192001342773, 0.03272806549072266, 0.0326932487487793, 0.03275980758666992, 0.03283967971801758, 0.032628734588623046, 0.03258367919921875, 0.03262771224975586, 0.03241062545776367, 0.03150643157958984, 0.032704513549804685, 0.032538623809814454, 0.032661502838134765, 0.0325560302734375, 0.032694271087646484, 0.032589824676513675, 0.03251507186889648, 0.03257958221435547, 0.03243417739868164, 0.032486400604248046, 0.03266867065429688, 0.032435199737548825, 0.03246694564819336, 0.03281203079223633, 0.03252838516235351, 0.03253145599365234, 0.032593952178955075, 0.033390560150146485, 0.11712921905517579, 0.033188865661621096, 0.03282636642456055, 0.03250688171386719, 0.03269222259521484, 0.0325662727355957, 0.03255091094970703, 0.03261337661743164, 0.032672767639160154, 0.032481281280517575, 0.032694271087646484, 0.032538623809814454, 0.031939584732055666, 0.03247206497192383, 0.031926271438598636, 0.03246387100219727, 0.03254988861083984, 0.03264614486694336, 0.032492576599121095, 0.03260105514526367, 0.0326901741027832, 0.03251609420776367, 0.03258883285522461, 0.03308643341064453, 0.033972225189208984, 0.03269734573364258, 0.03275161743164062, 0.03252326583862305, 0.03183718490600586, 0.032791553497314455, 0.03274649429321289, 0.03262464141845703, 0.03310182571411133, 0.032740352630615234, 0.032892929077148435, 0.03301792144775391, 0.032632766723632814, 0.03248230361938476, 0.0327014389038086, 0.03286937713623047, 0.032723968505859374, 0.032553985595703126, 0.03254476928710937, 0.032727039337158204, 0.03263283157348633, 0.032502784729003906, 0.03277004623413086, 0.03271680068969727, 0.03257958221435547, 0.03259494400024414, 0.032625663757324216, 0.03252838516235351, 0.03238918304443359, 0.03274233627319336, 0.03260927963256836, 0.032510974884033206, 0.0325478401184082, 0.03242496109008789, 0.03259904098510742, 0.03225804901123047, 0.03284479904174804, 0.03301990509033203, 0.032489471435546875, 0.11785625457763672, 0.03262259292602539, 0.032399360656738284, 0.03247206497192383, 0.03338444900512695, 0.03304550552368164, 0.03244134521484375, 0.03262054443359375, 0.0325560302734375, 0.03259494400024414, 0.03259084701538086, 0.03267583847045898, 0.032510974884033206, 0.032525310516357424, 0.03270041656494141, 0.03266764831542969, 0.03266048049926758, 0.03247411346435547, 0.03246284866333008, 0.03279974365234375, 0.032072704315185545, 0.03166720008850098, 0.031513599395751955, 0.03147776031494141, 0.03221299362182617, 0.0315156478881836, 0.03204710388183594, 0.03258265686035156, 0.032950271606445314, 0.03267686462402344, 0.032712703704833986, 0.03294003295898437, 0.03276800155639648, 0.03243315124511719, 0.03263488006591797, 0.03282534408569336, 0.03263078308105469, 0.03286937713623047, 0.03333222579956055, 0.03263078308105469, 0.0325928955078125, 0.03272294235229492, 0.033329151153564454, 0.03259187316894531, 0.032851966857910156, 0.03259084701538086, 0.03258777618408203, 0.03288780975341797, 0.03152179145812988, 0.031718399047851564, 0.032307201385498044, 0.031902719497680664, 0.03260825729370117, 0.03197337532043457, 0.033296382904052735, 0.037326847076416016, 0.033637374877929685, 0.03275059127807617, 0.03273420715332031, 0.03268710327148437, 0.03267379379272461, 0.0331141128540039, 0.03388313674926758, 0.11669606781005859, 0.03168262481689453, 0.03217407989501953, 0.032575424194335935, 0.03260723114013672, 0.03229695892333984, 0.032456703186035156, 0.03256729507446289, 0.03263897705078125, 0.032500736236572264, 0.03258060836791992, 0.032715774536132815, 0.032584705352783204, 0.032546817779541014, 0.03245772933959961, 0.03235635375976562, 0.032747520446777346, 0.03261337661743164, 0.032510974884033206, 0.03265126419067383, 0.032500736236572264, 0.032586753845214846, 0.03274137496948242, 0.032737281799316405, 0.032718849182128903, 0.03263283157348633, 0.03275161743164062, 0.03289702224731445, 0.03328102493286133, 0.033519615173339845, 0.03277926254272461, 0.032833534240722655, 0.032645118713378905, 0.03291545486450195, 0.03264716720581055, 0.03267692947387695, 0.03261433410644531, 0.032683006286621095, 0.03263385772705078, 0.03263488006591797, 0.0328089599609375, 0.03256940841674805, 0.03246176147460938, 0.03256729507446289, 0.03262054443359375, 0.03282227325439453, 0.03258879852294922, 0.03260416030883789, 0.03288678359985352, 0.032729087829589845, 0.032541694641113283, 0.0328908805847168, 0.032686080932617184, 0.03269222259521484, 0.03297484970092773, 0.03277414321899414, 0.032551937103271485, 0.0325478401184082, 0.032451583862304685, 0.032527359008789065, 0.03270553588867187, 0.03250175857543945, 0.03236044692993164]",tokens/s,29.697606667742548,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,/,/,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 373, in cached_file raise EnvironmentError( OSError: / does not appear to have a file named config.json. Checkout 'https://huggingface.co///tree/None' for available files. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,TencentARC/Mistral_Pro_8B_v0.1,,cuda,0,42,,,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,d35829e539df8480b726c647eeabf91e41eae047,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,,MB,1273.401344,19665.518592,0.0,19019.071488,18747.268096,s,10,1.1008828201293945,0.11008828201293945,0.0024261382345637157,0.10917001342773437,0.11099395599365235,0.11411876983642578,0.11661862091064454,"[0.11724358367919922, 0.10893663787841797, 0.10915385437011718, 0.10922672271728516, 0.10884496307373047, 0.11029955291748048, 0.1088978271484375, 0.10918617248535156, 0.109942626953125, 0.10915087890625]",tokens/s,2325.4064403503953,kWh,1.2924292993142412e-06,7.081907045294824e-07,7.784082314217451e-06,9.784702318061173e-06,tokens/kWh,26163289.559403386,MB,1273.401344,19665.518592,0.0,19019.071488,18774.938112,s,10,38.20687036132812,3.8206870361328122,0.009300846714813165,3.8190885009765623,3.832152001953125,3.835840771484375,3.838791787109375,"[3.831332275390625, 3.8135830078125, 3.82000732421875, 3.80987060546875, 3.818169677734375, 3.8121875, 3.811255859375, 3.839529541015625, 3.822679443359375, 3.828255126953125]",tokens/s,16.489180978237556,kWh,4.4929080872576014e-05,2.4622353401834478e-05,0.0001603518021943793,0.0002299032364687898,tokens/kWh,274028.33021253475,,s,629,39.12849711990357,0.06220746759921075,0.01302544399920906,0.06040678405761719,0.0615464973449707,0.061790821838378904,0.16975769592285156,"[0.06134272003173828, 0.0619950065612793, 0.061774848937988285, 0.06159564971923828, 0.06201446533203125, 0.062325759887695314, 0.06233804702758789, 0.061328384399414064, 0.060622848510742185, 0.06049689483642578, 0.06022655868530274, 0.0613488655090332, 0.06141747283935547, 0.06154035186767578, 0.06064332962036133, 0.060407806396484375, 0.06013030242919922, 0.06017340850830078, 0.060087230682373045, 0.06018966293334961, 0.06021324920654297, 0.060456993103027344, 0.06024905776977539, 0.06022041702270508, 0.060268543243408204, 0.060680191040039064, 0.06026956939697266, 0.06055116653442383, 0.06043852615356445, 0.0637757453918457, 0.061446144104003904, 0.061497344970703124, 0.06033407974243164, 0.06032691192626953, 0.06030950546264648, 0.06028595352172852, 0.06125363159179688, 0.06104985427856445, 0.06138163375854492, 0.06055731201171875, 0.06028287887573242, 0.060662784576416016, 0.06029312133789062, 0.060247039794921874, 0.060371967315673826, 0.060488704681396485, 0.06159360122680664, 0.060967937469482425, 0.06091571044921875, 0.06028799819946289, 0.06044467163085938, 0.06030847930908203, 0.06021529769897461, 0.06033817672729492, 0.060470272064208984, 0.06047129440307617, 0.06045183944702148, 0.0603770866394043, 0.06053171157836914, 0.06042316818237305, 0.060903423309326174, 0.06041292953491211, 0.1694924774169922, 0.06027478408813477, 0.06027356719970703, 0.0603054084777832, 0.06034124755859375, 0.060386302947998044, 0.06021222305297851, 0.06032793426513672, 0.06028902435302735, 0.06040063858032227, 0.06031465530395508, 0.06060335922241211, 0.060214271545410154, 0.060189697265625, 0.06023168182373047, 0.06027264022827149, 0.06023680114746094, 0.060911617279052734, 0.06027471923828125, 0.060278751373291015, 0.06024806213378906, 0.060246017456054686, 0.060235774993896485, 0.06028499221801758, 0.060525505065917966, 0.06073651123046875, 0.06159158325195312, 0.06137750244140625, 0.060680191040039064, 0.06032486343383789, 0.06043545532226562, 0.06034431838989258, 0.06042726516723633, 0.0603955192565918, 0.06041907119750976, 0.06037913513183594, 0.06034431838989258, 0.06019174575805664, 0.060393470764160156, 0.060229633331298826, 0.06048255920410156, 0.06029619216918945, 0.060655616760253904, 0.06044467163085938, 0.06025113677978516, 0.060265472412109375, 0.06167654418945313, 0.06060748672485351, 0.06104678344726563, 0.060919807434082034, 0.061067264556884764, 0.060439552307128906, 0.06025625610351563, 0.060486656188964844, 0.060286975860595705, 0.060870655059814455, 0.06133964920043945, 0.06048153686523437, 0.06125056076049805, 0.06043648147583008, 0.06093107223510742, 0.06026444625854492, 0.06207385635375977, 0.17065676879882813, 0.06058291244506836, 0.06093209457397461, 0.06032691192626953, 0.06032486343383789, 0.06081433486938476, 0.06026137542724609, 0.060290046691894535, 0.060418048858642576, 0.06025830459594726, 0.06055321502685547, 0.060846080780029295, 0.06034431838989258, 0.06034124755859375, 0.06014361572265625, 0.060300289154052736, 0.060319744110107425, 0.06026963043212891, 0.06040057754516601, 0.06020608139038086, 0.06024192047119141, 0.060316673278808595, 0.060401664733886716, 0.060319744110107425, 0.060044288635253906, 0.06027980804443359, 0.06083379364013672, 0.06154342269897461, 0.0626319351196289, 0.06146662521362305, 0.06138982391357422, 0.0615464973449707, 0.06143078231811523, 0.06135504150390625, 0.06046819305419922, 0.061312000274658204, 0.06117171096801758, 0.06123212814331055, 0.06126489639282227, 0.06147174453735352, 0.06051430511474609, 0.06050406265258789, 0.06032486343383789, 0.061238273620605466, 0.061795326232910154, 0.060472320556640625, 0.06027980804443359, 0.06030131149291992, 0.06030745697021484, 0.06014976119995117, 0.06030233764648438, 0.06030847930908203, 0.06020505523681641, 0.060186622619628906, 0.06035763168334961, 0.06024192047119141, 0.06024806213378906, 0.060271648406982424, 0.06025212860107422, 0.060375038146972655, 0.06031155014038086, 0.06063417434692383, 0.06038214492797851, 0.1700823059082031, 0.060647422790527344, 0.06041907119750976, 0.06032281494140625, 0.060295169830322265, 0.060175361633300783, 0.06024911880493164, 0.060349407196044924, 0.06031564712524414, 0.06074367904663086, 0.061982719421386716, 0.060865535736083984, 0.060165119171142575, 0.060216320037841796, 0.06033817672729492, 0.0603422737121582, 0.060859390258789066, 0.060728321075439455, 0.06251219177246094, 0.060615615844726564, 0.060788734436035156, 0.060470272064208984, 0.060249088287353515, 0.06025932693481445, 0.060249088287353515, 0.06035763168334961, 0.060903423309326174, 0.06058086395263672, 0.06045286560058594, 0.06036275100708008, 0.06034124755859375, 0.06034636688232422, 0.06025625610351563, 0.060252159118652344, 0.0603422737121582, 0.0599818229675293, 0.06021017456054688, 0.060375038146972655, 0.06026342391967773, 0.060400737762451175, 0.060367774963378903, 0.0603351058959961, 0.060217342376708984, 0.06029926300048828, 0.060418048858642576, 0.06038118362426758, 0.06060748672485351, 0.06020710372924805, 0.06031772613525391, 0.060523487091064455, 0.06047129440307617, 0.060273662567138675, 0.06057164764404297, 0.060641281127929686, 0.06106623840332031, 0.06032998275756836, 0.06033817672729492, 0.06018252944946289, 0.060181503295898435, 0.06055321502685547, 0.06041190338134766, 0.060381214141845704, 0.0603616943359375, 0.16966554260253905, 0.06024192047119141, 0.060243968963623044, 0.06023884963989258, 0.06034329605102539, 0.06096384048461914, 0.06052249526977539, 0.06031769561767578, 0.06015999984741211, 0.060237823486328126, 0.060175361633300783, 0.06040678405761719, 0.060265472412109375, 0.060388351440429686, 0.06024499130249023, 0.060388351440429686, 0.06059622573852539, 0.06225612640380859, 0.06142566299438477, 0.06144716644287109, 0.06152703857421875, 0.06159360122680664, 0.061505535125732425, 0.06107033538818359, 0.06093414306640625, 0.060711936950683595, 0.06042316818237305, 0.060611583709716796, 0.0607825927734375, 0.06037094497680664, 0.061341697692871094, 0.060352512359619144, 0.06077337646484375, 0.0627322883605957, 0.06178406524658203, 0.06041702270507812, 0.06040678405761719, 0.06035968017578125, 0.06044979095458984, 0.06058905410766602, 0.06035763168334961, 0.06024294281005859, 0.060158977508544924, 0.060368896484375, 0.06032896041870117, 0.060336128234863284, 0.060368896484375, 0.06030745697021484, 0.060303359985351565, 0.060382209777832034, 0.06030438232421875, 0.06035148620605469, 0.06030643081665039, 0.06042316818237305, 0.06047641754150391, 0.06032998275756836, 0.0603351058959961, 0.06040371322631836, 0.06028083038330078, 0.06037913513183594, 0.06031872177124024, 0.06038118362426758, 0.06036070251464844, 0.17123532104492187, 0.060281856536865235, 0.06057984161376953, 0.06063616180419922, 0.06180249786376953, 0.06084710311889648, 0.06022655868530274, 0.06032588958740234, 0.060262401580810546, 0.06033919906616211, 0.06041702270507812, 0.06036896133422852, 0.06052345657348633, 0.06038323211669922, 0.06023891067504883, 0.0606627197265625, 0.06034739303588867, 0.060308544158935544, 0.06044255828857422, 0.060418048858642576, 0.060464126586914066, 0.06022553634643555, 0.06031052780151367, 0.060252159118652344, 0.06023987197875977, 0.06025932693481445, 0.0603351058959961, 0.06032793426513672, 0.0605030403137207, 0.06064441680908203, 0.06087366485595703, 0.060816383361816405, 0.060864513397216796, 0.06024499130249023, 0.06022655868530274, 0.06037913513183594, 0.06034022521972656, 0.06054707336425781, 0.060508159637451174, 0.060450817108154295, 0.060184574127197264, 0.061416446685791014, 0.06073548889160156, 0.06056345748901367, 0.06003609466552735, 0.060402687072753904, 0.06054608154296875, 0.060288993835449216, 0.06035865783691406, 0.06022348785400391, 0.06042726516723633, 0.06077132797241211, 0.060939262390136716, 0.061104129791259766, 0.061758464813232425, 0.06065459060668945, 0.060249088287353515, 0.0605747184753418, 0.060864513397216796, 0.06015488052368164, 0.06032281494140625, 0.06022252655029297, 0.060238784790039065, 0.1697935333251953, 0.0602531852722168, 0.060286975860595705, 0.06022655868530274, 0.060260353088378904, 0.06086860656738281, 0.060252159118652344, 0.06023884963989258, 0.0603135986328125, 0.06023372650146484, 0.06019174575805664, 0.060286975860595705, 0.060268543243408204, 0.060355583190917966, 0.06023680114746094, 0.06026444625854492, 0.06043852615356445, 0.060418048858642576, 0.06031052780151367, 0.06030847930908203, 0.06016716766357422, 0.060314624786376954, 0.06029312133789062, 0.06028287887573242, 0.060281856536865235, 0.06035763168334961, 0.06027980804443359, 0.060300289154052736, 0.060252159118652344, 0.06033919906616211, 0.06029926300048828, 0.06051942443847656, 0.060240894317626956, 0.06065971374511719, 0.06016204833984375, 0.06014976119995117, 0.06032793426513672, 0.06025523376464844, 0.060230655670166014, 0.06022348785400391, 0.060249088287353515, 0.06029209518432617, 0.06046515274047851, 0.06028595352172852, 0.060286975860595705, 0.06018048095703125, 0.060129280090332034, 0.06018867111206055, 0.06028083038330078, 0.05994188690185547, 0.06065356826782226, 0.06021945571899414, 0.06165190505981445, 0.06201753616333008, 0.061628414154052735, 0.06147686386108398, 0.061459457397460934, 0.061538303375244144, 0.06137855911254883, 0.0613939208984375, 0.0615464973449707, 0.06094131088256836, 0.060267520904541017, 0.17014790344238281, 0.060321727752685544, 0.060237823486328126, 0.06024399948120117, 0.06152291107177734, 0.06187519836425781, 0.060644351959228515, 0.060270591735839846, 0.06026956939697266, 0.06032281494140625, 0.06021222305297851, 0.06032486343383789, 0.060352512359619144, 0.06030131149291992, 0.06027775955200195, 0.06031769561767578, 0.06046515274047851, 0.06033715057373047, 0.06044467163085938, 0.06029414367675781, 0.060194816589355465, 0.06018252944946289, 0.06015590286254883, 0.06026956939697266, 0.06026137542724609, 0.0602501106262207, 0.06033715057373047, 0.0602347526550293, 0.06036684799194336, 0.060249088287353515, 0.0602716178894043, 0.06039961624145508, 0.061306880950927733, 0.06060134506225586, 0.061873214721679684, 0.06122796630859375, 0.06044876861572265, 0.060346431732177734, 0.060787647247314454, 0.06117171096801758, 0.06175948715209961, 0.06170828628540039, 0.06187724685668945, 0.06137753677368164, 0.06173593521118164, 0.06162124633789062, 0.061459457397460934, 0.0616519660949707, 0.06177382278442383, 0.06181990432739258, 0.061661182403564455, 0.06186700820922852, 0.061615169525146486, 0.0616242561340332, 0.0617625617980957, 0.06176563262939453, 0.0616099853515625, 0.061827072143554686, 0.06152601623535156, 0.06144307327270508, 0.06170624160766602, 0.06178406524658203, 0.06160179138183594, 0.17100697326660155, 0.06032896041870117, 0.06029926300048828, 0.06025113677978516, 0.060382209777832034, 0.060409854888916016, 0.060777473449707034, 0.06056755065917969, 0.060181503295898435, 0.060453887939453124, 0.06197043228149414, 0.061328384399414064, 0.061316158294677736, 0.0614901123046875, 0.06159872055053711, 0.061369342803955076, 0.060572673797607425, 0.06111641693115234, 0.06030438232421875, 0.060590080261230465, 0.06036787033081055, 0.06038118362426758, 0.06074982452392578, 0.061195262908935545, 0.061925376892089844, 0.061521919250488284, 0.06075596618652344, 0.060524543762207034, 0.060224510192871096, 0.06027468872070312, 0.06028083038330078, 0.06027167892456055, 0.06030227279663086, 0.06042931365966797, 0.06030649566650391, 0.06048147201538086, 0.06070169448852539, 0.06052556610107422, 0.060793857574462894, 0.06031769561767578, 0.06141132736206055, 0.06138265609741211, 0.06138374328613281, 0.060510208129882816, 0.06024800109863281, 0.06029312133789062, 0.06010060882568359, 0.060268543243408204, 0.06032486343383789, 0.06125260925292969, 0.061052928924560546, 0.060303359985351565, 0.0606668815612793, 0.06055833435058594, 0.06085222244262695, 0.06076316833496094, 0.06034735870361328, 0.06028902435302735, 0.06038425445556641, 0.060426239013671876, 0.06058598327636719, 0.06035456085205078, 0.06030233764648438, 0.16998503112792968, 0.060443679809570314, 0.060333023071289064, 0.060104705810546874, 0.060641281127929686, 0.06022860717773437, 0.06033820724487305, 0.06049276733398438, 0.061276161193847656, 0.06052249526977539, 0.06084505462646484, 0.06114713668823242, 0.06113385772705078, 0.06115331268310547, 0.06088185501098633, 0.06097715377807617, 0.06134988784790039, 0.06095667266845703, 0.06141747283935547, 0.06155263900756836, 0.060527614593505856, 0.06099148941040039, 0.06169804763793945, 0.061142017364501956, 0.060662784576416016, 0.06107955169677735, 0.06043859100341797, 0.06115116882324219, 0.06056755065917969, 0.06081126403808594, 0.06098227310180664, 0.060351585388183596, 0.06113987350463867, 0.06030137634277344, 0.061045696258544925, 0.06038937759399414, 0.060767230987548826, 0.06071603012084961, 0.06045798492431641, 0.06154751968383789, 0.0620840950012207, 0.06156595230102539, 0.06044467163085938, 0.06040063858032227, 0.0603873291015625, 0.060421119689941405, 0.060385280609130856, 0.06074265670776367, 0.06043852615356445, 0.060486656188964844, 0.061080577850341794, 0.06103859329223633, 0.06031769561767578, 0.06070374298095703, 0.06041292953491211, 0.06031769561767578, 0.06035968017578125, 0.06057984161376953, 0.06048255920410156, 0.06036172866821289, 0.06051430511474609, 0.060652542114257815, 0.06068121719360352]",tokens/s,16.075240458955562,,,main,False,False,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-160m,EleutherAI/pythia-160m,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1276.112896,1147.666432,0.0,501.219328,454.728704,s,18,0.17622518444061278,0.00979028802447849,0.0003311221958476073,0.00963588809967041,0.01004348497390747,0.010192420911788938,0.010792147817611694,"[0.010942079544067384, 0.009920319557189942, 0.009861920356750488, 0.009587936401367188, 0.009930879592895507, 0.009594655990600586, 0.009879967689514161, 0.010036352157592774, 0.010060128211975097, 0.009802687644958495, 0.009627488136291504, 0.0095600004196167, 0.009626111984252929, 0.009532832145690917, 0.009532320022583007, 0.009644288063049317, 0.009565216064453125, 0.009520000457763671]",tokens/s,26148.36247513123,kWh,1.172617049904202e-07,6.425279449331356e-08,2.612984343472248e-07,4.4281293383095857e-07,tokens/kWh,578122228.240349,MB,1276.112896,1147.666432,0.0,501.219328,469.748736,s,18,10.280933654785159,0.5711629808213976,0.009247880941270055,0.5692927551269531,0.5811380493164062,0.5899757080078125,0.5981793701171875,"[0.6002302856445313, 0.572284423828125, 0.5696863403320312, 0.57040966796875, 0.5737970581054688, 0.56949755859375, 0.57095556640625, 0.5781260375976562, 0.5881660766601563, 0.56612841796875, 0.5648677978515625, 0.5669077758789063, 0.5627127075195313, 0.5634201049804688, 0.5633443603515625, 0.5645104370117188, 0.5668010864257812, 0.5690879516601562]",tokens/s,110.30126621546586,kWh,6.747217727512963e-06,3.6971926956577298e-06,1.1739112400541698e-05,2.2183522823712388e-05,tokens/kWh,2839945.6885476327,,s,1133,10.436439044952387,0.009211331901987991,0.0012552244200118663,0.008954879760742187,0.00932971534729004,0.009483263778686522,0.01900560600280762,"[0.009881600379943848, 0.0098088960647583, 0.009772031784057618, 0.009533439636230469, 0.009441280364990234, 0.009479167938232422, 0.00941055965423584, 0.0095283203125, 0.00982528018951416, 0.009777152061462402, 0.009639936447143555, 0.009496671676635742, 0.009455519676208496, 0.009660415649414063, 0.0098088960647583, 0.009852928161621094, 0.009687040328979492, 0.00962771224975586, 0.009784255981445313, 0.009752639770507813, 0.009404352188110351, 0.009387007713317871, 0.009478143692016602, 0.009404416084289552, 0.009727999687194825, 0.00953651237487793, 0.009561087608337402, 0.009534496307373047, 0.009504735946655274, 0.009562111854553223, 0.0095283842086792, 0.009438176155090332, 0.009414624214172364, 0.009407487869262696, 0.009444352149963378, 0.009541631698608399, 0.009444352149963378, 0.009408512115478516, 0.009456640243530273, 0.009432064056396485, 0.009445376396179199, 0.009663488388061523, 0.009561087608337402, 0.009589759826660157, 0.009451519966125489, 0.009422847747802734, 0.009636863708496094, 0.009337856292724609, 0.009251839637756347, 0.009408512115478516, 0.009323519706726074, 0.009018367767333984, 0.009038847923278808, 0.009225215911865235, 0.009522175788879395, 0.009539584159851074, 0.009669631958007812, 0.009662464141845703, 0.009332736015319825, 0.009241600036621094, 0.009463808059692384, 0.009225215911865235, 0.019723264694213868, 0.009153535842895508, 0.009170944213867188, 0.009274368286132812, 0.009189375877380371, 0.009183232307434081, 0.00919961643218994, 0.009189375877380371, 0.009220095634460449, 0.009343999862670899, 0.00919961643218994, 0.00923136043548584, 0.009187328338623046, 0.009275391578674316, 0.009218048095703125, 0.009225279808044433, 0.00923027229309082, 0.009179136276245118, 0.00912384033203125, 0.009063424110412598, 0.00892518424987793, 0.008942591667175292, 0.00890880012512207, 0.009026559829711914, 0.00893337631225586, 0.008969216346740723, 0.008928256034851074, 0.008910880088806153, 0.00896713638305664, 0.008943615913391113, 0.00897433567047119, 0.008953856468200684, 0.008950783729553222, 0.008951807975769043, 0.008962047576904298, 0.008934399604797364, 0.008961088180541992, 0.00893945598602295, 0.008966143608093263, 0.008918016433715821, 0.008963071823120117, 0.008990752220153808, 0.008932319641113282, 0.008897567749023437, 0.008908767700195313, 0.008931327819824218, 0.009082880020141602, 0.009260031700134277, 0.008994815826416015, 0.008944640159606934, 0.008931327819824218, 0.008944640159606934, 0.009037887573242188, 0.009296832084655761, 0.009174015998840332, 0.009154560089111329, 0.009236479759216308, 0.008972288131713867, 0.009174015998840332, 0.009256959915161133, 0.009192447662353515, 0.009192447662353515, 0.009225215911865235, 0.019504127502441407, 0.009028608322143555, 0.008922112464904786, 0.008922112464904786, 0.008904704093933105, 0.00891596794128418, 0.008911871910095214, 0.008984576225280762, 0.00892518424987793, 0.008927231788635253, 0.008991744041442871, 0.008929280281066895, 0.008910847663879394, 0.008960000038146973, 0.008940544128417969, 0.008939519882202148, 0.00894156837463379, 0.008923135757446288, 0.008960000038146973, 0.008993791580200196, 0.00890777587890625, 0.008914943695068359, 0.008977408409118653, 0.008962047576904298, 0.008918016433715821, 0.00894156837463379, 0.008954879760742187, 0.00897644805908203, 0.00893126392364502, 0.008914943695068359, 0.008936448097229004, 0.00910540771484375, 0.009194496154785157, 0.009159680366516113, 0.008923135757446288, 0.008961024284362793, 0.008913920402526856, 0.009186304092407227, 0.00913920021057129, 0.00909721565246582, 0.008931327819824218, 0.008928256034851074, 0.00890675163269043, 0.008979455947875976, 0.009170944213867188, 0.00921497631072998, 0.009194496154785157, 0.009061375617980956, 0.008963071823120117, 0.008910847663879394, 0.009191424369812011, 0.009201663970947266, 0.009202688217163087, 0.00921497631072998, 0.009234432220458985, 0.009260031700134277, 0.009177087783813476, 0.00924675178527832, 0.009285599708557128, 0.009208831787109375, 0.00923033618927002, 0.009206784248352052, 0.009226240158081055, 0.018976768493652343, 0.00890060806274414, 0.00899071979522705, 0.008931327819824218, 0.0089169921875, 0.008942591667175292, 0.008910847663879394, 0.008972288131713867, 0.008965120315551758, 0.008884223937988281, 0.008945664405822755, 0.008859647750854491, 0.008872960090637207, 0.008901632308959961, 0.008820735931396484, 0.00881766414642334, 0.008902655601501466, 0.009004032135009766, 0.00892518424987793, 0.00894156837463379, 0.008943615913391113, 0.008938528060913086, 0.008933343887329102, 0.008901632308959961, 0.009233407974243164, 0.009194496154785157, 0.009187328338623046, 0.009206784248352052, 0.009259008407592773, 0.009190400123596192, 0.009211903572082519, 0.009177087783813476, 0.009181183815002441, 0.009094143867492676, 0.009048064231872559, 0.00923136043548584, 0.00920576000213623, 0.008881152153015137, 0.008851455688476563, 0.008851455688476563, 0.00890060806274414, 0.008994815826416015, 0.009095168113708496, 0.00890777587890625, 0.008926207542419434, 0.008910880088806153, 0.008997856140136718, 0.008896512031555176, 0.008946687698364257, 0.008910847663879394, 0.008910847663879394, 0.00893545627593994, 0.008927200317382812, 0.00961638355255127, 0.009489407539367676, 0.009220128059387206, 0.009099231719970704, 0.009400320053100587, 0.009209856033325196, 0.009805824279785156, 0.009313280105590821, 0.009256959915161133, 0.009243647575378417, 0.019607551574707033, 0.00919961643218994, 0.009338879585266113, 0.009190400123596192, 0.009269248008728028, 0.009232383728027344, 0.009252863883972168, 0.009169919967651367, 0.009242624282836913, 0.009203712463378906, 0.00918835163116455, 0.009161727905273438, 0.009211903572082519, 0.009215999603271484, 0.009174015998840332, 0.009174015998840332, 0.008920063972473144, 0.00891596794128418, 0.008894463539123536, 0.0089303035736084, 0.008983551979064941, 0.009185279846191406, 0.009203712463378906, 0.009168895721435547, 0.009206784248352052, 0.009146368026733399, 0.009296895980834961, 0.009198592185974122, 0.009153535842895508, 0.009219072341918945, 0.009194496154785157, 0.009215999603271484, 0.00921497631072998, 0.009227264404296874, 0.009156607627868652, 0.008936448097229004, 0.009210880279541016, 0.009232383728027344, 0.009181247711181641, 0.009132991790771485, 0.008947711944580078, 0.008931327819824218, 0.00898252773284912, 0.00893545627593994, 0.008943584442138672, 0.008987648010253906, 0.008937472343444825, 0.008932352066040039, 0.008961024284362793, 0.008918016433715821, 0.00891596794128418, 0.008905728340148926, 0.008979455947875976, 0.009030655860900879, 0.008939519882202148, 0.008919039726257324, 0.008903679847717285, 0.008946687698364257, 0.009577471733093262, 0.009234432220458985, 0.00910540771484375, 0.008940544128417969, 0.008943615913391113, 0.019324928283691405, 0.009176063537597656, 0.009275391578674316, 0.009064448356628419, 0.008910847663879394, 0.008938495635986327, 0.009251839637756347, 0.009166848182678223, 0.009054207801818847, 0.009000991821289063, 0.008952799797058106, 0.00898252773284912, 0.009194496154785157, 0.009248767852783203, 0.008919039726257324, 0.008935423851013183, 0.008910847663879394, 0.008958975791931152, 0.008969216346740723, 0.00890777587890625, 0.00893337631225586, 0.008952832221984864, 0.008970239639282226, 0.008913920402526856, 0.008926207542419434, 0.008943615913391113, 0.008929280281066895, 0.008919039726257324, 0.009005056381225587, 0.008942591667175292, 0.00923852825164795, 0.009070591926574707, 0.00889958381652832, 0.00893337631225586, 0.009168895721435547, 0.008943615913391113, 0.009005056381225587, 0.008965120315551758, 0.008977408409118653, 0.008985600471496581, 0.00892518424987793, 0.008904704093933105, 0.008903679847717285, 0.008931327819824218, 0.008896512031555176, 0.008914943695068359, 0.008909824371337891, 0.00889350414276123, 0.00889030361175537, 0.00890060806274414, 0.00923750400543213, 0.009146368026733399, 0.009132032394409179, 0.009132032394409179, 0.009182208061218262, 0.009168895721435547, 0.009200639724731445, 0.00900607967376709, 0.008929280281066895, 0.009415679931640625, 0.009255935668945312, 0.009272319793701172, 0.009187328338623046, 0.019438592910766602, 0.008918016433715821, 0.00890880012512207, 0.008913920402526856, 0.00892518424987793, 0.008905728340148926, 0.008919039726257324, 0.008956928253173829, 0.008966143608093263, 0.008902655601501466, 0.008929280281066895, 0.00890060806274414, 0.008926207542419434, 0.00892518424987793, 0.008939519882202148, 0.00894156837463379, 0.00892422389984131, 0.008897472381591796, 0.008947711944580078, 0.00897433567047119, 0.008926207542419434, 0.009004032135009766, 0.00890675163269043, 0.008956928253173829, 0.008890368461608887, 0.008965120315551758, 0.008888319969177246, 0.008994815826416015, 0.008923135757446288, 0.009013248443603515, 0.008935423851013183, 0.008964096069335938, 0.008913920402526856, 0.008922112464904786, 0.00890060806274414, 0.009037823677062988, 0.00922111988067627, 0.009232383728027344, 0.009218048095703125, 0.009263104438781738, 0.008979455947875976, 0.00922111988067627, 0.009264127731323242, 0.009250816345214843, 0.009285632133483887, 0.009240575790405273, 0.009207807540893554, 0.009294848442077636, 0.009203712463378906, 0.009194496154785157, 0.009227264404296874, 0.009203712463378906, 0.009166848182678223, 0.009276415824890137, 0.009186304092407227, 0.00918835163116455, 0.009158656120300293, 0.00920473575592041, 0.009247743606567382, 0.009174015998840332, 0.009217023849487305, 0.00918015956878662, 0.009158656120300293, 0.01967616081237793, 0.00927948760986328, 0.009332736015319825, 0.009197567939758301, 0.009159680366516113, 0.009116671562194823, 0.009142271995544434, 0.009150464057922364, 0.009201663970947266, 0.009250816345214843, 0.009208831787109375, 0.009172032356262207, 0.009252799987792968, 0.009141247749328613, 0.009211903572082519, 0.00921292781829834, 0.009431039810180664, 0.009178112030029297, 0.009144319534301757, 0.00917407989501953, 0.009192383766174317, 0.00921292781829834, 0.009177087783813476, 0.009143296241760255, 0.008985600471496581, 0.009044992446899413, 0.009219072341918945, 0.009134079933166504, 0.00893337631225586, 0.008943615913391113, 0.008920063972473144, 0.008901632308959961, 0.009057279586791991, 0.009197567939758301, 0.009187328338623046, 0.009208831787109375, 0.009165823936462402, 0.009170944213867188, 0.009192447662353515, 0.009161727905273438, 0.009185279846191406, 0.009118720054626465, 0.009160736083984374, 0.009204704284667969, 0.00922111988067627, 0.009120767593383788, 0.008970272064208985, 0.008934368133544921, 0.008938495635986327, 0.008880127906799316, 0.008894463539123536, 0.009210880279541016, 0.009358336448669433, 0.009310208320617675, 0.009340928077697755, 0.009309184074401856, 0.009300031661987306, 0.009349056243896484, 0.009318400382995605, 0.009288703918457031, 0.009307135581970214, 0.00935321617126465, 0.009315327644348144, 0.019891199111938478, 0.009323583602905274, 0.009331647872924804, 0.00927948760986328, 0.009292799949645996, 0.009305088043212891, 0.009280511856079102, 0.009398271560668945, 0.009327615737915039, 0.009302016258239745, 0.009329728126525879, 0.009317312240600586, 0.009289728164672852, 0.009329664230346679, 0.00932044792175293, 0.009323519706726074, 0.009361408233642577, 0.009302016258239745, 0.009325568199157714, 0.009357312202453612, 0.009275391578674316, 0.009318400382995605, 0.009308192253112793, 0.009307104110717774, 0.009296895980834961, 0.009333760261535644, 0.009325568199157714, 0.009340959548950195, 0.00937775993347168, 0.009346048355102539, 0.00931430435180664, 0.009335807800292969, 0.009308159828186035, 0.009264127731323242, 0.009264127731323242, 0.009673727989196777, 0.009313280105590821, 0.009370623588562011, 0.00932863998413086, 0.009283583641052246, 0.009409536361694336, 0.009313280105590821, 0.009256959915161133, 0.00931123161315918, 0.009313280105590821, 0.009317376136779786, 0.009343999862670899, 0.009373696327209472, 0.009334783554077148, 0.009378815650939941, 0.009321503639221192, 0.009284576416015626, 0.009332736015319825, 0.009300992012023926, 0.00929792022705078, 0.009334783554077148, 0.009296895980834961, 0.009362431526184082, 0.009437184333801269, 0.009296928405761719, 0.009306079864501953, 0.009360383987426758, 0.009326592445373535, 0.019224576950073242, 0.008909824371337891, 0.0089169921875, 0.008895487785339355, 0.008895487785339355, 0.008914943695068359, 0.008929280281066895, 0.008911871910095214, 0.008928256034851074, 0.008977408409118653, 0.008911871910095214, 0.008918016433715821, 0.008914943695068359, 0.008929280281066895, 0.008932352066040039, 0.008926207542419434, 0.00892416000366211, 0.008902655601501466, 0.008893440246582032, 0.00892416000366211, 0.008910847663879394, 0.008932352066040039, 0.008931327819824218, 0.008896512031555176, 0.008880127906799316, 0.008942591667175292, 0.008950783729553222, 0.00890777587890625, 0.00890777587890625, 0.008945664405822755, 0.0088985595703125, 0.00891596794128418, 0.008926207542419434, 0.008918016433715821, 0.00892518424987793, 0.008954879760742187, 0.008902655601501466, 0.008935423851013183, 0.008931327819824218, 0.008934399604797364, 0.008980480194091797, 0.009018367767333984, 0.009036800384521485, 0.01044480037689209, 0.009338879585266113, 0.009308192253112793, 0.009228256225585938, 0.009224191665649414, 0.009118720054626465, 0.008940608024597167, 0.008926143646240234, 0.00897433567047119, 0.008977408409118653, 0.008932352066040039, 0.008904704093933105, 0.008927231788635253, 0.008923135757446288, 0.008951807975769043, 0.008940544128417969, 0.008901632308959961, 0.008893440246582032, 0.008976384162902832, 0.008966143608093263, 0.019070976257324217, 0.008905728340148926, 0.008931327819824218, 0.008920063972473144, 0.008926239967346192, 0.009015263557434082, 0.008915007591247559, 0.008929216384887695, 0.008940544128417969, 0.008942591667175292, 0.008919039726257324, 0.008911871910095214, 0.008905728340148926, 0.008948736190795899, 0.008932352066040039, 0.00892416000366211, 0.008929280281066895, 0.008871935844421386, 0.008965120315551758, 0.008940544128417969, 0.008912896156311035, 0.008942591667175292, 0.00902143955230713, 0.008879103660583497, 0.008920063972473144, 0.008913920402526856, 0.0089169921875, 0.008902655601501466, 0.00889958381652832, 0.008914943695068359, 0.008920063972473144, 0.00903987216949463, 0.008928256034851074, 0.008934399604797364, 0.008901632308959961, 0.008937503814697266, 0.008902624130249023, 0.008945664405822755, 0.008910847663879394, 0.008944640159606934, 0.00889958381652832, 0.008909824371337891, 0.008895487785339355, 0.009209919929504394, 0.009177023887634278, 0.009071616172790528, 0.009143296241760255, 0.009193471908569336, 0.009182208061218262, 0.009183232307434081, 0.009142271995544434, 0.008952832221984864, 0.0089169921875, 0.008904704093933105, 0.008948736190795899, 0.00893337631225586, 0.008932352066040039, 0.00892416000366211, 0.009006112098693848, 0.008916959762573243, 0.00891596794128418, 0.008890368461608887, 0.008955904006958008, 0.018922496795654296, 0.008934399604797364, 0.00894976043701172, 0.008928256034851074, 0.008914943695068359, 0.00890880012512207, 0.008939519882202148, 0.008923135757446288, 0.00890675163269043, 0.008948736190795899, 0.008962047576904298, 0.00892416000366211, 0.00894156837463379, 0.008911871910095214, 0.008947711944580078, 0.008946687698364257, 0.008946687698364257, 0.008940544128417969, 0.008901632308959961, 0.008934399604797364, 0.00893337631225586, 0.0089303035736084, 0.008910847663879394, 0.008934399604797364, 0.008919039726257324, 0.008954879760742187, 0.008920127868652344, 0.008944576263427734, 0.008938495635986327, 0.008962047576904298, 0.009143296241760255, 0.009061375617980956, 0.010007552146911621, 0.009490431785583496, 0.009315327644348144, 0.009257984161376954, 0.009264127731323242, 0.009258048057556152, 0.00922105598449707, 0.009163776397705077, 0.009189375877380371, 0.009136128425598144, 0.00890060806274414, 0.008882176399230958, 0.008921088218688965, 0.008863743782043456, 0.008821760177612305, 0.00879923152923584, 0.008847359657287598, 0.008897536277770996, 0.00889241600036621, 0.008914943695068359, 0.00890880012512207, 0.00892416000366211, 0.008955904006958008, 0.008928256034851074, 0.008940544128417969, 0.008921088218688965, 0.008948736190795899, 0.008926207542419434, 0.008928256034851074, 0.008943615913391113, 0.008928256034851074, 0.019016767501831056, 0.00891385555267334, 0.008902655601501466, 0.00890880012512207, 0.008960000038146973, 0.008894463539123536, 0.008836095809936523, 0.0088340482711792, 0.008896512031555176, 0.0089169921875, 0.008942591667175292, 0.00881766414642334, 0.008828927993774414, 0.008882176399230958, 0.008952832221984864, 0.009177087783813476, 0.009170944213867188, 0.009135104179382325, 0.008999936103820801, 0.008936448097229004, 0.008877056121826172, 0.008909824371337891, 0.008922112464904786, 0.00890777587890625, 0.00889241600036621, 0.008904704093933105, 0.00890880012512207, 0.008987648010253906, 0.008938495635986327, 0.008890368461608887, 0.008890368461608887, 0.008886272430419923, 0.00890880012512207, 0.008887295722961425, 0.008918016433715821, 0.008893440246582032, 0.00888319969177246, 0.008943615913391113, 0.00888319969177246, 0.00888323211669922, 0.00895792007446289, 0.0089303035736084, 0.008890368461608887, 0.008897536277770996, 0.008953920364379883, 0.008920063972473144, 0.00893331241607666, 0.008918016433715821, 0.008923135757446288, 0.008928256034851074, 0.008940544128417969, 0.008880160331726073, 0.008882176399230958, 0.008926176071166992, 0.008901632308959961, 0.008901632308959961, 0.008878080368041993, 0.008885248184204102, 0.008960000038146973, 0.009125887870788574, 0.00898252773284912, 0.008919039726257324, 0.008910847663879394, 0.018911231994628908, 0.008936448097229004, 0.008946687698364257, 0.008905728340148926, 0.008966143608093263, 0.008955904006958008, 0.009013248443603515, 0.009174015998840332, 0.009162752151489258, 0.009236479759216308, 0.00900607967376709, 0.008918016433715821, 0.008926207542419434, 0.008956928253173829, 0.008936448097229004, 0.008927231788635253, 0.008918016433715821, 0.008858624458312989, 0.008816639900207519, 0.00882380771636963, 0.008828927993774414, 0.008831999778747558, 0.008956928253173829, 0.008884223937988281, 0.0088985595703125, 0.009042943954467773, 0.009096192359924317, 0.008903679847717285, 0.008902655601501466, 0.008935423851013183, 0.008903679847717285, 0.008928256034851074, 0.00894976043701172, 0.008912896156311035, 0.00892518424987793, 0.0089303035736084, 0.008896512031555176, 0.008926207542419434, 0.00890777587890625, 0.008932352066040039, 0.008945664405822755, 0.008887295722961425, 0.00892518424987793, 0.008880127906799316, 0.00892518424987793, 0.008929280281066895, 0.0089303035736084, 0.00890777587890625, 0.008910847663879394, 0.008904704093933105, 0.00890060806274414, 0.008944640159606934, 0.008951807975769043, 0.008926207542419434, 0.008889344215393067, 0.008922112464904786, 0.00892518424987793, 0.008928256034851074, 0.008920063972473144, 0.008939519882202148, 0.008919039726257324, 0.008957951545715333, 0.008923135757446288, 0.018956287384033203, 0.008893440246582032, 0.008922112464904786, 0.008897536277770996, 0.008893440246582032, 0.008903679847717285, 0.008954879760742187, 0.008958975791931152, 0.00899891185760498, 0.008960000038146973, 0.008909824371337891, 0.008938528060913086, 0.008979424476623536, 0.008940544128417969, 0.008896512031555176, 0.008933440208435058, 0.00890873622894287, 0.008930368423461914, 0.008996800422668457, 0.00891596794128418, 0.008881152153015137, 0.008921088218688965, 0.0089169921875, 0.008885248184204102, 0.008909824371337891, 0.009086976051330567, 0.00900812816619873, 0.008913920402526856, 0.008903679847717285, 0.008933440208435058, 0.00892307186126709, 0.00900710391998291, 0.0090316801071167, 0.00894159984588623, 0.008897503852844239, 0.008929280281066895, 0.008889344215393067, 0.008951807975769043, 0.008914943695068359, 0.008902655601501466, 0.008882176399230958, 0.00889139175415039, 0.008911871910095214, 0.009109567642211915, 0.00903059196472168, 0.008918016433715821, 0.008871935844421386, 0.008988672256469727, 0.008914943695068359, 0.008914943695068359, 0.008958975791931152, 0.00891596794128418, 0.009012224197387696, 0.008886272430419923, 0.008911871910095214, 0.00890777587890625, 0.008935423851013183, 0.008905728340148926, 0.008918016433715821, 0.00894156837463379, 0.008905728340148926, 0.00889958381652832, 0.0089303035736084, 0.01902284812927246, 0.008904704093933105, 0.008942591667175292, 0.008911871910095214, 0.008929280281066895, 0.008910847663879394, 0.00889139175415039, 0.00892518424987793, 0.008921088218688965, 0.008976384162902832, 0.008932352066040039, 0.008989695549011231, 0.008977408409118653, 0.008927231788635253, 0.0089303035736084, 0.009013248443603515, 0.008960000038146973, 0.008960000038146973, 0.008910847663879394, 0.008945664405822755, 0.008939519882202148, 0.008961024284362793, 0.008936448097229004, 0.009037823677062988, 0.00920576000213623, 0.009206784248352052, 0.008856575965881347, 0.008847359657287598, 0.008865792274475098, 0.008903679847717285, 0.008907808303833007, 0.008943584442138672, 0.00894156837463379, 0.008938495635986327, 0.008940544128417969, 0.00913920021057129, 0.00919654369354248, 0.009150464057922364, 0.008953856468200684, 0.008921088218688965, 0.00890675163269043, 0.008878080368041993, 0.008942591667175292, 0.008922112464904786, 0.008936448097229004, 0.008946687698364257, 0.00892416000366211, 0.008934399604797364, 0.008950783729553222, 0.008932352066040039, 0.008910847663879394, 0.008928256034851074, 0.008928256034851074, 0.008942591667175292, 0.008947775840759278, 0.008911808013916016, 0.00889958381652832, 0.008952863693237305, 0.008958944320678711, 0.009027615547180175, 0.008932319641113282, 0.008992768287658692, 0.008905728340148926, 0.019042303085327148, 0.008913920402526856, 0.008936448097229004, 0.008945664405822755, 0.008919039726257324, 0.00892518424987793, 0.008938495635986327, 0.008945664405822755, 0.008929280281066895, 0.008956928253173829, 0.008939519882202148, 0.008902655601501466, 0.009009152412414552, 0.00902143955230713, 0.008921088218688965, 0.008969216346740723, 0.008935423851013183, 0.008927231788635253, 0.008947711944580078, 0.009337856292724609, 0.00920473575592041, 0.009240575790405273, 0.009218048095703125, 0.009151488304138184, 0.009164799690246582, 0.009184255599975585, 0.009158656120300293, 0.009134079933166504, 0.009259008407592773, 0.009044992446899413, 0.008912896156311035, 0.00890880012512207, 0.009035776138305664, 0.008896512031555176, 0.00892416000366211, 0.008927231788635253, 0.00891596794128418, 0.008960000038146973, 0.009072640419006347, 0.008947711944580078, 0.00899891185760498, 0.00891596794128418, 0.008902655601501466, 0.0088985595703125, 0.008950783729553222, 0.00894156837463379, 0.008910847663879394, 0.008940544128417969, 0.008980480194091797, 0.00894976043701172, 0.008930335998535157, 0.008893407821655274, 0.008950783729553222, 0.008961024284362793, 0.008914943695068359, 0.00892518424987793, 0.008901632308959961, 0.008912896156311035, 0.008920063972473144, 0.009104384422302245, 0.009118720054626465, 0.0088985595703125, 0.00894976043701172, 0.018981887817382814, 0.009211903572082519, 0.009185279846191406, 0.008970239639282226, 0.00892518424987793, 0.008911871910095214, 0.008910847663879394, 0.009165823936462402, 0.009174015998840332, 0.009192447662353515, 0.00890777587890625, 0.00892416000366211, 0.008922112464904786, 0.008939519882202148, 0.008955904006958008, 0.008952832221984864, 0.008929280281066895, 0.008896512031555176, 0.00890777587890625, 0.008917056083679199, 0.008950752258300781, 0.008923104286193848, 0.008928256034851074, 0.00889241600036621, 0.008886272430419923, 0.008954943656921387, 0.008918975830078125, 0.00889241600036621, 0.008905728340148926, 0.008920063972473144, 0.008942591667175292, 0.008954879760742187, 0.009131072044372558, 0.009196479797363282, 0.009163776397705077, 0.009189375877380371, 0.009155584335327148, 0.009184255599975585, 0.009208831787109375, 0.00913920021057129, 0.009218048095703125, 0.009137151718139648, 0.009404416084289552, 0.00894156837463379, 0.009177087783813476, 0.009005056381225587, 0.008921088218688965, 0.008913920402526856, 0.00890777587890625, 0.008905728340148926, 0.00892416000366211, 0.009224191665649414, 0.009186304092407227, 0.009162752151489258, 0.009161727905273438, 0.009209856033325196, 0.00909721565246582, 0.009022463798522949, 0.00897433567047119, 0.00890777587890625, 0.008904704093933105, 0.008936448097229004, 0.009060352325439454]",tokens/s,108.56193334909362,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-6.7b,,cuda,0,42,,,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,217063f5c507ed7cc255df7e1f64c4333a0b4dfe,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,,MB,1301.676032,15125.184512,0.0,14478.737408,14237.625344,s,10,0.8224414138793944,0.08224414138793945,0.00023158737850503415,0.08215292739868164,0.08237123489379883,0.08264656257629394,0.08286682472229004,"[0.08292189025878906, 0.0821987533569336, 0.08213279724121093, 0.08231005096435547, 0.0821382064819336, 0.08213455963134765, 0.08213699340820313, 0.08216111755371094, 0.08214473724365234, 0.08216230773925781]",tokens/s,3112.683720442374,kWh,9.732241499727975e-07,5.332684545643789e-07,5.104102443934566e-06,6.610595048471742e-06,tokens/kWh,38725712.000644006,MB,1301.676032,15125.184512,0.0,14478.737408,14440.977408,s,10,21.55573461914063,2.155573461914062,0.003128784082911401,2.1546517333984374,2.1563899902343753,2.160598193359375,2.163964755859375,"[2.164806396484375, 2.155454833984375, 2.155191162109375, 2.154422119140625, 2.153969482421875, 2.15465673828125, 2.155087646484375, 2.153739501953125, 2.154646728515625, 2.153760009765625]",tokens/s,29.22656133651717,kWh,2.539409045693219e-05,1.3916760766133236e-05,0.00011640616143086489,0.00015571701265393033,tokens/kWh,404580.0707724396,,s,629,22.260371433258044,0.03539009766813682,0.009800577973651038,0.03420876693725586,0.034320641326904294,0.034373221588134766,0.11664900054931641,"[0.034751487731933595, 0.03479142379760742, 0.03473206329345703, 0.03477910232543945, 0.03483855819702148, 0.03484259033203125, 0.03480575942993164, 0.034825214385986326, 0.03483443069458008, 0.034772991180419925, 0.034618366241455076, 0.0346060791015625, 0.03485184097290039, 0.0344453125, 0.03424870300292969, 0.03413094329833984, 0.034099262237548826, 0.03411040115356445, 0.03410227203369141, 0.03416883087158203, 0.03420569610595703, 0.034121726989746096, 0.03416064071655273, 0.03413094329833984, 0.03420467376708984, 0.034146305084228515, 0.0341923828125, 0.03416883087158203, 0.03421184158325195, 0.034146305084228515, 0.03413094329833984, 0.03419750213623047, 0.034179073333740234, 0.03422208023071289, 0.03421699142456055, 0.03435107040405273, 0.034342910766601564, 0.03427635192871094, 0.03420876693725586, 0.03420159912109375, 0.034253822326660154, 0.034236415863037106, 0.034260990142822266, 0.03425791931152344, 0.03429171371459961, 0.0342476806640625, 0.034293758392333985, 0.03426201629638672, 0.03425484848022461, 0.03429683303833008, 0.03427635192871094, 0.03431731033325195, 0.03426508712768555, 0.03428044891357422, 0.03424563217163086, 0.03429683303833008, 0.03427532958984375, 0.034283519744873044, 0.03433881759643555, 0.03430297470092773, 0.034367488861083983, 0.03433574295043945, 0.11684864044189452, 0.03409107208251953, 0.03410835266113281, 0.0340684814453125, 0.034070526123046875, 0.034078720092773435, 0.034105342864990236, 0.03407564926147461, 0.03406438446044922, 0.0340582389831543, 0.03409305572509766, 0.0340684814453125, 0.034132991790771484, 0.03428966522216797, 0.03419340896606445, 0.0341319694519043, 0.0341104621887207, 0.03411251068115234, 0.034149375915527344, 0.03407974243164062, 0.03413094329833984, 0.03412684631347656, 0.03414527893066406, 0.034127872467041014, 0.034127872467041014, 0.034187263488769534, 0.03425484848022461, 0.03421286392211914, 0.03422822570800781, 0.03421184158325195, 0.03422316741943359, 0.03420358276367187, 0.03418931198120117, 0.03420467376708984, 0.034200576782226565, 0.03420774459838867, 0.03424051284790039, 0.03422003173828125, 0.03422003173828125, 0.03421798324584961, 0.03425177764892578, 0.034206718444824216, 0.034233345031738284, 0.03421286392211914, 0.034255870819091795, 0.03424460983276367, 0.03428147125244141, 0.03421286392211914, 0.03427123260498047, 0.03437977600097656, 0.034531326293945314, 0.03434188842773438, 0.03427942276000977, 0.03427942276000977, 0.03428659057617187, 0.03436851119995117, 0.034372608184814454, 0.03430809783935547, 0.03434905624389648, 0.034283519744873044, 0.034304000854492187, 0.03430096054077148, 0.03432956695556641, 0.11668479919433594, 0.03402035140991211, 0.034113536834716796, 0.034344959259033206, 0.034233345031738284, 0.034086910247802735, 0.03409100723266602, 0.034138111114501955, 0.03405721664428711, 0.03408281707763672, 0.03409622573852539, 0.034100128173828126, 0.0341104621887207, 0.034124801635742184, 0.03410227203369141, 0.03410227203369141, 0.034141185760498044, 0.03410739135742188, 0.034113536834716796, 0.03413401412963867, 0.034118656158447266, 0.03419136047363281, 0.034132991790771484, 0.034165760040283204, 0.03415654373168946, 0.034154495239257815, 0.034165760040283204, 0.03419033432006836, 0.034149375915527344, 0.03422003173828125, 0.03416678237915039, 0.034223102569580076, 0.03419852828979492, 0.03422822570800781, 0.03417292785644531, 0.03422208023071289, 0.03421184158325195, 0.034285568237304685, 0.034301952362060545, 0.03424051284790039, 0.03424460983276367, 0.03424256134033203, 0.03421388626098633, 0.03427532958984375, 0.03423436737060547, 0.03428966522216797, 0.03422003173828125, 0.03429785537719727, 0.034320384979248046, 0.03427328109741211, 0.03424665451049805, 0.034304000854492187, 0.034298881530761716, 0.03428044891357422, 0.03427840042114258, 0.0342999038696289, 0.03428761672973633, 0.03429280090332031, 0.03427731323242188, 0.034367488861083983, 0.03428659057617187, 0.03435007858276367, 0.034288639068603514, 0.11707904052734375, 0.03401529693603516, 0.034060222625732425, 0.03401113510131836, 0.03401932907104492, 0.03405619049072266, 0.03409100723266602, 0.03403059387207031, 0.034088958740234376, 0.034062335968017575, 0.034081790924072264, 0.034080768585205076, 0.034136062622070314, 0.03410432052612305, 0.03412070465087891, 0.03410124969482422, 0.03408588790893555, 0.03412377548217774, 0.03409408187866211, 0.03404800033569336, 0.034121726989746096, 0.034119678497314454, 0.034100223541259765, 0.0341739501953125, 0.03414527893066406, 0.03419443130493164, 0.034187263488769534, 0.03416371154785156, 0.03420876693725586, 0.034184192657470705, 0.03418316650390625, 0.03414220809936523, 0.03417292785644531, 0.034206718444824216, 0.0342108154296875, 0.0342108154296875, 0.034200576782226565, 0.03419551849365234, 0.03425273513793945, 0.03422617721557617, 0.03421388626098633, 0.03424051284790039, 0.03423436737060547, 0.03421798324584961, 0.03424256134033203, 0.03423846435546875, 0.034216960906982424, 0.034301952362060545, 0.034272254943847655, 0.03424256134033203, 0.03426303863525391, 0.034298881530761716, 0.03427123260498047, 0.034285568237304685, 0.03439923095703125, 0.034536449432373044, 0.03436236953735351, 0.03432755279541016, 0.03426201629638672, 0.034339839935302735, 0.03430604934692383, 0.034272254943847655, 0.034315265655517575, 0.11667865753173828, 0.034067455291748046, 0.034028545379638675, 0.03402956771850586, 0.0340469741821289, 0.03402751922607422, 0.03409408187866211, 0.03408998489379883, 0.03406643295288086, 0.034114593505859374, 0.03410940933227539, 0.034062335968017575, 0.0341104621887207, 0.03403776168823242, 0.034121726989746096, 0.034049022674560545, 0.034111488342285154, 0.03409305572509766, 0.03410739135742188, 0.03408793640136719, 0.034105342864990236, 0.03410739135742188, 0.034118656158447266, 0.03410943984985351, 0.03415039825439453, 0.034121726989746096, 0.034165760040283204, 0.03410851287841797, 0.034177982330322265, 0.03413808059692383, 0.03417190551757812, 0.03414425659179687, 0.034184192657470705, 0.034154495239257815, 0.034260990142822266, 0.03418009567260742, 0.03423436737060547, 0.03415859222412109, 0.03424870300292969, 0.03418214416503906, 0.03425689697265625, 0.034202625274658206, 0.03434393692016602, 0.03455692672729492, 0.034293758392333985, 0.03424153518676758, 0.03426201629638672, 0.03425996780395508, 0.034260990142822266, 0.03426713562011719, 0.03428147125244141, 0.03424358367919922, 0.034344959259033206, 0.03424665451049805, 0.03430809783935547, 0.03425177764892578, 0.03431628799438476, 0.03433065414428711, 0.03431216049194336, 0.03428761672973633, 0.03431321716308594, 0.03427532958984375, 0.03433779144287109, 0.11667250823974609, 0.03399168014526367, 0.03406335830688476, 0.03410739135742188, 0.034149375915527344, 0.03417190551757812, 0.034105342864990236, 0.034078720092773435, 0.034080768585205076, 0.03407564926147461, 0.03413401412963867, 0.034044929504394535, 0.03409612655639648, 0.03410124969482422, 0.034119678497314454, 0.03410432052612305, 0.03413094329833984, 0.0341473274230957, 0.034095104217529294, 0.034136062622070314, 0.03410739135742188, 0.034116607666015625, 0.03413913726806641, 0.034132991790771484, 0.03415859222412109, 0.03416883087158203, 0.03421593475341797, 0.03419033432006836, 0.034233345031738284, 0.0341923828125, 0.03420569610595703, 0.03415961456298828, 0.03422719955444336, 0.03420774459838867, 0.034200576782226565, 0.03416678237915039, 0.03421286392211914, 0.03421491241455078, 0.034229248046875, 0.03426713562011719, 0.034219009399414066, 0.03421798324584961, 0.03425689697265625, 0.03422003173828125, 0.03423440170288086, 0.03426710510253906, 0.034331649780273435, 0.03427635192871094, 0.03429177474975586, 0.034270145416259765, 0.034255870819091795, 0.03422617721557617, 0.03427532958984375, 0.03428761672973633, 0.03429683303833008, 0.034256961822509764, 0.03433465576171875, 0.03434188842773438, 0.03427532958984375, 0.034334720611572264, 0.034374656677246096, 0.034285568237304685, 0.03434905624389648, 0.1166223373413086, 0.034045951843261715, 0.034049022674560545, 0.034088958740234376, 0.03411763381958008, 0.034146305084228515, 0.03407564926147461, 0.0341288948059082, 0.03422412872314453, 0.03416166305541992, 0.034083839416503905, 0.034167808532714845, 0.034098175048828124, 0.034108417510986325, 0.034136062622070314, 0.034252799987792966, 0.03415654373168946, 0.03413708877563477, 0.03408793640136719, 0.034118656158447266, 0.034121726989746096, 0.034132991790771484, 0.03410739135742188, 0.034141185760498044, 0.03416678237915039, 0.03412275314331055, 0.034157569885253904, 0.03419340896606445, 0.03417705535888672, 0.03416572952270508, 0.03416678237915039, 0.03415961456298828, 0.03418316650390625, 0.03429177474975586, 0.03432236862182617, 0.03438284683227539, 0.034229248046875, 0.03425689697265625, 0.034233345031738284, 0.034239486694335936, 0.03424870300292969, 0.03423846435546875, 0.03420876693725586, 0.03423436737060547, 0.03424460983276367, 0.03426611328125, 0.034236415863037106, 0.034237438201904294, 0.03427635192871094, 0.034290687561035156, 0.03423955154418945, 0.034319297790527344, 0.03427942276000977, 0.03432243347167969, 0.034260990142822266, 0.03427945709228516, 0.034317279815673826, 0.034282497406005856, 0.034301952362060545, 0.034298881530761716, 0.034298881530761716, 0.034304000854492187, 0.034310142517089845, 0.11665100860595704, 0.03404800033569336, 0.03415552139282227, 0.034195457458496094, 0.03405926513671875, 0.03407974243164062, 0.034050048828125, 0.03405926513671875, 0.03406643295288086, 0.03404390335083008, 0.034044929504394535, 0.03410636901855469, 0.034098175048828124, 0.03405414581298828, 0.03412384033203125, 0.034119617462158205, 0.034121726989746096, 0.03410432052612305, 0.034098175048828124, 0.03407974243164062, 0.03412684631347656, 0.034095104217529294, 0.03411455917358398, 0.03413401412963867, 0.034165760040283204, 0.03410636901855469, 0.034179073333740234, 0.034148353576660156, 0.034132991790771484, 0.03417702484130859, 0.034200576782226565, 0.034233345031738284, 0.03420979309082031, 0.03422719955444336, 0.03417599868774414, 0.03421798324584961, 0.034170879364013675, 0.03416678237915039, 0.034187297821044925, 0.034245601654052736, 0.03422515106201172, 0.03424153518676758, 0.03425689697265625, 0.0342108154296875, 0.03424153518676758, 0.034239486694335936, 0.03422515106201172, 0.03426406478881836, 0.03427532958984375, 0.03425075149536133, 0.03427635192871094, 0.03428659057617187, 0.03424358367919922, 0.03429683303833008, 0.0342476806640625, 0.034275360107421875, 0.03432444763183594, 0.034305023193359374, 0.034290687561035156, 0.03428764724731445, 0.03430806350708008, 0.03426713562011719, 0.034321407318115234, 0.11664383697509766, 0.03402956771850586, 0.03400396728515625, 0.03403776168823242, 0.03405926513671875, 0.03405516815185547, 0.03407257461547852, 0.03407462310791016, 0.03410227203369141, 0.03407155227661133, 0.034162689208984375, 0.03407974243164062, 0.03409305572509766, 0.03406950378417969, 0.03409612655639648, 0.03406438446044922, 0.03415244674682617, 0.034105342864990236, 0.03418828964233398, 0.0340582389831543, 0.03437363052368164, 0.03469219207763672, 0.03427318572998047, 0.034111488342285154, 0.034200576782226565, 0.03415039825439453, 0.03415244674682617, 0.03416883087158203, 0.034195457458496094, 0.034154495239257815, 0.034219009399414066, 0.03416371154785156, 0.034200576782226565, 0.03416678237915039, 0.034203647613525394, 0.03419443130493164, 0.03426508712768555, 0.03418828964233398, 0.03423539352416992, 0.03419443130493164, 0.03427635192871094, 0.03419340896606445, 0.03424460983276367, 0.03423846435546875, 0.034255870819091795, 0.03425996780395508, 0.03426003265380859, 0.034242496490478516, 0.03429478454589844, 0.0342476806640625, 0.0342927360534668, 0.03421184158325195, 0.034275390625, 0.03422304153442383, 0.03430092620849609, 0.03425996780395508, 0.034283519744873044, 0.03424870300292969, 0.03431935882568359, 0.03425791931152344, 0.03431423950195313, 0.034282497406005856, 0.03431219100952149, 0.11666841888427734, 0.034004032135009764, 0.03404281616210938, 0.034045951843261715, 0.034044929504394535, 0.03406854248046875, 0.03409196853637695, 0.03405311965942383, 0.034095104217529294, 0.034051071166992186, 0.034113536834716796, 0.034032638549804685, 0.0341104621887207, 0.034086910247802735, 0.03411558532714844, 0.0340469741821289, 0.034138111114501955, 0.034141185760498044, 0.0340766716003418, 0.03410432052612305, 0.03411455917358398, 0.03409100723266602, 0.034148353576660156, 0.034141185760498044, 0.034136062622070314, 0.034127872467041014, 0.03415654373168946, 0.03419852828979492, 0.03413708877563477, 0.03415552139282227, 0.03416371154785156, 0.03413913726806641, 0.034203647613525394, 0.03420876693725586, 0.03427328109741211, 0.03417292785644531, 0.034223102569580076, 0.034253822326660154, 0.03431628799438476, 0.03421798324584961, 0.03423539352416992, 0.03420979309082031, 0.03422003173828125, 0.034233345031738284, 0.03423846435546875, 0.03421798324584961, 0.034277374267578126, 0.03427328109741211, 0.03425996780395508, 0.03428761672973633, 0.03426611328125, 0.03426303863525391, 0.034269184112548826, 0.034260990142822266, 0.03430297470092773, 0.03426406478881836, 0.03432044982910156, 0.034289600372314454, 0.034285568237304685, 0.03431219100952149, 0.03433881759643555, 0.034277374267578126, 0.03433779144287109]",tokens/s,28.256491671124767,,,main,False,False,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-13b,facebook/opt-13b,cuda,0,42,,,True,,,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",opt,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-6.7b,,cuda,0,42,,,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,d35829e539df8480b726c647eeabf91e41eae047,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,,MB,1192.951808,14382.792704,0.0,13736.3456,13487.531008,s,10,0.8513474197387696,0.08513474197387695,0.00021429432968233625,0.08507120132446289,0.08527548065185546,0.08551233215332031,0.08570181335449219,"[0.08574918365478515, 0.08507942199707032, 0.08506537628173828, 0.08504169464111327, 0.0849843521118164, 0.084989501953125, 0.08503981018066406, 0.0850770263671875, 0.08522284698486328, 0.08509820556640625]",tokens/s,3006.9980135554056,kWh,1.0067580769403281e-06,5.516551041567917e-07,5.112296933525468e-06,6.670710114622587e-06,tokens/kWh,38376723.85715473,MB,1192.951808,14382.792704,0.0,13736.3456,13661.254656,s,10,19.876877197265628,1.9876877197265628,0.0030016916208529323,1.9866903076171876,1.9884474975585937,1.9924959045410158,1.9957346301269532,"[1.9965443115234376, 1.98677587890625, 1.9857510986328124, 1.9863314208984375, 1.986604736328125, 1.9861082763671876, 1.986561767578125, 1.9872159423828124, 1.9875478515625, 1.9874359130859376]",tokens/s,31.695119597894696,kWh,2.3404805134076856e-05,1.2826450611971834e-05,0.00010996662846287472,0.00014619788420892343,tokens/kWh,430922.7889370145,,s,629,20.61038695144655,0.0327669108926018,0.010142886768973333,0.031545343399047854,0.03165429801940918,0.03173130226135254,0.11686746459960938,"[0.03244851303100586, 0.03200102233886719, 0.032031742095947266, 0.031970304489135744, 0.03204710388183594, 0.03201638412475586, 0.03205120086669922, 0.03203583908081055, 0.0320634880065918, 0.03206047821044922, 0.03205318450927734, 0.03189145660400391, 0.031936511993408204, 0.03190169525146484, 0.03164159965515137, 0.03146956825256347, 0.03146444892883301, 0.031455232620239255, 0.03146342468261719, 0.03145219230651856, 0.03151049613952637, 0.03150643157958984, 0.0314654712677002, 0.03148595237731933, 0.031510528564453126, 0.0315043830871582, 0.031526912689208986, 0.03151155281066895, 0.0315729923248291, 0.03154022407531738, 0.03153510475158691, 0.03153408050537109, 0.03156480026245117, 0.031541248321533204, 0.03159654426574707, 0.03154947280883789, 0.031594463348388675, 0.03157811164855957, 0.03160272026062012, 0.03156374359130859, 0.031591424942016604, 0.03159654426574707, 0.03158425521850586, 0.03158732795715332, 0.03161497688293457, 0.03161907196044922, 0.03164262390136719, 0.03162623977661133, 0.03163039970397949, 0.03161087989807129, 0.03163030433654785, 0.031619104385375976, 0.03164255905151367, 0.03164057540893555, 0.03165593528747559, 0.03161292839050293, 0.03168364715576172, 0.031652799606323244, 0.03166924858093262, 0.031681535720825195, 0.031661056518554685, 0.03163545608520508, 0.11689984130859375, 0.03139583969116211, 0.03134576034545898, 0.03139676856994629, 0.031409151077270506, 0.03140403175354004, 0.03139379119873047, 0.03142758369445801, 0.03139993667602539, 0.03140505599975586, 0.031406080245971676, 0.03146956825256347, 0.03141119956970215, 0.03147776031494141, 0.031455232620239255, 0.03143680000305176, 0.031425535202026365, 0.03143270492553711, 0.031438848495483396, 0.03144704055786133, 0.031459327697753905, 0.031457279205322264, 0.03146444892883301, 0.031493120193481446, 0.031442943572998046, 0.03144806480407715, 0.031473663330078124, 0.0315167350769043, 0.03149100875854492, 0.03152076721191406, 0.03173785591125488, 0.03190169525146484, 0.0316682243347168, 0.03155148887634277, 0.03151974487304687, 0.03153510475158691, 0.031541248321533204, 0.03157401657104492, 0.031635488510131836, 0.03158012771606445, 0.03154944038391113, 0.03155353546142578, 0.03157094383239746, 0.03155558395385742, 0.03154841613769531, 0.03160063934326172, 0.03159449577331543, 0.03163443183898926, 0.03160985565185547, 0.031632383346557616, 0.03158527946472168, 0.03158425521850586, 0.03159347152709961, 0.031640640258789064, 0.031636415481567384, 0.03162521553039551, 0.03160063934326172, 0.03164672088623047, 0.03160479927062988, 0.03164358329772949, 0.031627264022827145, 0.03163340759277344, 0.03165388870239258, 0.11698585510253906, 0.03139174461364746, 0.03136511993408203, 0.03137740707397461, 0.031357952117919925, 0.03141427230834961, 0.03139993667602539, 0.03142348861694336, 0.03139379119873047, 0.031409151077270506, 0.03139891242980957, 0.0314337272644043, 0.03140812873840332, 0.03145113563537598, 0.03147980880737305, 0.0314521598815918, 0.03143276786804199, 0.03142444801330566, 0.031456256866455076, 0.03143577575683594, 0.03146240043640137, 0.031503360748291014, 0.031440895080566404, 0.03144908714294434, 0.031453184127807614, 0.03147980880737305, 0.0314839038848877, 0.03152179145812988, 0.031488000869750975, 0.0315156478881836, 0.031494144439697266, 0.031513599395751955, 0.03150643157958984, 0.031514623641967776, 0.03149516868591309, 0.03156070327758789, 0.03151155281066895, 0.03156073570251465, 0.031523807525634766, 0.03158323287963867, 0.03160371208190918, 0.031547391891479495, 0.0315729923248291, 0.03155763244628906, 0.03159449577331543, 0.0315863037109375, 0.03159859275817871, 0.031632383346557616, 0.03156172752380371, 0.03159552001953125, 0.03159859275817871, 0.03160063934326172, 0.03160678482055664, 0.03161087989807129, 0.03164672088623047, 0.031692800521850584, 0.03163750457763672, 0.03163955116271973, 0.03158937644958496, 0.03161087989807129, 0.03162521553039551, 0.03162521553039551, 0.03162524795532227, 0.1168803482055664, 0.03137638473510742, 0.03139795112609863, 0.031378368377685546, 0.03136102485656738, 0.03139993667602539, 0.03140403175354004, 0.03141119956970215, 0.03143065643310547, 0.031460351943969726, 0.031437824249267575, 0.03142758369445801, 0.031440895080566404, 0.0314204158782959, 0.03138662338256836, 0.03142860794067383, 0.03145011138916016, 0.031476736068725586, 0.03146240043640137, 0.03148595237731933, 0.031454208374023435, 0.031457279205322264, 0.031470592498779294, 0.03143475151062012, 0.03143577575683594, 0.031514623641967776, 0.03149004745483398, 0.03149619293212891, 0.03148185539245606, 0.031498239517211916, 0.03148902320861816, 0.031514656066894534, 0.031561695098876956, 0.03159040069580078, 0.031514623641967776, 0.03158425521850586, 0.031535167694091794, 0.03154118347167969, 0.03156172752380371, 0.03156172752380371, 0.031604736328125, 0.031558656692504884, 0.03156991958618164, 0.03159449577331543, 0.03158118438720703, 0.031644672393798826, 0.03164672088623047, 0.03157606315612793, 0.03158323287963867, 0.03160371208190918, 0.03162931251525879, 0.031615999221801756, 0.03163545608520508, 0.03162009620666504, 0.03165081596374512, 0.03165184020996094, 0.03163852882385254, 0.03162623977661133, 0.03167849540710449, 0.03163644790649414, 0.03162931251525879, 0.031677440643310545, 0.031631359100341795, 0.11683433532714843, 0.031388639450073245, 0.03139481544494629, 0.03139072036743164, 0.03138150405883789, 0.03142451286315918, 0.031543296813964845, 0.03148595237731933, 0.03141734313964844, 0.03141836738586426, 0.03145113563537598, 0.031440895080566404, 0.031389696121215824, 0.03143475151062012, 0.03141836738586426, 0.03141119956970215, 0.03143168067932129, 0.03145113563537598, 0.031456256866455076, 0.031426559448242186, 0.03143577575683594, 0.031456256866455076, 0.03151974487304687, 0.031444000244140624, 0.03147977638244629, 0.031486976623535154, 0.031541248321533204, 0.031475711822509765, 0.0315156478881836, 0.03160063934326172, 0.03160166358947754, 0.031525888442993165, 0.03153510475158691, 0.031541248321533204, 0.03153305625915527, 0.03153919982910156, 0.03152383995056152, 0.03160985565185547, 0.03159040069580078, 0.031541248321533204, 0.031546367645263675, 0.03157708740234375, 0.03153715133666992, 0.0315729923248291, 0.03159654426574707, 0.03159145545959473, 0.03159651184082031, 0.03160272026062012, 0.03160367965698242, 0.0316231689453125, 0.03159040069580078, 0.031631359100341795, 0.0316180477142334, 0.03165081596374512, 0.031661056518554685, 0.03163340759277344, 0.03160576057434082, 0.03165388870239258, 0.0315996150970459, 0.03161497688293457, 0.03169183921813965, 0.03169683265686035, 0.031659008026123044, 0.11681996917724609, 0.03136511993408203, 0.03139686393737793, 0.03136204719543457, 0.03136307144165039, 0.03138047981262207, 0.03142451286315918, 0.031425535202026365, 0.031389696121215824, 0.03141939163208008, 0.03141734313964844, 0.03141734313964844, 0.03140300750732422, 0.03141119956970215, 0.03140812873840332, 0.031474687576293944, 0.031454208374023435, 0.0314839038848877, 0.031454208374023435, 0.03144908714294434, 0.03147878456115723, 0.031508480072021484, 0.0314654712677002, 0.03150233650207519, 0.031453184127807614, 0.03155148887634277, 0.0314747200012207, 0.031484895706176755, 0.03148185539245606, 0.03151155281066895, 0.0314901123046875, 0.03150534439086914, 0.031579135894775394, 0.03163545608520508, 0.031528959274291994, 0.03157094383239746, 0.0315545597076416, 0.03156991958618164, 0.03156889533996582, 0.03156787109375, 0.03158937644958496, 0.03155251121520996, 0.031542272567749025, 0.031579135894775394, 0.031536191940307615, 0.031571903228759766, 0.03160063934326172, 0.03158835220336914, 0.03158118438720703, 0.031604736328125, 0.031615999221801756, 0.03164057540893555, 0.03163955116271973, 0.03162112045288086, 0.03161292839050293, 0.03164057540893555, 0.031615999221801756, 0.031611904144287106, 0.031643648147583005, 0.03161907196044922, 0.03162214469909668, 0.03164262390136719, 0.03165798377990723, 0.11688038635253906, 0.03139072036743164, 0.03137542343139649, 0.031378368377685546, 0.03137740707397461, 0.03141427230834961, 0.03137433624267578, 0.0314337272644043, 0.03139993667602539, 0.03142451286315918, 0.03138051223754883, 0.03139068794250488, 0.03142348861694336, 0.03143987274169922, 0.031455232620239255, 0.031421440124511715, 0.03142758369445801, 0.03141836738586426, 0.03141529655456543, 0.03164163208007813, 0.031812576293945315, 0.0315729923248291, 0.03155763244628906, 0.031497215270996096, 0.03148492813110351, 0.031458303451538085, 0.031470592498779294, 0.03148287963867188, 0.031470592498779294, 0.03151974487304687, 0.0315545597076416, 0.031494144439697266, 0.031525888442993165, 0.03155353546142578, 0.031527936935424807, 0.031531007766723636, 0.03154022407531738, 0.03156684875488281, 0.03155660820007324, 0.03156377601623535, 0.031541248321533204, 0.031559680938720705, 0.031545343399047854, 0.03155353546142578, 0.031559680938720705, 0.031631359100341795, 0.0316364803314209, 0.03159449577331543, 0.03161292839050293, 0.0315996150970459, 0.03161702346801758, 0.03161907196044922, 0.03161708831787109, 0.03163027191162109, 0.0316231689453125, 0.031632383346557616, 0.03162931251525879, 0.0316364803314209, 0.03160576057434082, 0.03163852882385254, 0.03160576057434082, 0.031673408508300784, 0.03167328071594238, 0.11695104217529297, 0.03136614418029785, 0.03140812873840332, 0.031341567993164066, 0.03134464073181152, 0.03138662338256836, 0.03139583969116211, 0.03141427230834961, 0.03141839981079102, 0.031402975082397463, 0.03142758369445801, 0.03140403175354004, 0.03143270492553711, 0.03143680000305176, 0.03141222381591797, 0.0314726390838623, 0.031498239517211916, 0.03145939254760742, 0.03145414352416992, 0.03146342468261719, 0.031473663330078124, 0.03148185539245606, 0.03149926376342774, 0.031480831146240236, 0.03144908714294434, 0.03149004745483398, 0.03147878456115723, 0.031512575149536134, 0.03147878456115723, 0.031486976623535154, 0.03150035285949707, 0.031532991409301755, 0.03156070327758789, 0.03155046463012695, 0.03151667213439941, 0.031545343399047854, 0.031528959274291994, 0.03162931251525879, 0.03155660820007324, 0.031591424942016604, 0.03156793594360351, 0.03155756759643555, 0.03153305625915527, 0.031719423294067385, 0.031751167297363284, 0.03162521553039551, 0.031562751770019534, 0.03163443183898926, 0.03162214469909668, 0.03162931251525879, 0.03162931251525879, 0.03163340759277344, 0.03163033676147461, 0.031662080764770506, 0.03163443183898926, 0.031635488510131836, 0.031657951354980465, 0.03162521553039551, 0.031645696640014646, 0.03162623977661133, 0.031665151596069335, 0.03172147178649903, 0.031932416915893554, 0.11726643371582031, 0.03140096092224121, 0.0313753604888916, 0.03136511993408203, 0.031369216918945314, 0.03141119956970215, 0.03139174461364746, 0.03143680000305176, 0.03141119956970215, 0.03142860794067383, 0.03138355255126953, 0.03143270492553711, 0.031388671875, 0.03146444892883301, 0.03147980880737305, 0.03142860794067383, 0.03144806480407715, 0.03144704055786133, 0.03144704055786133, 0.031438911437988284, 0.03145414352416992, 0.03148902320861816, 0.031471616744995115, 0.03142963218688965, 0.03143577575683594, 0.031475711822509765, 0.0314654712677002, 0.03149926376342774, 0.031473663330078124, 0.03168460845947266, 0.0315863037109375, 0.03153510475158691, 0.03156380844116211, 0.031540191650390625, 0.031558656692504884, 0.031547391891479495, 0.031525888442993165, 0.03157196807861328, 0.03155660820007324, 0.03160063934326172, 0.03164777565002441, 0.031593439102172854, 0.031543296813964845, 0.03155558395385742, 0.03156172752380371, 0.03161702346801758, 0.031600671768188475, 0.03158320045471191, 0.0316180477142334, 0.03159756851196289, 0.03162214469909668, 0.03169385528564453, 0.031658975601196286, 0.031666175842285156, 0.03163852882385254, 0.031644672393798826, 0.03165184020996094, 0.031628288269042966, 0.03160780715942383, 0.031698944091796875, 0.03174710464477539, 0.032173023223876954, 0.031937536239624024, 0.11699404907226563, 0.03139276885986328, 0.03143168067932129, 0.031413280487060546, 0.031399904251098634, 0.0313948802947998, 0.031426496505737304, 0.031406080245971676, 0.03138252830505371, 0.03141119956970215, 0.031442943572998046, 0.03140300750732422, 0.031409151077270506, 0.031412256240844726, 0.03142857551574707, 0.03144908714294434, 0.031456256866455076, 0.031719423294067385, 0.03181260871887207, 0.03155558395385742, 0.03146854400634766, 0.031473663330078124, 0.0314839038848877, 0.03146240043640137, 0.03145113563537598, 0.03151155281066895, 0.03166924858093262, 0.03152383995056152, 0.031513599395751955, 0.031493120193481446, 0.031486976623535154, 0.031525888442993165, 0.03151974487304687, 0.03158937644958496, 0.0315361270904541, 0.03159040069580078, 0.031562751770019534, 0.03158732795715332, 0.03156377601623535, 0.03157196807861328, 0.03156377601623535, 0.031562751770019534, 0.0315863037109375, 0.0316180477142334, 0.03159347152709961, 0.0316231689453125, 0.03157708740234375, 0.031579135894775394, 0.03163443183898926, 0.03165593528747559, 0.03164057540893555, 0.03163545608520508, 0.03159043121337891, 0.03164463996887207, 0.03160371208190918, 0.031665151596069335, 0.031628288269042966, 0.031643648147583005, 0.031678464889526366, 0.03165388870239258, 0.03162419128417969, 0.031648767471313476, 0.03162419128417969]",tokens/s,30.518592469019794,,,main,False,False,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,microsoft/rho-math-1b-v0.1,microsoft/rho-math-1b-v0.1,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1249.23904,3041.394688,0.0,2394.947584,2279.417856,s,10,0.23634982109069824,0.023634982109069825,0.0011165227742452248,0.02325393581390381,0.024405410957336424,0.025520129299163814,0.026411903972625734,"[0.02663484764099121, 0.022848447799682616, 0.02291916847229004, 0.023588703155517577, 0.02277120018005371, 0.02387081527709961, 0.02384486389160156, 0.022809791564941406, 0.022904287338256835, 0.024157695770263672]",tokens/s,10831.402317912527,kWh,2.7556789257896285e-07,1.5098250052967787e-07,1.0927326794172656e-06,1.5192830725259062e-06,tokens/kWh,168500528.0644531,MB,1249.23904,3041.394688,0.0,2394.947584,2279.420416,s,10,14.361950073242188,1.4361950073242187,0.016421711028924653,1.44034375,1.456474658203125,1.4572763305664063,1.4579176684570314,"[1.4580780029296876, 1.4076146240234375, 1.4368963623046875, 1.4487359619140625, 1.4437911376953125, 1.4562965087890625, 1.42836181640625, 1.4149393310546876, 1.42232568359375, 1.44491064453125]",tokens/s,43.86590934985603,kWh,1.6920164502275267e-05,9.272085132958566e-06,3.6294880563983074e-05,6.248713019921691e-05,tokens/kWh,1008207.6068967801,,s,629,14.545633255004864,0.02312501312401412,0.002848130619724419,0.022890495300292968,0.02322396125793457,0.02381926383972168,0.04538929153442383,"[0.024168447494506837, 0.023833599090576172, 0.024000511169433594, 0.024071199417114258, 0.024100831985473633, 0.023974912643432617, 0.023829504013061522, 0.023569408416748046, 0.024196096420288086, 0.02433843231201172, 0.02389094352722168, 0.024004608154296874, 0.023888896942138672, 0.02342092704772949, 0.024335359573364256, 0.024337408065795898, 0.024024063110351563, 0.023837696075439452, 0.023723007202148438, 0.023476224899291992, 0.023793664932250977, 0.02409164810180664, 0.02411827278137207, 0.023204864501953124, 0.023071840286254884, 0.022964128494262694, 0.02274508857727051, 0.02290790367126465, 0.022785024642944338, 0.022833152770996092, 0.022840320587158205, 0.022329343795776366, 0.02259868812561035, 0.02262118339538574, 0.0225515193939209, 0.022330368041992187, 0.022416383743286132, 0.022361087799072265, 0.022545408248901368, 0.022195199966430663, 0.02283417510986328, 0.0224399356842041, 0.022724607467651366, 0.022737920761108397, 0.022588415145874022, 0.022801408767700194, 0.022829055786132812, 0.02285875129699707, 0.02249932861328125, 0.02245427131652832, 0.022384639739990234, 0.0233123836517334, 0.02289254379272461, 0.022740991592407226, 0.02285158348083496, 0.02206515121459961, 0.02205388832092285, 0.02208358383178711, 0.02208665657043457, 0.022124544143676757, 0.02206719970703125, 0.021986303329467775, 0.04530688095092773, 0.022137855529785155, 0.023468032836914062, 0.022943744659423827, 0.022657024383544923, 0.022776832580566408, 0.02229145622253418, 0.021932031631469725, 0.02179596710205078, 0.02183359909057617, 0.021983327865600585, 0.021810079574584963, 0.021769216537475586, 0.02206924819946289, 0.02227097511291504, 0.02283622360229492, 0.022611967086791994, 0.021876735687255858, 0.022788095474243163, 0.021967872619628907, 0.021715967178344727, 0.021936128616333008, 0.02261299133300781, 0.022573055267333983, 0.022879232406616212, 0.022656000137329102, 0.02188806343078613, 0.022639551162719727, 0.022740991592407226, 0.022524927139282228, 0.022437887191772463, 0.022565887451171874, 0.0224532470703125, 0.02265395164489746, 0.02245734405517578, 0.02249318313598633, 0.022495231628417968, 0.02215936088562012, 0.022544416427612304, 0.022628320693969726, 0.0224849910736084, 0.022598655700683593, 0.022882303237915038, 0.022681600570678712, 0.02271334457397461, 0.022803455352783202, 0.023249919891357423, 0.02208665657043457, 0.02208563232421875, 0.02206003189086914, 0.022147071838378905, 0.02208768081665039, 0.02205900764465332, 0.021980159759521483, 0.02206208038330078, 0.022031423568725585, 0.022041536331176757, 0.02189516830444336, 0.02210201644897461, 0.02206208038330078, 0.022120447158813478, 0.021985279083251954, 0.02210201644897461, 0.04544409561157227, 0.02203647994995117, 0.02207027244567871, 0.021993471145629884, 0.022034431457519533, 0.022012928009033202, 0.022021120071411132, 0.021978111267089845, 0.021995519638061522, 0.022785024642944338, 0.024576000213623047, 0.023803903579711915, 0.02328371238708496, 0.02311884880065918, 0.02310348892211914, 0.023012351989746094, 0.022912031173706056, 0.02313315200805664, 0.0228351993560791, 0.022789119720458984, 0.02302566337585449, 0.023000160217285157, 0.022814624786376952, 0.022785024642944338, 0.023013376235961915, 0.022848575592041016, 0.02274604797363281, 0.023004159927368165, 0.02286899185180664, 0.022907968521118163, 0.022862783432006838, 0.022986751556396484, 0.022915071487426757, 0.022924287796020508, 0.023036928176879884, 0.022841344833374022, 0.022763519287109374, 0.023130111694335938, 0.02269696044921875, 0.022973440170288087, 0.022906879425048828, 0.023001087188720702, 0.022689792633056642, 0.022736928939819337, 0.02296623992919922, 0.02269900894165039, 0.02262937545776367, 0.02269593620300293, 0.02282700729370117, 0.022908992767333984, 0.022900672912597658, 0.02288128089904785, 0.022965248107910157, 0.02289151954650879, 0.02271334457397461, 0.022838432312011717, 0.02296201515197754, 0.022776832580566408, 0.022622207641601562, 0.02282598304748535, 0.023010303497314453, 0.022676479339599608, 0.022632448196411133, 0.04679270553588867, 0.023007232666015624, 0.022987775802612305, 0.022890495300292968, 0.02267136001586914, 0.023059455871582032, 0.023000064849853515, 0.022953983306884765, 0.023043071746826172, 0.02288844871520996, 0.022968320846557616, 0.02306764793395996, 0.02313216018676758, 0.023203840255737306, 0.022889471054077147, 0.022983680725097655, 0.02307174491882324, 0.023069696426391603, 0.02289151954650879, 0.02305536079406738, 0.022995967864990235, 0.02282598304748535, 0.022982656478881838, 0.02287001609802246, 0.022978559494018554, 0.02295910453796387, 0.022866943359375, 0.022900800704956054, 0.02282080078125, 0.02301644706726074, 0.02287615966796875, 0.022994943618774414, 0.023143423080444335, 0.023207935333251953, 0.022983680725097655, 0.023053312301635744, 0.022890495300292968, 0.02309119987487793, 0.02301644706726074, 0.023012351989746094, 0.023003135681152344, 0.02294272041320801, 0.023088127136230468, 0.022969343185424804, 0.022924287796020508, 0.022958080291748048, 0.02285977554321289, 0.023053312301635744, 0.02289151954650879, 0.022972415924072266, 0.022862848281860353, 0.023405567169189453, 0.023079935073852538, 0.023006208419799806, 0.022952959060668944, 0.02321820831298828, 0.02294576072692871, 0.023007232666015624, 0.02286591911315918, 0.023034879684448242, 0.02292736053466797, 0.023045183181762696, 0.02317305564880371, 0.045399040222167966, 0.02202726364135742, 0.02188595199584961, 0.022016000747680665, 0.021954559326171876, 0.021967872619628907, 0.021979167938232423, 0.02193404769897461, 0.021968896865844727, 0.021843967437744142, 0.021977088928222657, 0.0224716796875, 0.02294272041320801, 0.023184383392333984, 0.023021568298339845, 0.02288844871520996, 0.023000064849853515, 0.022980607986450196, 0.02307276725769043, 0.022993919372558593, 0.023187456130981447, 0.023020544052124024, 0.022916095733642578, 0.02289459228515625, 0.02304614448547363, 0.023000064849853515, 0.022955007553100586, 0.022985727310180663, 0.023059455871582032, 0.023012351989746094, 0.023007232666015624, 0.023012351989746094, 0.02308710479736328, 0.023085056304931642, 0.02349056053161621, 0.023397375106811523, 0.02307788848876953, 0.02309222412109375, 0.023023712158203126, 0.023012256622314452, 0.02304102325439453, 0.02304819107055664, 0.02310041618347168, 0.02307072067260742, 0.02311884880065918, 0.02304204750061035, 0.023038976669311522, 0.023027711868286133, 0.023771135330200196, 0.024211456298828125, 0.023362560272216795, 0.02332159996032715, 0.023136255264282226, 0.02310758399963379, 0.02303385543823242, 0.023235584259033205, 0.023372800827026367, 0.023008256912231444, 0.023219200134277345, 0.023198720932006835, 0.023018495559692383, 0.023166015625, 0.023118783950805664, 0.04736102294921875, 0.02308915138244629, 0.022986751556396484, 0.023022592544555662, 0.023036928176879884, 0.02307583999633789, 0.022930431365966796, 0.022977535247802734, 0.023137344360351562, 0.022892511367797852, 0.022995935440063477, 0.023034879684448242, 0.023630847930908205, 0.02322131156921387, 0.02318124771118164, 0.022940671920776368, 0.023382015228271484, 0.02300726318359375, 0.02354991912841797, 0.02329702377319336, 0.023030784606933592, 0.023171072006225587, 0.023145471572875977, 0.02328883171081543, 0.022921215057373046, 0.023037952423095705, 0.023007232666015624, 0.023138303756713868, 0.023030784606933592, 0.022920192718505858, 0.02304921531677246, 0.023076864242553712, 0.022953983306884765, 0.02305740737915039, 0.02303385543823242, 0.023335935592651368, 0.023045120239257814, 0.023117824554443358, 0.023126016616821288, 0.022991872787475585, 0.02303696060180664, 0.023012319564819337, 0.023026687622070312, 0.02305740737915039, 0.023234560012817384, 0.02303385543823242, 0.023044095993041993, 0.02308403205871582, 0.023047168731689452, 0.023010303497314453, 0.023031808853149413, 0.02330521583557129, 0.023013376235961915, 0.022871040344238282, 0.023047168731689452, 0.02301644706726074, 0.023536640167236327, 0.023426111221313477, 0.02383148765563965, 0.02305740737915039, 0.02305740737915039, 0.02307583999633789, 0.023111679077148437, 0.0474152946472168, 0.023185407638549805, 0.022985727310180663, 0.022974464416503908, 0.02311065673828125, 0.023009279251098632, 0.022993919372558593, 0.02307583999633789, 0.022982656478881838, 0.02287820816040039, 0.022958080291748048, 0.023014400482177736, 0.022993919372558593, 0.023219232559204103, 0.02313007926940918, 0.024027135848999022, 0.023109632492065428, 0.022940671920776368, 0.022960128784179686, 0.022971391677856445, 0.022973440170288087, 0.023010303497314453, 0.022939647674560547, 0.02305740737915039, 0.022106111526489256, 0.021953535079956055, 0.022006784439086914, 0.022008832931518556, 0.022041599273681642, 0.023181312561035155, 0.023447551727294923, 0.022946815490722656, 0.021950464248657226, 0.022077503204345702, 0.022082496643066406, 0.0220897274017334, 0.02211942481994629, 0.02209280014038086, 0.022639616012573242, 0.0228853759765625, 0.022597631454467772, 0.021970943450927736, 0.02211123275756836, 0.021993471145629884, 0.02209382438659668, 0.022916095733642578, 0.02286899185180664, 0.022793216705322264, 0.022760543823242187, 0.02236406326293945, 0.02205695915222168, 0.021949472427368163, 0.02201193618774414, 0.022840255737304686, 0.022786048889160155, 0.022808576583862306, 0.02262937545776367, 0.022751232147216797, 0.02225971221923828, 0.02187571144104004, 0.022591487884521484, 0.022837247848510742, 0.022833152770996092, 0.04536422348022461, 0.022022144317626953, 0.021947391510009767, 0.022206464767456056, 0.022924287796020508, 0.022586368560791017, 0.022774784088134766, 0.022719520568847656, 0.02273276710510254, 0.022074432373046876, 0.02202822494506836, 0.022815744400024415, 0.02285875129699707, 0.02211840057373047, 0.022150144577026368, 0.02207334327697754, 0.022185983657836913, 0.022718463897705078, 0.022920192718505858, 0.022847488403320314, 0.022322175979614257, 0.021969919204711915, 0.022023168563842774, 0.021993471145629884, 0.02205388832092285, 0.02225766372680664, 0.023045120239257814, 0.022797311782836914, 0.02272153663635254, 0.022168575286865236, 0.021926912307739257, 0.022141952514648438, 0.02210406494140625, 0.022666240692138673, 0.022813695907592774, 0.022072320938110353, 0.02206617546081543, 0.021908479690551756, 0.022237184524536133, 0.022983680725097655, 0.022715391159057616, 0.022424575805664062, 0.022754304885864256, 0.022716415405273437, 0.022787071228027343, 0.02270515251159668, 0.02267136001586914, 0.022408191680908202, 0.022803455352783202, 0.02273587226867676, 0.02205286407470703, 0.021961727142333985, 0.022080575942993164, 0.02199443244934082, 0.021876735687255858, 0.022608896255493165, 0.022923263549804687, 0.024328224182128905, 0.02312495994567871, 0.02269388771057129, 0.02209587287902832, 0.022236160278320313, 0.022725631713867187, 0.04547792053222656, 0.02200060844421387, 0.021989376068115234, 0.022024192810058595, 0.02204364776611328, 0.021934080123901366, 0.0226693115234375, 0.022880319595336915, 0.022864831924438476, 0.022709247589111328, 0.021977088928222657, 0.02228428840637207, 0.022017023086547852, 0.022760448455810548, 0.022492223739624024, 0.02292627143859863, 0.022723583221435546, 0.022821887969970703, 0.022403072357177735, 0.02246963119506836, 0.02287001609802246, 0.022354944229125977, 0.02249932861328125, 0.02209689521789551, 0.022041599273681642, 0.022785152435302734, 0.022750080108642576, 0.022714368820190428, 0.022681600570678712, 0.022766592025756836, 0.02247382354736328, 0.022752159118652342, 0.022737920761108397, 0.02269388771057129, 0.02281881523132324, 0.022822912216186524, 0.02270310401916504, 0.02281062316894531, 0.022529024124145508, 0.021980159759521483, 0.022739967346191405, 0.022734848022460938, 0.022769664764404295, 0.022610944747924806, 0.022855680465698244, 0.022797311782836914, 0.022676479339599608, 0.022820863723754883, 0.02269900894165039, 0.022768640518188478, 0.022964223861694336, 0.022920192718505858, 0.022584320068359375, 0.022871040344238282, 0.022611967086791994, 0.022205440521240235, 0.022966272354125978, 0.02265292739868164, 0.022626399993896484, 0.022563743591308593, 0.022518783569335937, 0.02223308753967285, 0.02273587226867676, 0.048899070739746094, 0.023196672439575194, 0.022738943099975584, 0.02209791946411133, 0.022170623779296874, 0.022762496948242186, 0.02271232032775879, 0.022033407211303712, 0.021991424560546875, 0.02203545570373535, 0.02206003189086914, 0.021820415496826173, 0.02205695915222168, 0.022757375717163086, 0.022955007553100586, 0.022831104278564454, 0.022687744140625, 0.023044095993041993, 0.022787071228027343, 0.023138303756713868, 0.022961215972900392, 0.02306656074523926, 0.023003135681152344, 0.02309836769104004, 0.02311884880065918, 0.023021568298339845, 0.0231147518157959, 0.02314854431152344, 0.022990848541259764, 0.02309836769104004, 0.023121919631958008, 0.02308198356628418, 0.023641088485717773, 0.023085056304931642, 0.023145471572875977, 0.022998016357421876, 0.02310553550720215, 0.02306355285644531, 0.022983680725097655, 0.02301644706726074, 0.022969343185424804, 0.023060480117797853, 0.0231014404296875, 0.022990848541259764, 0.02330624008178711, 0.023191551208496093, 0.023002111434936523, 0.02308915138244629, 0.02293452835083008, 0.023055423736572267, 0.02302764892578125, 0.02306764793395996, 0.023051263809204102, 0.022965248107910157, 0.02307276725769043, 0.022982656478881838, 0.02309529685974121, 0.02310246467590332, 0.02307583999633789, 0.02310246467590332, 0.023128063201904296, 0.0230328311920166, 0.023051263809204102]",tokens/s,43.24321870163836,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-2.7b,,cuda,0,42,,,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,217063f5c507ed7cc255df7e1f64c4333a0b4dfe,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,,MB,1296.32256,6797.39392,0.0,6150.946816,5988.307968,s,10,0.3439705619812012,0.03439705619812013,0.00028379952317182644,0.0342990894317627,0.034425238037109375,0.03483617134094238,0.03516491798400879,"[0.03524710464477539, 0.03430201721191406, 0.034300289154052734, 0.03429308700561524, 0.034333919525146486, 0.03429167938232422, 0.03429788970947266, 0.03428041458129883, 0.034293888092041015, 0.03433027267456055]",tokens/s,7442.497361561742,kWh,4.077723770762769e-07,2.2344095077453947e-07,2.018158702648284e-06,2.6493720304991006e-06,tokens/kWh,96626671.1707429,MB,1296.65024,6797.39392,0.0,6150.946816,6117.056512,s,10,13.51379443359375,1.351379443359375,0.008485435861613749,1.3481452026367187,1.3610734985351562,1.364829229736328,1.3678338146972657,"[1.3685849609375, 1.3602388916015624, 1.357792724609375, 1.34811376953125, 1.3554085693359375, 1.346927978515625, 1.3481766357421876, 1.3412799072265624, 1.347203857421875, 1.340067138671875]",tokens/s,46.619030879579746,kWh,1.590264242727638e-05,8.714526017867747e-06,5.493133130135079e-05,7.95484997464949e-05,tokens/kWh,791969.6813990001,,s,629,13.799259147644053,0.021938408819783855,0.0041494085879454174,0.021323776245117186,0.021876121139526368,0.02208463363647461,0.05605527572631836,"[0.02187264060974121, 0.02148863983154297, 0.021340160369873046, 0.02130636787414551, 0.021355520248413085, 0.021312511444091797, 0.0212992000579834, 0.021531648635864258, 0.021313535690307618, 0.021332992553710937, 0.0212490234375, 0.021316608428955077, 0.0212674560546875, 0.02126131248474121, 0.0212541446685791, 0.021355520248413085, 0.021545984268188476, 0.02186137580871582, 0.02180607986450195, 0.021324800491333007, 0.02125721549987793, 0.02126848030090332, 0.02130636787414551, 0.021323776245117186, 0.021534719467163087, 0.026009599685668947, 0.02509926414489746, 0.022369279861450195, 0.021420032501220702, 0.021321727752685548, 0.021312511444091797, 0.02145280075073242, 0.021909503936767577, 0.021819391250610352, 0.02183475112915039, 0.021774335861206053, 0.02190438461303711, 0.021741567611694337, 0.022090751647949217, 0.02190438461303711, 0.021555200576782226, 0.02184294319152832, 0.0216494083404541, 0.022227968215942383, 0.021748735427856446, 0.02184294319152832, 0.021765119552612306, 0.021227519989013673, 0.021348352432250976, 0.021321727752685548, 0.021817344665527344, 0.021710847854614256, 0.021734399795532225, 0.021401599884033205, 0.021566463470458985, 0.02187059211730957, 0.02180607986450195, 0.021603328704833984, 0.021304319381713867, 0.021731327056884766, 0.021570560455322265, 0.021326847076416015, 0.056668159484863284, 0.02170675277709961, 0.021202943801879884, 0.021296127319335938, 0.021344255447387696, 0.021948415756225585, 0.021393407821655275, 0.021251071929931642, 0.02128486442565918, 0.021387264251708983, 0.021530624389648437, 0.02103091239929199, 0.02106675148010254, 0.02127667236328125, 0.021757951736450197, 0.021809152603149414, 0.021536800384521486, 0.02173641586303711, 0.021575679779052736, 0.02186956787109375, 0.021369855880737306, 0.022303743362426756, 0.02187468719482422, 0.021839872360229492, 0.02165043258666992, 0.021279743194580078, 0.021585920333862304, 0.02104319953918457, 0.021575679779052736, 0.0218603515625, 0.021320703506469727, 0.02148863983154297, 0.021336063385009766, 0.021746688842773438, 0.02167807960510254, 0.021291072845458985, 0.021342144012451172, 0.021585920333862304, 0.02186956787109375, 0.02186444854736328, 0.021569536209106444, 0.021316608428955077, 0.02173030471801758, 0.021602304458618164, 0.021315584182739256, 0.02132275199890137, 0.022208511352539064, 0.022006784439086914, 0.021720064163208007, 0.02106879997253418, 0.021312511444091797, 0.02168115234375, 0.0216760311126709, 0.021801984786987305, 0.021167104721069335, 0.021179391860961915, 0.021773311614990236, 0.021902336120605468, 0.02185625648498535, 0.02190336036682129, 0.021968896865844727, 0.022013952255249023, 0.021881856918334962, 0.056581119537353515, 0.021971967697143553, 0.02183888053894043, 0.021606367111206056, 0.02126335906982422, 0.021320703506469727, 0.021180416107177736, 0.021150720596313476, 0.02130227279663086, 0.021206016540527343, 0.021579776763916016, 0.02165760040283203, 0.02191155242919922, 0.02169856071472168, 0.021699583053588867, 0.021894144058227538, 0.021557247161865235, 0.02124595260620117, 0.02122444725036621, 0.02162483215332031, 0.021881856918334962, 0.021836799621582033, 0.021598207473754884, 0.02128281593322754, 0.021796863555908205, 0.021329919815063478, 0.021279743194580078, 0.021320703506469727, 0.02183884811401367, 0.021954559326171876, 0.021086208343505858, 0.02105753517150879, 0.021212160110473634, 0.02165350341796875, 0.021622783660888673, 0.021159936904907226, 0.021312511444091797, 0.021988351821899413, 0.023443456649780273, 0.022245376586914063, 0.021316608428955077, 0.021219327926635743, 0.021401599884033205, 0.021893119812011717, 0.021813247680664064, 0.02183475112915039, 0.021361663818359376, 0.021775360107421874, 0.02389299201965332, 0.02222489547729492, 0.02111187171936035, 0.021202943801879884, 0.02117318344116211, 0.021014528274536134, 0.021153791427612305, 0.02105753517150879, 0.021136383056640624, 0.02147225570678711, 0.021072895050048827, 0.02105855941772461, 0.02101759910583496, 0.021246976852416992, 0.021264415740966797, 0.05615100860595703, 0.021242879867553712, 0.02149580764770508, 0.02227097511291504, 0.022339584350585938, 0.02210304069519043, 0.02186751937866211, 0.021507072448730468, 0.021122047424316406, 0.02126438331604004, 0.021348352432250976, 0.021187583923339845, 0.02127769660949707, 0.021255168914794922, 0.02143027114868164, 0.02100223922729492, 0.02102681541442871, 0.021032960891723632, 0.021177343368530274, 0.02101862335205078, 0.02102374458312988, 0.02103091239929199, 0.021082111358642578, 0.021132287979125978, 0.02123472023010254, 0.021222368240356445, 0.021259263992309572, 0.021190656661987304, 0.021275648117065428, 0.021223424911499023, 0.02122444725036621, 0.021307392120361326, 0.0212807674407959, 0.021584896087646483, 0.02143129539489746, 0.021407743453979493, 0.021327871322631836, 0.021332992553710937, 0.021259263992309572, 0.021352447509765626, 0.021350400924682617, 0.02128895950317383, 0.02127872085571289, 0.02126540756225586, 0.02167296028137207, 0.021580799102783203, 0.021764095306396485, 0.021296127319335938, 0.021389312744140625, 0.021429248809814453, 0.021448703765869142, 0.021122047424316406, 0.021283840179443358, 0.02185113525390625, 0.02125619125366211, 0.02105548858642578, 0.02124185562133789, 0.02125312042236328, 0.023193599700927735, 0.022075456619262697, 0.021352415084838868, 0.021165023803710936, 0.021511167526245118, 0.05597491073608398, 0.02103910446166992, 0.021202943801879884, 0.021818368911743165, 0.021554176330566405, 0.021183488845825195, 0.021227519989013673, 0.02119987106323242, 0.02140880012512207, 0.021683168411254884, 0.021343231201171875, 0.021255168914794922, 0.02162073516845703, 0.021823488235473632, 0.021755903244018555, 0.021320703506469727, 0.021555200576782226, 0.021949440002441405, 0.021599231719970705, 0.02128998374938965, 0.02128998374938965, 0.021634143829345705, 0.021810079574584963, 0.02150297546386719, 0.0212674560546875, 0.021293119430541994, 0.021891008377075194, 0.021525503158569336, 0.021295103073120117, 0.021285888671875, 0.02142310333251953, 0.021819391250610352, 0.021799936294555664, 0.02143436813354492, 0.021356544494628905, 0.022255615234375, 0.022106111526489256, 0.021955583572387697, 0.02188800048828125, 0.02142207908630371, 0.02182246398925781, 0.021533695220947266, 0.02125721549987793, 0.02125209617614746, 0.021327871322631836, 0.02181427192687988, 0.02141187286376953, 0.021278688430786133, 0.02128998374938965, 0.021553152084350585, 0.021353471755981446, 0.021283840179443358, 0.02130227279663086, 0.02130636787414551, 0.02143129539489746, 0.021778432846069336, 0.02162380790710449, 0.02126950454711914, 0.021386240005493166, 0.021843967437744142, 0.021591039657592775, 0.021315584182739256, 0.02168012809753418, 0.056130561828613285, 0.021242879867553712, 0.021201919555664063, 0.021363712310791014, 0.021527551651000978, 0.021134336471557616, 0.02102681541442871, 0.02107187271118164, 0.021597183227539063, 0.0214783992767334, 0.021282880783081055, 0.021266368865966796, 0.02126848030090332, 0.0212490234375, 0.021234687805175782, 0.021230592727661132, 0.02127667236328125, 0.02130636787414551, 0.021352447509765626, 0.02127872085571289, 0.021390335083007812, 0.021444608688354492, 0.021352447509765626, 0.021309440612792968, 0.02128281593322754, 0.021323776245117186, 0.02130636787414551, 0.021325824737548828, 0.021337087631225587, 0.021346303939819337, 0.02126950454711914, 0.021326847076416015, 0.021340160369873046, 0.021313535690307618, 0.021296127319335938, 0.021259263992309572, 0.02131455993652344, 0.021983232498168945, 0.021578752517700195, 0.02125721549987793, 0.02128179168701172, 0.021349376678466796, 0.021382144927978516, 0.021325824737548828, 0.02131865692138672, 0.02128691291809082, 0.021331968307495116, 0.021326847076416015, 0.02127462387084961, 0.021308416366577147, 0.02127667236328125, 0.021349376678466796, 0.021326847076416015, 0.02122547149658203, 0.02126848030090332, 0.021954559326171876, 0.022030336380004883, 0.021574655532836915, 0.0213309440612793, 0.021441535949707033, 0.022102048873901367, 0.021575647354125975, 0.021755903244018555, 0.056215553283691405, 0.021843967437744142, 0.02228531265258789, 0.02168115234375, 0.021368831634521485, 0.021710847854614256, 0.021526527404785157, 0.021176319122314453, 0.021214208602905273, 0.02165452766418457, 0.021408767700195314, 0.02104319953918457, 0.02103193664550781, 0.021144575119018554, 0.021178367614746094, 0.020988927841186524, 0.02102681541442871, 0.02102579116821289, 0.021605375289916993, 0.021840896606445313, 0.02122444725036621, 0.021171199798583985, 0.02104729652404785, 0.021506048202514647, 0.021916671752929686, 0.021395456314086913, 0.02168217658996582, 0.021538816452026367, 0.02170163154602051, 0.021506048202514647, 0.020992000579833983, 0.02148044776916504, 0.021710847854614256, 0.021287935256958008, 0.021046272277832033, 0.02104422378540039, 0.020990976333618162, 0.02109644889831543, 0.021285888671875, 0.022010879516601564, 0.021547008514404296, 0.021948415756225585, 0.02198739242553711, 0.021816255569458008, 0.021598207473754884, 0.02126848030090332, 0.02145484733581543, 0.021703680038452147, 0.021735424041748046, 0.02148659133911133, 0.02112512016296387, 0.021238784790039062, 0.02146099281311035, 0.02104422378540039, 0.02122854423522949, 0.021242879867553712, 0.02130227279663086, 0.02105241584777832, 0.021078016281127928, 0.021000192642211913, 0.02109235191345215, 0.02122956848144531, 0.021246976852416992, 0.055943168640136716, 0.02109132766723633, 0.021082111358642578, 0.02102070426940918, 0.021032928466796875, 0.02104422378540039, 0.021506048202514647, 0.021737472534179687, 0.021202943801879884, 0.02109337615966797, 0.021355520248413085, 0.021815296173095702, 0.021771263122558594, 0.021189632415771483, 0.021731327056884766, 0.021550079345703126, 0.021564416885375977, 0.021155839920043946, 0.02106982421875, 0.02104729652404785, 0.02145996856689453, 0.021974016189575195, 0.021375999450683594, 0.021230592727661132, 0.021317632675170898, 0.021607456207275392, 0.021213151931762694, 0.021078016281127928, 0.02105036735534668, 0.02103603172302246, 0.021118976593017577, 0.021346303939819337, 0.021325824737548828, 0.021004287719726563, 0.021045248031616212, 0.02103193664550781, 0.02123161506652832, 0.02111692810058594, 0.021142528533935546, 0.02103500747680664, 0.021044288635253906, 0.021251007080078124, 0.021129215240478515, 0.021046272277832033, 0.021112831115722656, 0.021242879867553712, 0.021361663818359376, 0.02125312042236328, 0.02126540756225586, 0.0212490234375, 0.02128179168701172, 0.021308416366577147, 0.021304351806640625, 0.021446624755859376, 0.021368831634521485, 0.021235712051391603, 0.02127872085571289, 0.021576704025268553, 0.021835775375366212, 0.02129715156555176, 0.02123673629760742, 0.021315584182739256, 0.02140880012512207, 0.05616739273071289, 0.02126131248474121, 0.02145894432067871, 0.02130534362792969, 0.021348384857177733, 0.021228511810302733, 0.02123263931274414, 0.021809215545654296, 0.021246912002563477, 0.021209087371826172, 0.0212490234375, 0.02124595260620117, 0.021533695220947266, 0.021283840179443358, 0.02105036735534668, 0.02101759910583496, 0.02105753517150879, 0.021336063385009766, 0.02132275199890137, 0.02127257537841797, 0.02126540756225586, 0.02131046485900879, 0.021558271408081055, 0.02126848030090332, 0.021293119430541994, 0.021387199401855468, 0.021583871841430666, 0.02127462387084961, 0.02127667236328125, 0.02127359962463379, 0.021300224304199217, 0.02162892723083496, 0.021752832412719726, 0.021279743194580078, 0.021381120681762695, 0.021324800491333007, 0.022137855529785155, 0.02142720031738281, 0.021613567352294923, 0.02125619125366211, 0.021429248809814453, 0.02125823974609375, 0.02126335906982422, 0.021277727127075194, 0.02132374382019043, 0.021317663192749022, 0.021255136489868164, 0.021337087631225587, 0.021235712051391603, 0.021543935775756837, 0.02125823974609375, 0.0212490234375, 0.02146713638305664, 0.021336063385009766, 0.02163199996948242, 0.02172313690185547, 0.02125823974609375, 0.0212674560546875, 0.021343231201171875, 0.021354496002197267, 0.022286336898803712, 0.021815296173095702, 0.021078016281127928, 0.05608652877807617, 0.02123366355895996, 0.021243904113769533, 0.021344255447387696, 0.02128281593322754, 0.021098495483398438, 0.02103603172302246, 0.020993024826049804, 0.0210565128326416, 0.02123263931274414, 0.02106879997253418, 0.02112006378173828, 0.02122540855407715, 0.021175359725952147, 0.021110719680786132, 0.021013504028320314, 0.021016576766967773, 0.02104115104675293, 0.02142207908630371, 0.021167104721069335, 0.02106265640258789, 0.02101759910583496, 0.02105036735534668, 0.02105446434020996, 0.021016576766967773, 0.021012479782104493, 0.021038143157958985, 0.021285823822021484, 0.021600255966186522, 0.02124083137512207, 0.021130239486694336, 0.02106470489501953, 0.02148659133911133, 0.021191680908203125, 0.02129715156555176, 0.021136383056640624, 0.021172224044799806, 0.021572608947753907, 0.021213184356689452, 0.02109235191345215, 0.02107699203491211, 0.02128281593322754, 0.021521408081054686, 0.021321727752685548, 0.0210565128326416, 0.02106265640258789, 0.021334016799926758, 0.021308416366577147, 0.021106687545776368, 0.02109235191345215, 0.02101759910583496, 0.021126144409179686, 0.021061632156372072, 0.02307689666748047, 0.022600736618041992, 0.022022079467773438, 0.02145075225830078, 0.02128895950317383, 0.021338111877441408, 0.0212992000579834, 0.02165657615661621, 0.021303295135498047, 0.021315616607666017]",tokens/s,45.582157220910624,,,main,False,False,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen2-beta-72B,Qwen/Qwen2-beta-72B,cuda,0,42,,,True,,,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",qwen2,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-0.5B,,cuda,0,42,,,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,217063f5c507ed7cc255df7e1f64c4333a0b4dfe,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,,MB,1286.94272,2227.699712,0.0,1581.252608,1426.269184,s,10,0.22171405029296873,0.022171405029296876,0.0007901586888176377,0.021733263969421386,0.022868406486511228,0.023498507595062255,0.024002588481903075,"[0.02412860870361328, 0.02163865661621094, 0.02265983963012695, 0.022728384017944334, 0.02163350486755371, 0.02156620788574219, 0.021581632614135742, 0.02153798484802246, 0.021827871322631837, 0.022411359786987304]",tokens/s,11546.404012814093,kWh,2.58353698866741e-07,1.4156572396048704e-07,6.657914108471634e-07,1.0657108336743914e-06,tokens/kWh,240215255.31213298,MB,1286.94272,2229.796864,0.0,1583.34976,1426.271744,s,10,13.4040087890625,1.34040087890625,0.02574622815490189,1.3282482299804688,1.3791546752929686,1.3815338195800781,1.3834371350097656,"[1.33265380859375, 1.34905908203125, 1.3786259765625, 1.3839129638671874, 1.3238426513671875, 1.3176536865234374, 1.3119920654296875, 1.31819140625, 1.321446044921875, 1.366631103515625]",tokens/s,47.000864436471566,kWh,1.5748179307960294e-05,8.629782943563365e-06,2.7694675721951442e-05,5.2072637973475106e-05,tokens/kWh,1209848.4434779568,,s,629,13.578077138900747,0.021586768106360503,0.00270674205748266,0.021170175552368165,0.02204037170410156,0.02215116882324219,0.0427763720703125,"[0.02282592010498047, 0.021413888931274414, 0.02087321662902832, 0.02089779281616211, 0.020839424133300782, 0.02164735984802246, 0.022206464767456056, 0.02166374397277832, 0.021465087890625, 0.021335039138793945, 0.021251071929931642, 0.02082918357849121, 0.02084249687194824, 0.020741119384765624, 0.020836448669433592, 0.020779935836791993, 0.023398399353027344, 0.02121625518798828, 0.020567039489746093, 0.020692991256713866, 0.020815872192382814, 0.020929536819458007, 0.02084556770324707, 0.020759552001953126, 0.020818944931030273, 0.021089279174804687, 0.021726303100585938, 0.021909408569335938, 0.021780479431152345, 0.020914176940917968, 0.020914176940917968, 0.020839424133300782, 0.020580352783203124, 0.020666368484497072, 0.020884479522705078, 0.021156864166259767, 0.020952064514160155, 0.02186649513244629, 0.02088140869140625, 0.020790271759033203, 0.020813823699951172, 0.021003263473510742, 0.0210513916015625, 0.02105753517150879, 0.020983808517456053, 0.021801984786987305, 0.021172224044799806, 0.020937728881835937, 0.020918272018432618, 0.021581823348999024, 0.021009408950805664, 0.02069196891784668, 0.020777984619140624, 0.020746240615844725, 0.02088652801513672, 0.02085068893432617, 0.02085273551940918, 0.020998144149780275, 0.020970495223999023, 0.020938751220703124, 0.020568063735961914, 0.020759552001953126, 0.042820606231689456, 0.020738048553466795, 0.020600831985473633, 0.020550655364990233, 0.020511743545532226, 0.020468736648559572, 0.020509695053100584, 0.020709375381469726, 0.020641792297363282, 0.020562944412231447, 0.020680736541748047, 0.020835296630859375, 0.020753408432006838, 0.020724735260009765, 0.020695039749145508, 0.02066022491455078, 0.02063871955871582, 0.020463615417480468, 0.021644287109375, 0.021953535079956055, 0.021586944580078125, 0.02151628875732422, 0.02168524742126465, 0.021642240524291992, 0.021566463470458985, 0.021564416885375977, 0.021564416885375977, 0.021530624389648437, 0.021825536727905274, 0.022793216705322264, 0.02204364776611328, 0.022009855270385743, 0.02227916717529297, 0.020831232070922853, 0.020710399627685547, 0.02063871955871582, 0.02046668815612793, 0.02187775993347168, 0.02244812774658203, 0.021980159759521483, 0.021828607559204103, 0.02192793655395508, 0.02209791946411133, 0.021984256744384766, 0.021933055877685546, 0.02206515121459961, 0.021778432846069336, 0.022162431716918944, 0.021936128616333008, 0.021958656311035156, 0.02165862464904785, 0.021750783920288085, 0.021531648635864258, 0.021639167785644533, 0.021549055099487305, 0.021592063903808592, 0.021570560455322265, 0.021577728271484374, 0.02172313690185547, 0.021603328704833984, 0.021639167785644533, 0.021545984268188476, 0.02191360092163086, 0.04522598266601562, 0.022033407211303712, 0.02165452766418457, 0.021783552169799804, 0.021758975982666014, 0.02160950469970703, 0.021676000595092774, 0.021735424041748046, 0.02169241523742676, 0.022396928787231447, 0.02215116882324219, 0.022188032150268554, 0.02208460807800293, 0.021805055618286134, 0.022187007904052734, 0.02204364776611328, 0.02205900764465332, 0.022025215148925782, 0.02164121627807617, 0.021925888061523437, 0.02228428840637207, 0.0216944637298584, 0.021967872619628907, 0.022032384872436524, 0.02191155242919922, 0.021863456726074218, 0.021592031478881835, 0.021608448028564452, 0.02166886329650879, 0.02167500877380371, 0.021638143539428712, 0.02164531135559082, 0.021729280471801758, 0.02167091178894043, 0.021766143798828123, 0.02170675277709961, 0.021772287368774415, 0.021613567352294923, 0.021585920333862304, 0.021561344146728514, 0.021576704025268553, 0.02162483215332031, 0.021600255966186522, 0.02162892723083496, 0.02313932800292969, 0.022120447158813478, 0.02205900764465332, 0.02208460807800293, 0.02201907157897949, 0.021921791076660157, 0.0216760311126709, 0.021610496520996093, 0.02167398452758789, 0.02171801567077637, 0.022190080642700196, 0.02210099220275879, 0.02206617546081543, 0.02206105613708496, 0.022006784439086914, 0.021783552169799804, 0.02207539176940918, 0.021884927749633788, 0.022046720504760742, 0.04527001571655274, 0.022380544662475587, 0.02208665657043457, 0.022125568389892578, 0.022001663208007814, 0.02206617546081543, 0.021707775115966797, 0.021954559326171876, 0.021976064682006836, 0.02168832015991211, 0.02189619255065918, 0.021994495391845705, 0.022031360626220704, 0.02209689521789551, 0.021957632064819335, 0.021981184005737304, 0.02215116882324219, 0.021958656311035156, 0.022039552688598633, 0.021953535079956055, 0.022071296691894532, 0.021965824127197265, 0.021951488494873047, 0.021969919204711915, 0.02203647994995117, 0.021923904418945313, 0.02200876808166504, 0.021942272186279296, 0.02167296028137207, 0.02186444854736328, 0.021704704284667968, 0.02167500877380371, 0.021687295913696288, 0.02168320083618164, 0.021686304092407228, 0.021647327423095702, 0.02243071937561035, 0.022181888580322266, 0.021966848373413086, 0.021994495391845705, 0.02206515121459961, 0.022246400833129884, 0.0222423038482666, 0.02206822395324707, 0.02210304069519043, 0.022006784439086914, 0.02207334327697754, 0.02206208038330078, 0.0224768009185791, 0.021999616622924805, 0.021695487976074217, 0.021976064682006836, 0.021979135513305666, 0.02168627166748047, 0.02183884811401367, 0.021719039916992186, 0.021543935775756837, 0.021731327056884766, 0.021695487976074217, 0.022021120071411132, 0.02205081558227539, 0.022072320938110353, 0.021932031631469725, 0.04286675262451172, 0.020921279907226562, 0.020832256317138673, 0.020571168899536134, 0.021840864181518555, 0.021972991943359374, 0.021985279083251954, 0.02167807960510254, 0.021751808166503905, 0.021523456573486328, 0.02162892723083496, 0.02168422317504883, 0.021768192291259765, 0.021804031372070314, 0.02189516830444336, 0.021914655685424805, 0.02214806365966797, 0.02207539176940918, 0.02202828788757324, 0.021994495391845705, 0.021972991943359374, 0.020761600494384767, 0.020823040008544923, 0.02082611274719238, 0.02083839988708496, 0.020798463821411133, 0.02065715217590332, 0.02058243179321289, 0.02071753692626953, 0.020813823699951172, 0.020904960632324218, 0.020904960632324218, 0.020858879089355468, 0.02073811149597168, 0.020682687759399413, 0.02070528030395508, 0.020537343978881836, 0.02067046356201172, 0.02063564872741699, 0.020501504898071288, 0.020489215850830078, 0.020543487548828124, 0.020760576248168947, 0.020702207565307617, 0.020634624481201173, 0.020520959854125977, 0.020715520858764647, 0.020747264862060546, 0.020699136734008788, 0.020751359939575196, 0.020478975296020507, 0.020557823181152343, 0.020797439575195312, 0.020718624114990234, 0.02055369567871094, 0.020695039749145508, 0.020709375381469726, 0.02051481628417969, 0.020702207565307617, 0.0204902400970459, 0.02048307228088379, 0.020736000061035157, 0.020737024307250978, 0.042521598815917966, 0.02045337677001953, 0.020733951568603515, 0.02085990333557129, 0.020790271759033203, 0.020816896438598635, 0.020785152435302736, 0.021215232849121093, 0.020918272018432618, 0.020668415069580077, 0.021137407302856445, 0.020737024307250978, 0.020717567443847656, 0.020806655883789063, 0.020750335693359375, 0.021292032241821288, 0.021394432067871092, 0.020743167877197266, 0.020770816802978515, 0.02059878349304199, 0.020839424133300782, 0.02068889617919922, 0.02106777572631836, 0.021342208862304687, 0.020783103942871094, 0.02072889518737793, 0.02069702339172363, 0.020701183319091796, 0.020753408432006838, 0.020877344131469727, 0.02132169532775879, 0.02124083137512207, 0.020779008865356444, 0.02071244812011719, 0.020694015502929687, 0.020512767791748047, 0.020840448379516603, 0.021346303939819337, 0.021507072448730468, 0.02086297607421875, 0.02207539176940918, 0.02229452705383301, 0.020931583404541015, 0.020738048553466795, 0.020675615310668947, 0.020640735626220704, 0.020719615936279297, 0.02084249687194824, 0.020719615936279297, 0.02070528030395508, 0.020707328796386718, 0.020924415588378906, 0.02110771179199219, 0.021029951095581055, 0.020664255142211915, 0.021153791427612305, 0.021403648376464843, 0.020918272018432618, 0.020746240615844725, 0.020822015762329102, 0.020802560806274413, 0.0208721923828125, 0.020752384185791017, 0.04279500961303711, 0.020806655883789063, 0.021196800231933592, 0.02126848030090332, 0.020810752868652343, 0.020727807998657227, 0.020716543197631835, 0.021303295135498047, 0.020576255798339844, 0.020754432678222655, 0.020952064514160155, 0.020580352783203124, 0.020445184707641603, 0.020612096786499022, 0.020501504898071288, 0.020562944412231447, 0.020716543197631835, 0.021386240005493166, 0.021148672103881837, 0.02050662422180176, 0.02049843215942383, 0.02069708824157715, 0.020685823440551757, 0.020699136734008788, 0.02085171127319336, 0.021170175552368165, 0.020762624740600585, 0.020737024307250978, 0.020746240615844725, 0.020718591690063477, 0.020720640182495118, 0.020939775466918945, 0.021346303939819337, 0.02087321662902832, 0.021407743453979493, 0.02144972801208496, 0.020925439834594727, 0.02079641532897949, 0.02084659194946289, 0.021166080474853514, 0.020874271392822264, 0.02092848014831543, 0.02082713508605957, 0.020686847686767578, 0.020990976333618162, 0.020783103942871094, 0.020788223266601562, 0.020767744064331056, 0.020703231811523438, 0.020774911880493165, 0.020817920684814452, 0.02073401641845703, 0.0207523193359375, 0.020783103942871094, 0.02069196891784668, 0.020561920166015626, 0.020538368225097657, 0.020607999801635742, 0.020724735260009765, 0.020690944671630858, 0.020692991256713866, 0.020786176681518553, 0.02069811248779297, 0.042728446960449216, 0.020674560546875, 0.02132275199890137, 0.020990976333618162, 0.020756479263305663, 0.020703231811523438, 0.020617216110229493, 0.02106879997253418, 0.02086604881286621, 0.02070528030395508, 0.020728832244873048, 0.02224332809448242, 0.021517311096191406, 0.021385215759277345, 0.02068377685546875, 0.020716543197631835, 0.020714496612548827, 0.02065510368347168, 0.020586496353149415, 0.02050662422180176, 0.020569087982177735, 0.02082815933227539, 0.020764671325683593, 0.021315584182739256, 0.020853792190551758, 0.020830175399780274, 0.020766719818115235, 0.020764671325683593, 0.020989952087402345, 0.02206924819946289, 0.021493759155273438, 0.0213309440612793, 0.02103193664550781, 0.020517887115478514, 0.02085478401184082, 0.020799488067626954, 0.020576255798339844, 0.02053222465515137, 0.020485120773315428, 0.020734975814819336, 0.020765695571899414, 0.020731903076171874, 0.020754432678222655, 0.020738048553466795, 0.020724735260009765, 0.020706304550170897, 0.020728832244873048, 0.020765695571899414, 0.02086911964416504, 0.02084864044189453, 0.020789247512817383, 0.02065203285217285, 0.02187468719482422, 0.020880384445190428, 0.020864000320434572, 0.020763647079467772, 0.020784128189086915, 0.020818944931030273, 0.02065407943725586, 0.021141504287719725, 0.021745664596557617, 0.021542911529541017, 0.021327871322631836, 0.04310323333740235, 0.020983808517456053, 0.021437440872192383, 0.0216013126373291, 0.020961248397827148, 0.0208855037689209, 0.02087424087524414, 0.020824064254760744, 0.020823040008544923, 0.020896768569946288, 0.020695104598999023, 0.02080556869506836, 0.020779008865356444, 0.020741119384765624, 0.02088243293762207, 0.020831232070922853, 0.020816896438598635, 0.020825120925903322, 0.02072675132751465, 0.020765695571899414, 0.020779008865356444, 0.020753408432006838, 0.020765695571899414, 0.020766719818115235, 0.020726783752441406, 0.020722688674926756, 0.020785152435302736, 0.020760576248168947, 0.02083328056335449, 0.020769792556762694, 0.020495359420776366, 0.02058956718444824, 0.02081279945373535, 0.020783103942871094, 0.02090188789367676, 0.020722688674926756, 0.020761600494384767, 0.020732927322387695, 0.020867071151733398, 0.020769792556762694, 0.020690944671630858, 0.020538400650024415, 0.0207410888671875, 0.02082713508605957, 0.020563968658447264, 0.020633600234985353, 0.020763647079467772, 0.020764671325683593, 0.020808704376220705, 0.020799488067626954, 0.020797439575195312, 0.02129408073425293, 0.02250444793701172, 0.021850112915039063, 0.02166476821899414, 0.02166374397277832, 0.021569536209106444, 0.02169753646850586, 0.02167091178894043, 0.02143846321105957, 0.021215232849121093, 0.02165760040283203, 0.021784608840942382, 0.044536800384521485, 0.021275648117065428, 0.021359615325927735, 0.02167398452758789, 0.02131865692138672, 0.021580799102783203, 0.021612543106079102, 0.02167807960510254, 0.02167705535888672, 0.02164019203186035, 0.021755903244018555, 0.021708831787109376, 0.021796831130981444, 0.021691423416137695, 0.02171183967590332, 0.021710847854614256, 0.021700639724731446, 0.02171900749206543, 0.021760000228881835, 0.02166169548034668, 0.02170572853088379, 0.021704704284667968, 0.02169241523742676, 0.02163711929321289, 0.02169856071472168, 0.021679103851318358, 0.02168115234375, 0.02169241523742676, 0.021801984786987305, 0.0216494083404541, 0.02164531135559082, 0.0217262077331543, 0.021902336120605468, 0.021751808166503905, 0.021735424041748046, 0.021768192291259765, 0.02171392059326172, 0.021777408599853516, 0.021760000228881835, 0.021745664596557617, 0.02165043258666992, 0.021687295913696288, 0.02182044792175293, 0.02176918411254883, 0.021805055618286134, 0.021702655792236326, 0.02171392059326172, 0.02169856071472168, 0.021708799362182618, 0.021775360107421874, 0.02182143974304199, 0.02183065605163574, 0.021659648895263672, 0.021405696868896484, 0.021526527404785157, 0.021728256225585937, 0.02168524742126465, 0.02168115234375, 0.0217262077331543, 0.021727264404296873, 0.021676000595092774, 0.02166374397277832, 0.021584896087646483]",tokens/s,46.32467422047082,,,main,False,False,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-1.8B,,cuda,0,42,,,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,217063f5c507ed7cc255df7e1f64c4333a0b4dfe,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,,MB,1282.637824,5222.432768,0.0,4575.985664,4408.404992,s,10,0.3124974708557129,0.031249747085571287,0.00090185139925193,0.03089788818359375,0.031724757766723635,0.03277908306121826,0.03362254329681397,"[0.03383340835571289, 0.030997472763061522, 0.030746112823486327, 0.03069455909729004, 0.03073587226867676, 0.030798303604125978, 0.030700031280517577, 0.031203039169311525, 0.03129820823669434, 0.03149046325683594]",tokens/s,8192.066300536588,kWh,3.658403691492583e-07,2.0044497673116097e-07,1.426686727863879e-06,1.992972073744298e-06,tokens/kWh,128451373.38981362,MB,1282.637824,5222.432768,0.0,4575.985664,4408.407552,s,10,13.857858032226563,1.3857858032226564,0.02388665195845957,1.3779736328125,1.4190756225585939,1.4192181701660156,1.419332208251953,"[1.3689766845703124, 1.357552978515625, 1.363571044921875, 1.3694896240234375, 1.3864576416015626, 1.3621201171875, 1.3937149658203125, 1.4193607177734375, 1.4190439453125, 1.4175703125]",tokens/s,45.461571228030316,kWh,1.6344615843212395e-05,8.955108905606538e-06,4.0334874459137036e-05,6.563459920795597e-05,tokens/kWh,959859.5978378943,,s,629,14.112389108657833,0.02243623069738925,0.0037539084422681035,0.02188800048828125,0.02259947509765625,0.022733823776245117,0.052650106048583986,"[0.022372352600097657, 0.0216760311126709, 0.021629951477050782, 0.021507072448730468, 0.021522432327270507, 0.021604352951049805, 0.021731327056884766, 0.021742591857910155, 0.021642240524291992, 0.02170675277709961, 0.021566463470458985, 0.02167705535888672, 0.021573631286621094, 0.021543935775756837, 0.02165555191040039, 0.021982208251953125, 0.02168320083618164, 0.021559295654296876, 0.021627935409545898, 0.021552095413208006, 0.021612543106079102, 0.021593088150024413, 0.021786624908447266, 0.021634048461914062, 0.02232729530334473, 0.022024192810058595, 0.021949440002441405, 0.021518335342407227, 0.021592063903808592, 0.021707775115966797, 0.02170675277709961, 0.021607423782348634, 0.021778432846069336, 0.021921791076660157, 0.021601280212402343, 0.021580799102783203, 0.021659648895263672, 0.021572608947753907, 0.02164121627807617, 0.02162892723083496, 0.021610496520996093, 0.02162892723083496, 0.02162073516845703, 0.021614591598510743, 0.02163609504699707, 0.021518335342407227, 0.021348352432250976, 0.021364736557006835, 0.021840927124023437, 0.021446624755859376, 0.021568544387817384, 0.021591007232666017, 0.021532703399658203, 0.021665760040283203, 0.021561344146728514, 0.021549055099487305, 0.02154806327819824, 0.02170159912109375, 0.022352895736694335, 0.022076416015625, 0.0221265926361084, 0.02166476821899414, 0.05295513534545898, 0.02126131248474121, 0.02129817581176758, 0.021243904113769533, 0.021342208862304687, 0.02148454475402832, 0.021578752517700195, 0.021586944580078125, 0.021521408081054686, 0.021507072448730468, 0.021332992553710937, 0.021321727752685548, 0.021716991424560548, 0.021602304458618164, 0.021493759155273438, 0.02148659133911133, 0.021545984268188476, 0.02149990463256836, 0.021510143280029297, 0.021535743713378908, 0.02148454475402832, 0.021541887283325196, 0.021506048202514647, 0.021255168914794922, 0.021311487197875977, 0.02151628875732422, 0.021556224822998047, 0.021465087890625, 0.021351423263549805, 0.02151219177246094, 0.021522432327270507, 0.0214968318939209, 0.021481472015380858, 0.02152038383483887, 0.02146611213684082, 0.021334016799926758, 0.021778495788574218, 0.021539840698242187, 0.02157049560546875, 0.021584896087646483, 0.021557247161865235, 0.021527551651000978, 0.02168422317504883, 0.021573631286621094, 0.02149990463256836, 0.02142310333251953, 0.021574655532836915, 0.021584896087646483, 0.021592096328735353, 0.021636064529418946, 0.021582847595214845, 0.021587968826293946, 0.021550079345703126, 0.021562368392944335, 0.021560319900512694, 0.02168524742126465, 0.021560319900512694, 0.021613567352294923, 0.021599231719970705, 0.022232063293457033, 0.021622783660888673, 0.02227302360534668, 0.02187161636352539, 0.05274214553833008, 0.021585920333862304, 0.021545984268188476, 0.021562368392944335, 0.021581823348999024, 0.02165862464904785, 0.0216760311126709, 0.021578752517700195, 0.02166579246520996, 0.021597183227539063, 0.02167705535888672, 0.02165350341796875, 0.02345062446594238, 0.022935552597045897, 0.02208768081665039, 0.021605375289916993, 0.021394432067871092, 0.021523456573486328, 0.021617664337158202, 0.021595136642456055, 0.02169139289855957, 0.021554176330566405, 0.021329919815063478, 0.021559295654296876, 0.021551103591918946, 0.021563392639160156, 0.021553152084350585, 0.021413888931274414, 0.021517311096191406, 0.021614591598510743, 0.021595136642456055, 0.021639167785644533, 0.021570560455322265, 0.021539840698242187, 0.02153267288208008, 0.021610496520996093, 0.021578752517700195, 0.021319679260253906, 0.021301248550415038, 0.021562368392944335, 0.021409791946411134, 0.021564416885375977, 0.021617664337158202, 0.021520416259765626, 0.021544927597045897, 0.021525503158569336, 0.021448768615722657, 0.02136569595336914, 0.0212992000579834, 0.021315584182739256, 0.02127462387084961, 0.021303295135498047, 0.021382144927978516, 0.02287615966796875, 0.02290995216369629, 0.02188800048828125, 0.02164531135559082, 0.021621759414672852, 0.021595136642456055, 0.021391359329223633, 0.021307392120361326, 0.021584896087646483, 0.021634048461914062, 0.05267763137817383, 0.021609472274780273, 0.02164735984802246, 0.021588991165161133, 0.021571584701538086, 0.021325824737548828, 0.021432319641113282, 0.021404672622680664, 0.021336063385009766, 0.021283840179443358, 0.021357568740844726, 0.021364736557006835, 0.021574655532836915, 0.021527551651000978, 0.021453855514526367, 0.02133705520629883, 0.021335039138793945, 0.021346303939819337, 0.021518335342407227, 0.022123519897460937, 0.021798912048339843, 0.02167193603515625, 0.02157267189025879, 0.021685184478759764, 0.02145587158203125, 0.021910528182983398, 0.021559295654296876, 0.02162073516845703, 0.021716032028198242, 0.021711807250976562, 0.021378047943115236, 0.021375999450683594, 0.021617664337158202, 0.021646335601806642, 0.021579839706420897, 0.021561279296875, 0.022375423431396483, 0.022809600830078124, 0.02261203193664551, 0.021896160125732422, 0.02276348876953125, 0.022368255615234374, 0.02208460807800293, 0.0219289608001709, 0.02229555130004883, 0.02228326416015625, 0.021921791076660157, 0.02190028762817383, 0.02185830307006836, 0.021998592376708984, 0.022133760452270508, 0.022190080642700196, 0.02182143974304199, 0.02184806442260742, 0.022074367523193358, 0.021595136642456055, 0.02147327995300293, 0.02165657615661621, 0.02148761558532715, 0.02152448081970215, 0.02143846321105957, 0.021534719467163087, 0.02166067123413086, 0.05257932662963867, 0.021349376678466796, 0.021563392639160156, 0.02165760040283203, 0.02194534492492676, 0.02221670341491699, 0.022200319290161134, 0.02184499168395996, 0.02282598304748535, 0.022593536376953126, 0.022253568649291993, 0.022000640869140626, 0.022063104629516602, 0.021719039916992186, 0.02247065544128418, 0.022402048110961914, 0.022245376586914063, 0.022148096084594726, 0.022108160018920898, 0.02225663948059082, 0.02204979133605957, 0.022314016342163085, 0.021767135620117187, 0.02144972801208496, 0.021420032501220702, 0.021962751388549806, 0.02246963119506836, 0.022093856811523437, 0.021812192916870116, 0.02168934440612793, 0.02150399971008301, 0.021642240524291992, 0.02164019203186035, 0.021881856918334962, 0.021926912307739257, 0.02173030471801758, 0.021744640350341796, 0.021955583572387697, 0.021824512481689453, 0.021747711181640626, 0.021955583572387697, 0.02232729530334473, 0.022207487106323243, 0.021943296432495117, 0.02211327934265137, 0.022131711959838866, 0.021986303329467775, 0.022254592895507814, 0.022163455963134765, 0.02248806381225586, 0.022064128875732423, 0.022153215408325197, 0.022185983657836913, 0.022196224212646484, 0.021977088928222657, 0.022047744750976563, 0.022108160018920898, 0.02207744026184082, 0.022007808685302735, 0.022030336380004883, 0.022774784088134766, 0.02146918487548828, 0.02149478340148926, 0.05274009704589844, 0.02150297546386719, 0.02149478340148926, 0.021556224822998047, 0.02148761558532715, 0.02131052780151367, 0.02138105583190918, 0.021562368392944335, 0.021601280212402343, 0.021513216018676756, 0.021497919082641603, 0.0217322883605957, 0.021792768478393554, 0.021832704544067383, 0.021501951217651367, 0.02152038383483887, 0.021972991943359374, 0.022002687454223634, 0.021607423782348634, 0.021977088928222657, 0.021557247161865235, 0.02147737693786621, 0.021711872100830077, 0.021969919204711915, 0.021360639572143555, 0.02147737693786621, 0.021312511444091797, 0.021296127319335938, 0.02126540756225586, 0.021377023696899415, 0.021608448028564452, 0.021786624908447266, 0.02208665657043457, 0.02167807960510254, 0.022737920761108397, 0.021587968826293946, 0.021522432327270507, 0.02127359962463379, 0.02129100799560547, 0.021399551391601563, 0.021543935775756837, 0.02171494483947754, 0.02267344093322754, 0.021657567977905273, 0.021626880645751953, 0.021543935775756837, 0.021766143798828123, 0.02182655906677246, 0.021319679260253906, 0.021392383575439454, 0.021324800491333007, 0.02149068832397461, 0.0216494083404541, 0.021960704803466798, 0.021374975204467773, 0.021301248550415038, 0.021445632934570313, 0.02164531135559082, 0.021733375549316408, 0.021797887802124022, 0.021596160888671875, 0.021603328704833984, 0.021598304748535156, 0.05255465698242188, 0.021364736557006835, 0.021296127319335938, 0.02146713638305664, 0.021332992553710937, 0.021378047943115236, 0.021559295654296876, 0.02186342430114746, 0.021747711181640626, 0.021415935516357423, 0.021235712051391603, 0.021343231201171875, 0.021389312744140625, 0.021335039138793945, 0.02150815963745117, 0.02179987144470215, 0.022276159286499023, 0.021540800094604493, 0.021358591079711914, 0.021932031631469725, 0.022183935165405275, 0.0223242244720459, 0.02260479927062988, 0.022443008422851563, 0.022587392807006838, 0.022371328353881836, 0.022503423690795898, 0.022544384002685547, 0.022553600311279298, 0.022544384002685547, 0.022560768127441407, 0.022429695129394533, 0.022245376586914063, 0.02225766372680664, 0.022632448196411133, 0.022525951385498046, 0.02231500816345215, 0.022190080642700196, 0.022188032150268554, 0.022208511352539064, 0.02244607925415039, 0.022509567260742186, 0.02247987174987793, 0.022399999618530272, 0.022503423690795898, 0.022564863204956053, 0.022585344314575196, 0.02253004837036133, 0.02247475242614746, 0.022371328353881836, 0.022107168197631834, 0.022197216033935548, 0.022140928268432617, 0.022163455963134765, 0.022143007278442383, 0.022192096710205077, 0.022195199966430663, 0.022156288146972656, 0.022260736465454102, 0.022486015319824217, 0.022487039566040038, 0.02252390480041504, 0.022575103759765625, 0.0541317138671875, 0.02244915199279785, 0.022375423431396483, 0.022486015319824217, 0.022543359756469726, 0.022551551818847656, 0.02248089599609375, 0.022444032669067384, 0.02249932861328125, 0.02242361640930176, 0.022444992065429686, 0.022492160797119142, 0.022545408248901368, 0.022405120849609376, 0.02241535949707031, 0.02249728012084961, 0.02252390480041504, 0.02254643249511719, 0.022487039566040038, 0.022487039566040038, 0.0224881591796875, 0.022746015548706054, 0.02229043197631836, 0.022188095092773436, 0.022513599395751954, 0.022322175979614257, 0.022155263900756835, 0.022579200744628908, 0.02246963119506836, 0.02249113655090332, 0.022413312911987306, 0.02245631980895996, 0.022500352859497072, 0.022510591506958007, 0.022557695388793944, 0.022432767868041992, 0.022519807815551757, 0.02248806381225586, 0.022537216186523438, 0.022619136810302733, 0.022733823776245117, 0.022590463638305663, 0.022623231887817383, 0.022579200744628908, 0.02253004837036133, 0.022564863204956053, 0.022626304626464845, 0.022535167694091796, 0.02264575958251953, 0.022776832580566408, 0.022605823516845702, 0.022603776931762694, 0.02264473533630371, 0.022519807815551757, 0.022561792373657227, 0.022920192718505858, 0.022542335510253905, 0.022419456481933595, 0.022602752685546876, 0.022605823516845702, 0.022566911697387695, 0.02260479927062988, 0.02269603157043457, 0.05462313461303711, 0.02305023956298828, 0.0226693115234375, 0.022591487884521484, 0.022729791641235352, 0.02257913589477539, 0.02266009521484375, 0.022684671401977538, 0.022577152252197266, 0.022566911697387695, 0.02249318313598633, 0.022183935165405275, 0.022468608856201173, 0.022467584609985353, 0.022510591506958007, 0.022271999359130858, 0.022334463119506837, 0.02264271926879883, 0.02297340774536133, 0.022625280380249024, 0.022567935943603516, 0.022558719635009765, 0.022486015319824217, 0.02253209686279297, 0.02246348762512207, 0.022701055526733398, 0.022558719635009765, 0.022589439392089843, 0.02253107261657715, 0.022579200744628908, 0.022594560623168947, 0.022559743881225586, 0.02266828727722168, 0.022598655700683593, 0.022466560363769532, 0.0225218563079834, 0.02168627166748047, 0.02164531135559082, 0.02163199996948242, 0.022923263549804687, 0.022977535247802734, 0.022550527572631835, 0.022459392547607423, 0.022211584091186523, 0.022404096603393556, 0.022323200225830078, 0.02247478485107422, 0.022393823623657227, 0.02253824043273926, 0.0224901123046875, 0.02250752067565918, 0.022733823776245117, 0.02262118339538574, 0.022607872009277344, 0.022718463897705078, 0.02253004837036133, 0.022487039566040038, 0.022581247329711913, 0.02254643249511719, 0.022260736465454102, 0.022360063552856444, 0.02254643249511719, 0.02248806381225586, 0.05481584167480469, 0.02262723159790039, 0.022412288665771486, 0.022508544921875, 0.022519807815551757, 0.022571008682250978, 0.022608959197998046, 0.022502336502075195, 0.022608896255493165, 0.022588415145874022, 0.022556671142578123, 0.022502399444580077, 0.022550527572631835, 0.022597631454467772, 0.022487039566040038, 0.02287308883666992, 0.024625152587890626, 0.02284441566467285, 0.022567935943603516, 0.022665279388427734, 0.022529983520507814, 0.022503423690795898, 0.022578176498413087, 0.02249830436706543, 0.022548479080200197, 0.022581247329711913, 0.02254643249511719, 0.02255462455749512, 0.022588415145874022, 0.022500352859497072, 0.0225218563079834, 0.02246348762512207, 0.02254745674133301, 0.02259660720825195, 0.022492160797119142, 0.022509567260742186, 0.022581247329711913, 0.0224849910736084, 0.022487039566040038, 0.022466560363769532, 0.022582271575927734, 0.022441984176635742, 0.0224768009185791, 0.022553600311279298, 0.022560768127441407, 0.022592512130737305, 0.02247475242614746, 0.022410240173339844, 0.022575103759765625, 0.022524927139282228, 0.02250649642944336, 0.022559743881225586, 0.022525951385498046, 0.022785024642944338, 0.02285875129699707, 0.02302566337585449, 0.02161664009094238, 0.021577728271484374, 0.021529600143432616, 0.02162483215332031, 0.021565439224243164, 0.021558271408081055, 0.021553152084350585]",tokens/s,44.57076651990226,,,main,False,False,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-70m,,cuda,0,42,,,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,217063f5c507ed7cc255df7e1f64c4333a0b4dfe,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,,MB,1276.506112,885.522432,0.0,239.075328,215.486464,s,33,0.17781625556945801,0.005388371380892667,0.0002118844441012189,0.005337183952331543,0.005426726531982422,0.005557625675201416,0.006254851970672607,"[0.0064943041801452635, 0.005746016025543213, 0.0053919677734375, 0.0052193918228149415, 0.005302624225616455, 0.005311456203460694, 0.005330815792083741, 0.005290080070495606, 0.005318880081176758, 0.005315072059631347, 0.0053012800216674805, 0.005337183952331543, 0.005354879856109619, 0.0053144960403442385, 0.005359424114227295, 0.005425600051879883, 0.005303552150726318, 0.005389023780822754, 0.005432032108306885, 0.005311071872711182, 0.005326687812805176, 0.005427008152008057, 0.005338592052459717, 0.00535814380645752, 0.00538972806930542, 0.005336095809936523, 0.005366975784301758, 0.005368607997894287, 0.005338304042816162, 0.005303135871887207, 0.005324960231781006, 0.005371967792510987, 0.005316895961761475]",tokens/s,47509.71711188728,kWh,6.558223184455779e-08,3.593580860049073e-08,1.197129096390998e-07,2.212309500841483e-07,tokens/kWh,1157161780.0431032,MB,1276.506112,885.522432,0.0,239.075328,220.896256,s,33,9.83883044433594,0.29814637710108904,0.006565352680415171,0.29647915649414064,0.29813798828125,0.30521984252929685,0.3255126794433594,"[0.3313179626464844, 0.3131764526367187, 0.29581048583984376, 0.29629986572265626, 0.29654156494140627, 0.29702188110351563, 0.2962219848632813, 0.2960429382324219, 0.29619210815429686, 0.29583963012695313, 0.29660260009765627, 0.29660693359375, 0.2959345397949219, 0.2961164245605469, 0.2976431884765625, 0.2959195251464844, 0.2982616882324219, 0.29685659790039065, 0.2957413024902344, 0.2961203918457031, 0.29991543579101565, 0.2964104309082031, 0.2968109436035156, 0.29647915649414064, 0.296353759765625, 0.29699728393554686, 0.29651690673828124, 0.29602044677734374, 0.29627545166015623, 0.29661376953125, 0.29703427124023435, 0.2958043212890625, 0.2973302001953125]",tokens/s,211.3056030147209,kWh,3.457817449008332e-06,1.8947154396265524e-06,5.381778419654403e-06,1.0734311308289286e-05,tokens/kWh,5869030.456695431,,s,2078,10.005302271842952,0.004814871160655898,0.0007238291203950634,0.0046929922103881834,0.004805609750747681,0.005073356771469115,0.010431877031326293,"[0.005583871841430664, 0.005627967834472656, 0.005525440216064453, 0.0054271998405456545, 0.005462016105651855, 0.005389311790466309, 0.005337088108062744, 0.005442560195922852, 0.005994495868682862, 0.006638591766357422, 0.005647359848022461, 0.005537792205810547, 0.005467135906219483, 0.005280767917633057, 0.005312511920928955, 0.005256192207336426, 0.005272575855255127, 0.005250048160552978, 0.0053821439743042, 0.005497856140136719, 0.0054241280555725096, 0.0052336640357971195, 0.005186560153961181, 0.005108736038208008, 0.0050063362121582035, 0.005053440093994141, 0.005228544235229492, 0.005208064079284668, 0.005173247814178467, 0.0051435518264770505, 0.005239808082580567, 0.0051333122253417966, 0.005075967788696289, 0.005111807823181152, 0.005185535907745361, 0.005169151782989502, 0.005142528057098389, 0.004968448162078858, 0.0049192957878112795, 0.004902912139892578, 0.004957183837890625, 0.005086207866668701, 0.004986879825592041, 0.0049725441932678225, 0.00506879997253418, 0.005091328144073487, 0.00506982421875, 0.005254144191741943, 0.005204991817474365, 0.005285888195037842, 0.0052715520858764645, 0.005254144191741943, 0.005164031982421875, 0.005105663776397705, 0.005022719860076904, 0.00520908784866333, 0.005216256141662597, 0.005078015804290771, 0.005079040050506592, 0.0051773438453674315, 0.00501043176651001, 0.004929535865783692, 0.01134499168395996, 0.005028768062591553, 0.005169151782989502, 0.005304384231567383, 0.004997056007385254, 0.004992000102996826, 0.0050462718009948735, 0.00505244779586792, 0.005082079887390137, 0.005191679954528809, 0.005058559894561767, 0.00502784013748169, 0.0050206718444824215, 0.005041152000427246, 0.0052111358642578124, 0.005254144191741943, 0.005246975898742676, 0.005255167961120606, 0.005253119945526123, 0.005144576072692871, 0.005121024131774903, 0.005161983966827393, 0.00502784013748169, 0.0049203200340270994, 0.005280767917633057, 0.004980735778808594, 0.004931583881378174, 0.005055488109588623, 0.004794367790222168, 0.004795392036437988, 0.005058559894561767, 0.004980735778808594, 0.00483843183517456, 0.004698080062866211, 0.0047216639518737795, 0.004690944194793701, 0.00470630407333374, 0.004716544151306152, 0.004958208084106445, 0.0048855037689208985, 0.0049500160217285155, 0.004902944087982177, 0.0050011839866638185, 0.005017600059509277, 0.005132287979125977, 0.005185535907745361, 0.004999167919158935, 0.004848639965057373, 0.004810751914978028, 0.004836351871490479, 0.004877312183380127, 0.004874239921569825, 0.004886623859405518, 0.004864927768707276, 0.0048568320274353025, 0.004850687980651855, 0.0048220157623291016, 0.0047626237869262695, 0.004685823917388916, 0.004707327842712402, 0.00470739221572876, 0.004694975852966308, 0.00469708776473999, 0.01054412841796875, 0.004687871932983399, 0.004682752132415771, 0.004678656101226806, 0.004679679870605469, 0.004678656101226806, 0.004687871932983399, 0.004698112010955811, 0.0047216639518737795, 0.004738048076629638, 0.004694015979766846, 0.004704256057739258, 0.004739071846008301, 0.004715519905090332, 0.004700160026550293, 0.004712448120117187, 0.0046929922103881834, 0.004712448120117187, 0.004689919948577881, 0.004700160026550293, 0.0046991357803344725, 0.004704288005828857, 0.004693984031677246, 0.004701183795928955, 0.004689919948577881, 0.004686848163604736, 0.0046919679641723635, 0.004695040225982666, 0.004676608085632325, 0.00471449613571167, 0.004695040225982666, 0.004685823917388916, 0.004683775901794434, 0.004681727886199951, 0.004686848163604736, 0.004720640182495117, 0.004687871932983399, 0.004682752132415771, 0.004690944194793701, 0.0046888961791992185, 0.004707327842712402, 0.004684800148010254, 0.004690976142883301, 0.004688864231109619, 0.004694015979766846, 0.00469708776473999, 0.004701183795928955, 0.0046888961791992185, 0.0046888961791992185, 0.004685823917388916, 0.00469708776473999, 0.004680704116821289, 0.0046919679641723635, 0.004695040225982666, 0.004677631855010986, 0.00464793586730957, 0.004636672019958496, 0.004655136108398438, 0.004660192012786865, 0.004644864082336426, 0.004642816066741944, 0.004631552219390869, 0.004644864082336426, 0.010299391746520996, 0.004680704116821289, 0.004685823917388916, 0.004695040225982666, 0.004709375858306885, 0.004702208042144776, 0.00470527982711792, 0.004707327842712402, 0.004690944194793701, 0.004695040225982666, 0.0047226881980895995, 0.004694015979766846, 0.004707327842712402, 0.004683775901794434, 0.0046919679641723635, 0.004682752132415771, 0.004681727886199951, 0.004709375858306885, 0.00470531177520752, 0.004685791969299316, 0.00470633602142334, 0.004689888000488281, 0.004712448120117187, 0.004677631855010986, 0.004677631855010986, 0.004678656101226806, 0.004689919948577881, 0.004676640033721924, 0.004706272125244141, 0.004686848163604736, 0.004695040225982666, 0.004690944194793701, 0.004682752132415771, 0.004695040225982666, 0.0046991357803344725, 0.004680704116821289, 0.004675583839416504, 0.004689919948577881, 0.004682752132415771, 0.004695040225982666, 0.004686848163604736, 0.0046929922103881834, 0.0046991357803344725, 0.004673567771911621, 0.00477180814743042, 0.00470630407333374, 0.00467251205444336, 0.004680704116821289, 0.004784128189086914, 0.004687871932983399, 0.004729856014251709, 0.004707327842712402, 0.004702208042144776, 0.004694143772125244, 0.004705152034759521, 0.004686848163604736, 0.0047185921669006346, 0.00471347188949585, 0.004712448120117187, 0.0046929922103881834, 0.0046991357803344725, 0.004682752132415771, 0.004711423873901367, 0.010620927810668946, 0.004698112010955811, 0.004701183795928955, 0.00470630407333374, 0.004683775901794434, 0.004685823917388916, 0.004684800148010254, 0.004696063995361328, 0.004724736213684082, 0.004690944194793701, 0.004698112010955811, 0.004684800148010254, 0.004686848163604736, 0.0047226881980895995, 0.004679679870605469, 0.004680704116821289, 0.004683775901794434, 0.004766719818115234, 0.004758528232574463, 0.004769792079925537, 0.004777984142303467, 0.00480460786819458, 0.004682752132415771, 0.004674560070037842, 0.004735007762908936, 0.0046806721687316895, 0.004733952045440673, 0.004700160026550293, 0.004708352088928222, 0.004687935829162598, 0.004704192161560059, 0.004700160026550293, 0.004702208042144776, 0.004683775901794434, 0.004652031898498535, 0.004664351940155029, 0.004706272125244141, 0.004684800148010254, 0.004672544002532959, 0.004675551891326905, 0.004674560070037842, 0.004673535823822021, 0.004716544151306152, 0.00469708776473999, 0.004681727886199951, 0.004683775901794434, 0.004676608085632325, 0.004709375858306885, 0.004673535823822021, 0.00469708776473999, 0.004667424201965332, 0.004667359828948975, 0.0046888961791992185, 0.0047554559707641605, 0.004682752132415771, 0.004690944194793701, 0.004686848163604736, 0.004685823917388916, 0.004677631855010986, 0.004698112010955811, 0.004687871932983399, 0.00468175983428955, 0.004668384075164795, 0.010424320220947265, 0.004699168205261231, 0.00467964792251587, 0.0046888961791992185, 0.004686848163604736, 0.004712448120117187, 0.004694015979766846, 0.0046919679641723635, 0.004678656101226806, 0.004683775901794434, 0.004680704116821289, 0.004696063995361328, 0.004676608085632325, 0.004731904029846192, 0.004689919948577881, 0.00469708776473999, 0.0046919679641723635, 0.004704256057739258, 0.004710400104522705, 0.0046910080909729, 0.004700096130371094, 0.004680704116821289, 0.004738048076629638, 0.0046929922103881834, 0.004696063995361328, 0.004686848163604736, 0.004694015979766846, 0.004689919948577881, 0.004702208042144776, 0.0046888961791992185, 0.0046929922103881834, 0.004701183795928955, 0.004675583839416504, 0.004681727886199951, 0.004683775901794434, 0.00469708776473999, 0.0046888961791992185, 0.004679679870605469, 0.005018623828887939, 0.004877312183380127, 0.004807680130004883, 0.004819968223571777, 0.004791296005249023, 0.004791296005249023, 0.004825088024139404, 0.004696063995361328, 0.004702208042144776, 0.004685823917388916, 0.004671487808227539, 0.004674560070037842, 0.004690944194793701, 0.0046919679641723635, 0.00470630407333374, 0.004670464038848877, 0.004704256057739258, 0.004679679870605469, 0.004677631855010986, 0.004682752132415771, 0.004684800148010254, 0.004684800148010254, 0.00467251205444336, 0.004693056106567383, 0.004682688236236572, 0.01042739200592041, 0.00467251205444336, 0.0046991357803344725, 0.0046919679641723635, 0.0046694397926330565, 0.004682752132415771, 0.004675583839416504, 0.004684800148010254, 0.004701183795928955, 0.004687935829162598, 0.004684735774993896, 0.004671520233154297, 0.00467964792251587, 0.0046929922103881834, 0.004684800148010254, 0.004676608085632325, 0.004686848163604736, 0.004673535823822021, 0.004684800148010254, 0.004731904029846192, 0.004683775901794434, 0.004701183795928955, 0.0046929922103881834, 0.004747263908386231, 0.004700160026550293, 0.004696063995361328, 0.004690944194793701, 0.00469708776473999, 0.004689919948577881, 0.004689919948577881, 0.004711423873901367, 0.0047329277992248535, 0.004712448120117187, 0.004694015979766846, 0.004709375858306885, 0.0046919679641723635, 0.0047185921669006346, 0.004686848163604736, 0.004715519905090332, 0.004707327842712402, 0.004711423873901367, 0.004700160026550293, 0.004704256057739258, 0.004781055927276612, 0.004710400104522705, 0.0046919679641723635, 0.004694015979766846, 0.004678656101226806, 0.00470527982711792, 0.004685823917388916, 0.004681727886199951, 0.004671487808227539, 0.004673535823822021, 0.004700160026550293, 0.004698112010955811, 0.004719615936279297, 0.004687871932983399, 0.004696063995361328, 0.004687871932983399, 0.00470630407333374, 0.004683775901794434, 0.004683775901794434, 0.0047001919746398925, 0.01039459228515625, 0.004685823917388916, 0.004671487808227539, 0.004682752132415771, 0.004668416023254395, 0.0046919679641723635, 0.004700160026550293, 0.004684800148010254, 0.004680704116821289, 0.004689919948577881, 0.004686848163604736, 0.004708352088928222, 0.004679679870605469, 0.004683775901794434, 0.004684800148010254, 0.004707327842712402, 0.004678656101226806, 0.0046919679641723635, 0.004673535823822021, 0.004685823917388916, 0.004892672061920166, 0.004717567920684815, 0.00470527982711792, 0.004687871932983399, 0.004712448120117187, 0.00469708776473999, 0.004692031860351563, 0.004689856052398682, 0.0047216639518737795, 0.004683775901794434, 0.00470527982711792, 0.004684800148010254, 0.004681727886199951, 0.004686848163604736, 0.004696063995361328, 0.0046991357803344725, 0.004696063995361328, 0.004686848163604736, 0.004681727886199951, 0.004671487808227539, 0.0046929922103881834, 0.00469708776473999, 0.004687871932983399, 0.004676608085632325, 0.004679679870605469, 0.004669472217559815, 0.004708320140838623, 0.004685823917388916, 0.004667391777038574, 0.004704288005828857, 0.004675551891326905, 0.004715519905090332, 0.004680704116821289, 0.004696063995361328, 0.004687871932983399, 0.004685823917388916, 0.004698112010955811, 0.004712448120117187, 0.004686848163604736, 0.004686848163604736, 0.004674560070037842, 0.004671487808227539, 0.004677631855010986, 0.01043660831451416, 0.004679679870605469, 0.004694015979766846, 0.0047084159851074215, 0.0046837120056152344, 0.004699168205261231, 0.004675551891326905, 0.004684800148010254, 0.004673535823822021, 0.004689919948577881, 0.004667391777038574, 0.0046888961791992185, 0.004681727886199951, 0.00469708776473999, 0.004686848163604736, 0.004698112010955811, 0.0046991357803344725, 0.004701183795928955, 0.00469708776473999, 0.004676608085632325, 0.004680704116821289, 0.004716544151306152, 0.004700160026550293, 0.0046929922103881834, 0.004689919948577881, 0.004687871932983399, 0.0047125439643859865, 0.004693920135498047, 0.004704256057739258, 0.004678656101226806, 0.004701183795928955, 0.004682752132415771, 0.004723743915557861, 0.004688864231109619, 0.0046919679641723635, 0.0046888961791992185, 0.004708352088928222, 0.004690976142883301, 0.004704224109649658, 0.0046991357803344725, 0.004709375858306885, 0.004694015979766846, 0.004709375858306885, 0.004686848163604736, 0.0047216639518737795, 0.004684800148010254, 0.004696063995361328, 0.004696063995361328, 0.0046919679641723635, 0.004733952045440673, 0.004686848163604736, 0.0046919679641723635, 0.0046929922103881834, 0.00470527982711792, 0.004681727886199951, 0.004720672130584717, 0.004699168205261231, 0.004696000099182129, 0.0046919679641723635, 0.00470527982711792, 0.004677631855010986, 0.004741119861602783, 0.004684800148010254, 0.010416128158569337, 0.004698112010955811, 0.004675583839416504, 0.004678656101226806, 0.004687871932983399, 0.004674560070037842, 0.00467251205444336, 0.004679679870605469, 0.004668416023254395, 0.004687871932983399, 0.004683775901794434, 0.004670464038848877, 0.004661248207092285, 0.0046888961791992185, 0.004670464038848877, 0.004694015979766846, 0.004673535823822021, 0.004679679870605469, 0.004677631855010986, 0.004700160026550293, 0.004675583839416504, 0.004680704116821289, 0.004673535823822021, 0.0046888961791992185, 0.004675583839416504, 0.004701183795928955, 0.004720640182495117, 0.004702208042144776, 0.0046694397926330565, 0.004681727886199951, 0.004715519905090332, 0.004684800148010254, 0.004671487808227539, 0.004677631855010986, 0.004680704116821289, 0.00468998384475708, 0.0047103362083435055, 0.004685823917388916, 0.004685823917388916, 0.004683775901794434, 0.004690944194793701, 0.004671487808227539, 0.004738048076629638, 0.004726784229278564, 0.004698112010955811, 0.004690944194793701, 0.004710400104522705, 0.004674623966217041, 0.004736959934234619, 0.004685823917388916, 0.004677631855010986, 0.004679679870605469, 0.004690944194793701, 0.004687871932983399, 0.004695040225982666, 0.0046888961791992185, 0.004680704116821289, 0.004691999912261963, 0.004701151847839356, 0.004733952045440673, 0.0047288317680358885, 0.004709375858306885, 0.004695040225982666, 0.01041305637359619, 0.004694015979766846, 0.004687871932983399, 0.004682752132415771, 0.004673535823822021, 0.00470630407333374, 0.004674560070037842, 0.00469708776473999, 0.004679679870605469, 0.004682752132415771, 0.004687871932983399, 0.0046919679641723635, 0.0046919679641723635, 0.004687871932983399, 0.004680704116821289, 0.00467251205444336, 0.004670464038848877, 0.004707327842712402, 0.004683775901794434, 0.004695104122161865, 0.004684735774993896, 0.004683839797973633, 0.00467142391204834, 0.004717567920684815, 0.004715519905090332, 0.004765696048736572, 0.004694015979766846, 0.004681727886199951, 0.004708352088928222, 0.00469708776473999, 0.004681727886199951, 0.00471347188949585, 0.004685823917388916, 0.004677631855010986, 0.0046929922103881834, 0.004731904029846192, 0.004719615936279297, 0.004703231811523437, 0.004702208042144776, 0.004717631816864014, 0.004703167915344238, 0.004996096134185791, 0.0047820801734924315, 0.004811776161193848, 0.004738048076629638, 0.004708352088928222, 0.004689919948577881, 0.004680704116821289, 0.004680704116821289, 0.0046694397926330565, 0.004698112010955811, 0.004677631855010986, 0.0046888961791992185, 0.004682752132415771, 0.004686848163604736, 0.004694015979766846, 0.0047226881980895995, 0.0046929922103881834, 0.004694047927856445, 0.004706272125244141, 0.004677631855010986, 0.004667391777038574, 0.004675583839416504, 0.010481663703918457, 0.004743167877197266, 0.004704256057739258, 0.004701183795928955, 0.004679679870605469, 0.004655104160308838, 0.004683775901794434, 0.004679679870605469, 0.004685823917388916, 0.004678656101226806, 0.0047185921669006346, 0.004696063995361328, 0.004686848163604736, 0.00466431999206543, 0.004700160026550293, 0.0046888961791992185, 0.004696063995361328, 0.0046694397926330565, 0.00469708776473999, 0.004682752132415771, 0.004720640182495117, 0.004703231811523437, 0.004694015979766846, 0.004689919948577881, 0.004678656101226806, 0.004700160026550293, 0.004685823917388916, 0.00470527982711792, 0.004737023830413818, 0.004719615936279297, 0.004709375858306885, 0.0047185921669006346, 0.004711423873901367, 0.004723711967468262, 0.0047185921669006346, 0.004704256057739258, 0.004683775901794434, 0.004734975814819336, 0.004698112010955811, 0.004689919948577881, 0.004689919948577881, 0.004711423873901367, 0.004715519905090332, 0.004701183795928955, 0.0046991357803344725, 0.004698112010955811, 0.004694015979766846, 0.00469708776473999, 0.004700160026550293, 0.004698112010955811, 0.004711423873901367, 0.004711423873901367, 0.004680704116821289, 0.0047226881980895995, 0.0047216639518737795, 0.004686848163604736, 0.004703231811523437, 0.004723711967468262, 0.004701183795928955, 0.0047226881980895995, 0.004742144107818603, 0.0047216639518737795, 0.00470527982711792, 0.01046121597290039, 0.004692959785461426, 0.004690944194793701, 0.004678656101226806, 0.004676608085632325, 0.004738048076629638, 0.004668416023254395, 0.004663360118865967, 0.004677567958831787, 0.00467251205444336, 0.004675583839416504, 0.00470527982711792, 0.004685823917388916, 0.0046888961791992185, 0.004702208042144776, 0.004682752132415771, 0.004680704116821289, 0.00469708776473999, 0.004687871932983399, 0.00469814395904541, 0.004688864231109619, 0.004676608085632325, 0.004674560070037842, 0.004703231811523437, 0.0046888961791992185, 0.004681727886199951, 0.0046888961791992185, 0.004689919948577881, 0.0047288317680358885, 0.004724736213684082, 0.004701183795928955, 0.004700160026550293, 0.004704256057739258, 0.00470630407333374, 0.004703231811523437, 0.004712448120117187, 0.004716544151306152, 0.004686912059783935, 0.004684735774993896, 0.004686848163604736, 0.00471449613571167, 0.004702271938323975, 0.004687808036804199, 0.004655168056488037, 0.004678592205047607, 0.004670464038848877, 0.0046919679641723635, 0.004684800148010254, 0.004680704116821289, 0.004700160026550293, 0.00469708776473999, 0.0046929922103881834, 0.004680704116821289, 0.004689919948577881, 0.0046919679641723635, 0.004680768013000488, 0.004683743953704834, 0.004704224109649658, 0.004677631855010986, 0.004682752132415771, 0.004689919948577881, 0.004674560070037842, 0.004717567920684815, 0.010430463790893555, 0.004687871932983399, 0.004675583839416504, 0.00471449613571167, 0.004682752132415771, 0.004678656101226806, 0.004683775901794434, 0.004678656101226806, 0.004675583839416504, 0.0046991357803344725, 0.004673535823822021, 0.004674560070037842, 0.004695040225982666, 0.004670464038848877, 0.004681727886199951, 0.004716544151306152, 0.0046704959869384765, 0.004743135929107666, 0.004686880111694336, 0.004685791969299316, 0.0046581759452819825, 0.0047185921669006346, 0.004676608085632325, 0.004687871932983399, 0.004676608085632325, 0.0046888961791992185, 0.004681727886199951, 0.004713503837585449, 0.004692959785461426, 0.004686880111694336, 0.004734943866729737, 0.0046991357803344725, 0.004707327842712402, 0.004679679870605469, 0.004695040225982666, 0.00467251205444336, 0.004675583839416504, 0.004679679870605469, 0.004696063995361328, 0.004676608085632325, 0.004673535823822021, 0.00469708776473999, 0.004726784229278564, 0.004678656101226806, 0.004677631855010986, 0.004682752132415771, 0.00470527982711792, 0.004683839797973633, 0.004682688236236572, 0.004670464038848877, 0.004708352088928222, 0.004703231811523437, 0.004696063995361328, 0.004677631855010986, 0.004696063995361328, 0.004674560070037842, 0.004711423873901367, 0.004671487808227539, 0.004677631855010986, 0.004682752132415771, 0.004667391777038574, 0.0049920320510864254, 0.004723680019378662, 0.010469375610351562, 0.004708352088928222, 0.004672544002532959, 0.004674560070037842, 0.004669407844543457, 0.004683807849884033, 0.00466431999206543, 0.004695007801055908, 0.004666368007659912, 0.004684800148010254, 0.00466534423828125, 0.004686848163604736, 0.004680704116821289, 0.0046929922103881834, 0.004667391777038574, 0.004686848163604736, 0.004675583839416504, 0.004679679870605469, 0.004684800148010254, 0.004700160026550293, 0.004707327842712402, 0.004704256057739258, 0.00469708776473999, 0.004687871932983399, 0.004673535823822021, 0.004921343803405762, 0.004710400104522705, 0.004712448120117187, 0.004746240139007568, 0.004986879825592041, 0.005346303939819336, 0.004842495918273926, 0.004809728145599365, 0.004803584098815918, 0.00486297607421875, 0.00479744005203247, 0.004696063995361328, 0.0046919679641723635, 0.004715583801269531, 0.004734911918640137, 0.004702208042144776, 0.004709375858306885, 0.004681727886199951, 0.004687871932983399, 0.0047257599830627444, 0.00470527982711792, 0.004684800148010254, 0.004667391777038574, 0.004672544002532959, 0.00466428804397583, 0.004671487808227539, 0.004683807849884033, 0.004665311813354492, 0.004677631855010986, 0.004666368007659912, 0.004682752132415771, 0.004684800148010254, 0.004716544151306152, 0.0046929922103881834, 0.0046919679641723635, 0.004683775901794434, 0.00467251205444336, 0.004690944194793701, 0.010534912109375, 0.004677631855010986, 0.00466431999206543, 0.004704256057739258, 0.004679679870605469, 0.004676608085632325, 0.004666368007659912, 0.0046694397926330565, 0.004680704116821289, 0.004717599868774414, 0.004698080062866211, 0.0046888961791992185, 0.004671487808227539, 0.0046919679641723635, 0.004710400104522705, 0.004671487808227539, 0.004677631855010986, 0.004675583839416504, 0.004684800148010254, 0.004666368007659912, 0.004709375858306885, 0.004932608127593994, 0.00470630407333374, 0.004670464038848877, 0.004673535823822021, 0.00467251205444336, 0.004690944194793701, 0.004682752132415771, 0.004727807998657227, 0.004684800148010254, 0.004678656101226806, 0.004718624114990235, 0.00467964792251587, 0.004668416023254395, 0.004671487808227539, 0.004740096092224121, 0.004678656101226806, 0.004711423873901367, 0.004674560070037842, 0.004703231811523437, 0.0046929922103881834, 0.00469708776473999, 0.004680704116821289, 0.004698112010955811, 0.00466534423828125, 0.0046694397926330565, 0.004678656101226806, 0.004666368007659912, 0.004667391777038574, 0.004707327842712402, 0.004690944194793701, 0.004684800148010254, 0.004666368007659912, 0.004681727886199951, 0.004696063995361328, 0.004715519905090332, 0.004709375858306885, 0.0046929922103881834, 0.004689919948577881, 0.004674560070037842, 0.004704256057739258, 0.004685823917388916, 0.004681727886199951, 0.010402815818786621, 0.00470527982711792, 0.004695040225982666, 0.004668447971343994, 0.004684768199920654, 0.004684800148010254, 0.004687871932983399, 0.004680704116821289, 0.004675583839416504, 0.00467251205444336, 0.004677631855010986, 0.004678656101226806, 0.0046991357803344725, 0.004681727886199951, 0.004701183795928955, 0.004690944194793701, 0.0046694397926330565, 0.004678656101226806, 0.004711423873901367, 0.0046674561500549315, 0.004687808036804199, 0.004674560070037842, 0.00466534423828125, 0.004668416023254395, 0.004695040225982666, 0.004682752132415771, 0.004694015979766846, 0.004700160026550293, 0.004690944194793701, 0.004678656101226806, 0.0046991357803344725, 0.0046888961791992185, 0.004677631855010986, 0.005040128231048584, 0.004827136039733886, 0.004811776161193848, 0.004884479999542236, 0.004869120121002197, 0.004808703899383545, 0.0051476478576660155, 0.005288959980010987, 0.0049725441932678225, 0.004731904029846192, 0.0046929922103881834, 0.004754432201385498, 0.004720640182495117, 0.0046888961791992185, 0.00466431999206543, 0.00471449613571167, 0.0047606401443481446, 0.0046796159744262695, 0.004685823917388916, 0.00471449613571167, 0.004683775901794434, 0.004687903881072998, 0.004676576137542725, 0.004682752132415771, 0.004676608085632325, 0.004681727886199951, 0.004677631855010986, 0.004687871932983399, 0.004678688049316406, 0.004670432090759277, 0.011083776473999024, 0.0048496642112731934, 0.004704256057739258, 0.004698112010955811, 0.00470630407333374, 0.004703231811523437, 0.004721695899963379, 0.004697055816650391, 0.004674560070037842, 0.004687871932983399, 0.0046919679641723635, 0.0047216639518737795, 0.004680704116821289, 0.00468998384475708, 0.004698048114776612, 0.004716544151306152, 0.004674623966217041, 0.004868031978607178, 0.004695040225982666, 0.004708352088928222, 0.004621312141418457, 0.004655104160308838, 0.004626431941986084, 0.0046592001914978025, 0.004641791820526123, 0.00464896011352539, 0.004668416023254395, 0.00465715217590332, 0.00467251205444336, 0.0047216639518737795, 0.004720640182495117, 0.004695040225982666, 0.004678656101226806, 0.004702208042144776, 0.004686848163604736, 0.004710400104522705, 0.004690944194793701, 0.004680704116821289, 0.004689919948577881, 0.0046888961791992185, 0.00470527982711792, 0.00471347188949585, 0.0046929922103881834, 0.004687871932983399, 0.004702208042144776, 0.004687871932983399, 0.004708352088928222, 0.0046888961791992185, 0.004704256057739258, 0.004695040225982666, 0.0046991357803344725, 0.004702208042144776, 0.004724736213684082, 0.004757503986358643, 0.004708352088928222, 0.004680704116821289, 0.004668416023254395, 0.004686848163604736, 0.0047257599830627444, 0.004703231811523437, 0.004675583839416504, 0.00470630407333374, 0.004678656101226806, 0.01054207992553711, 0.004671487808227539, 0.0047185921669006346, 0.004683839797973633, 0.0046908798217773435, 0.004686848163604736, 0.004673535823822021, 0.00468281602859497, 0.0046837120056152344, 0.004701183795928955, 0.0046694397926330565, 0.00470630407333374, 0.004687871932983399, 0.004666368007659912, 0.00466431999206543, 0.0046919679641723635, 0.004674560070037842, 0.004681727886199951, 0.004680704116821289, 0.004679743766784668, 0.004676544189453125, 0.004698112010955811, 0.004677631855010986, 0.004703231811523437, 0.004678656101226806, 0.0047001919746398925, 0.004693984031677246, 0.00471449613571167, 0.004687871932983399, 0.004727807998657227, 0.004686848163604736, 0.004684800148010254, 0.004679679870605469, 0.004695040225982666, 0.004737023830413818, 0.004687871932983399, 0.004670464038848877, 0.004678656101226806, 0.004700160026550293, 0.004734975814819336, 0.004684800148010254, 0.004684800148010254, 0.004667391777038574, 0.004685823917388916, 0.00471449613571167, 0.00467251205444336, 0.004670464038848877, 0.004673535823822021, 0.004660223960876465, 0.004660223960876465, 0.004700160026550293, 0.004666368007659912, 0.004675583839416504, 0.004700160026550293, 0.004685823917388916, 0.004695104122161865, 0.004699071884155273, 0.0046919679641723635, 0.00470527982711792, 0.004687935829162598, 0.004696063995361328, 0.004675519943237305, 0.004700160026550293, 0.010424320220947265, 0.004687871932983399, 0.004679679870605469, 0.004717567920684815, 0.004702208042144776, 0.004673535823822021, 0.004673567771911621, 0.004660192012786865, 0.004704256057739258, 0.004678656101226806, 0.00469708776473999, 0.004687871932983399, 0.004694047927856445, 0.004684768199920654, 0.004712448120117187, 0.004703231811523437, 0.004680704116821289, 0.004686848163604736, 0.004670464038848877, 0.004711423873901367, 0.004675615787506104, 0.004709343910217285, 0.004690944194793701, 0.004739071846008301, 0.004694015979766846, 0.004715583801269531, 0.004691904067993164, 0.00470527982711792, 0.004685823917388916, 0.004689919948577881, 0.004689919948577881, 0.004698112010955811, 0.004679679870605469, 0.004673600196838379, 0.0046908798217773435, 0.00469708776473999, 0.004711423873901367, 0.004700160026550293, 0.004700160026550293, 0.004675583839416504, 0.0046581759452819825, 0.0046581759452819825, 0.004642816066741944, 0.00466431999206543, 0.004655104160308838, 0.004681727886199951, 0.004684800148010254, 0.004687871932983399, 0.004727807998657227, 0.004837376117706299, 0.004679679870605469, 0.004690944194793701, 0.004678656101226806, 0.004666368007659912, 0.004686848163604736, 0.004674560070037842, 0.004690944194793701, 0.004694015979766846, 0.004684832096099854, 0.004665311813354492, 0.004851712226867676, 0.004775936126708984, 0.004679679870605469, 0.010490880012512208, 0.00469708776473999, 0.004686848163604736, 0.004673535823822021, 0.0047185921669006346, 0.004681727886199951, 0.004695040225982666, 0.004677631855010986, 0.00466431999206543, 0.004676608085632325, 0.004667391777038574, 0.004702208042144776, 0.004695040225982666, 0.004707327842712402, 0.004709375858306885, 0.0046929922103881834, 0.004790272235870361, 0.004701183795928955, 0.004694015979766846, 0.004684800148010254, 0.004679679870605469, 0.0046694397926330565, 0.004668447971343994, 0.004705247879028321, 0.004683775901794434, 0.004850687980651855, 0.004783103942871094, 0.004711423873901367, 0.004833280086517334, 0.0050728960037231445, 0.005371903896331787, 0.004867072105407715, 0.004851712226867676, 0.00481279993057251, 0.0048056321144104, 0.004820991992950439, 0.004770815849304199, 0.004766719818115234, 0.004716544151306152, 0.004701183795928955, 0.004687871932983399, 0.0046991357803344725, 0.004876287937164306, 0.004841472148895264, 0.004847616195678711, 0.004884479999542236, 0.004740096092224121, 0.004838399887084961, 0.0047554559707641605, 0.004966400146484375, 0.00481382417678833, 0.004712448120117187, 0.004730879783630371, 0.004701183795928955, 0.004690944194793701, 0.004712448120117187, 0.004743199825286865, 0.004686848163604736, 0.004720608234405518, 0.00469708776473999, 0.0046919679641723635, 0.004686848163604736, 0.0046888961791992185, 0.010582015991210938, 0.004707327842712402, 0.004676608085632325, 0.0046991357803344725, 0.004709375858306885, 0.004791296005249023, 0.004694015979766846, 0.004690944194793701, 0.004679679870605469, 0.004708352088928222, 0.004702208042144776, 0.00470630407333374, 0.004680704116821289, 0.004678656101226806, 0.004673535823822021, 0.0046929922103881834, 0.004666368007659912, 0.004679679870605469, 0.004677631855010986, 0.004670464038848877, 0.004671487808227539, 0.004708352088928222, 0.004703231811523437, 0.0046929922103881834, 0.004679679870605469, 0.004700160026550293, 0.004701183795928955, 0.0047513599395751956, 0.0047012481689453125, 0.0046888961791992185, 0.0047021441459655765, 0.004687871932983399, 0.004709375858306885, 0.004683775901794434, 0.0046919679641723635, 0.004694015979766846, 0.004694015979766846, 0.004694015979766846, 0.0046929922103881834, 0.004707327842712402, 0.00470527982711792, 0.004690944194793701, 0.0047288317680358885, 0.004689919948577881, 0.004730879783630371, 0.004702208042144776, 0.004701183795928955, 0.004677631855010986, 0.004690944194793701, 0.004694079875946045, 0.004694975852966308, 0.00470527982711792, 0.004700160026550293, 0.0046929922103881834, 0.004707327842712402, 0.004731904029846192, 0.0046991357803344725, 0.004687871932983399, 0.004707327842712402, 0.004696063995361328, 0.0046919679641723635, 0.004719615936279297, 0.00471347188949585, 0.010447872161865235, 0.004708352088928222, 0.004771840095520019, 0.004695040225982666, 0.004682752132415771, 0.0046888961791992185, 0.00468998384475708, 0.004712384223937988, 0.004706367969512939, 0.004699071884155273, 0.00466534423828125, 0.004680704116821289, 0.004719615936279297, 0.004676608085632325, 0.004696063995361328, 0.004687871932983399, 0.004673535823822021, 0.0047329277992248535, 0.004710400104522705, 0.004709375858306885, 0.004712448120117187, 0.004702208042144776, 0.004677631855010986, 0.004684800148010254, 0.00469708776473999, 0.004684800148010254, 0.0046929922103881834, 0.00470531177520752, 0.004707295894622803, 0.004707327842712402, 0.004712448120117187, 0.004716544151306152, 0.0046919679641723635, 0.0046991357803344725, 0.004696063995361328, 0.0047226881980895995, 0.004695072174072266, 0.004700128078460693, 0.0046991357803344725, 0.004720640182495117, 0.004704256057739258, 0.004703231811523437, 0.004701183795928955, 0.004708352088928222, 0.004700160026550293, 0.004680704116821289, 0.004687871932983399, 0.004694015979766846, 0.004680704116821289, 0.004678656101226806, 0.004675583839416504, 0.004687871932983399, 0.004663296222686767, 0.004698112010955811, 0.0046888961791992185, 0.004696063995361328, 0.004678656101226806, 0.004701183795928955, 0.004833280086517334, 0.0047861762046813965, 0.0047820801734924315, 0.004774911880493164, 0.0047851519584655765, 0.010476544380187988, 0.004695040225982666, 0.004681727886199951, 0.004694015979766846, 0.0046991357803344725, 0.00469708776473999, 0.004695040225982666, 0.0046888961791992185, 0.004694015979766846, 0.004709375858306885, 0.004696063995361328, 0.004719647884368896, 0.004704224109649658, 0.00466431999206543, 0.004673535823822021, 0.00470527982711792, 0.004678656101226806, 0.00467251205444336, 0.0046888961791992185, 0.004683775901794434, 0.004698112010955811, 0.0046929922103881834, 0.004690944194793701, 0.0046929922103881834, 0.004695040225982666, 0.0046888961791992185, 0.004695040225982666, 0.004720640182495117, 0.004717631816864014, 0.0047103681564331056, 0.004681695938110351, 0.004689919948577881, 0.004700160026550293, 0.004690944194793701, 0.004696063995361328, 0.0046356477737426755, 0.004641791820526123, 0.004651008129119873, 0.004653056144714355, 0.00464896011352539, 0.004766719818115234, 0.00471449613571167, 0.004709375858306885, 0.0046888961791992185, 0.004740096092224121, 0.004684800148010254, 0.004704256057739258, 0.004687871932983399, 0.004766719818115234, 0.004807680130004883, 0.004771840095520019, 0.004667391777038574, 0.004627456188201904, 0.00464793586730957, 0.00466534423828125, 0.004631552219390869, 0.0047226881980895995, 0.0049192957878112795, 0.004799488067626953, 0.004689919948577881, 0.004702208042144776, 0.004724736213684082, 0.004711423873901367, 0.010523648262023925, 0.004712448120117187, 0.0046991357803344725, 0.004678656101226806, 0.004682752132415771, 0.00469708776473999, 0.004703231811523437, 0.004717567920684815, 0.004686848163604736, 0.004676608085632325, 0.004678656101226806, 0.004680704116821289, 0.004723711967468262, 0.004700160026550293, 0.004702208042144776, 0.004684800148010254, 0.004678656101226806, 0.004677631855010986, 0.004726784229278564, 0.004819968223571777, 0.0047636480331420894, 0.004673567771911621, 0.00467964792251587, 0.004670464038848877, 0.0047185921669006346, 0.004697152137756348, 0.0047134079933166505, 0.004700160026550293, 0.004727807998657227, 0.004711455821990966, 0.004696032047271728, 0.0046929922103881834, 0.004695040225982666, 0.0046991357803344725, 0.004681727886199951, 0.00470630407333374, 0.004689919948577881, 0.004689919948577881, 0.004683775901794434, 0.004677631855010986, 0.004682752132415771, 0.004681727886199951, 0.004682752132415771, 0.004708352088928222, 0.0046888961791992185, 0.004682752132415771, 0.004666368007659912, 0.004723711967468262, 0.004684800148010254, 0.004676608085632325, 0.00466431999206543, 0.004663296222686767, 0.004695072174072266, 0.004697055816650391, 0.004683775901794434, 0.004674560070037842, 0.00485584020614624, 0.004669407844543457, 0.004727807998657227, 0.004678656101226806, 0.004694015979766846, 0.004685823917388916, 0.004698112010955811, 0.010440704345703124, 0.004686848163604736, 0.004678656101226806, 0.004677631855010986, 0.004700160026550293, 0.004689919948577881, 0.004675583839416504, 0.0050954241752624516, 0.004723711967468262, 0.004709407806396484, 0.004693984031677246, 0.004684800148010254, 0.004681727886199951, 0.004671487808227539, 0.00467251205444336, 0.004707327842712402, 0.004673535823822021, 0.004668416023254395, 0.004678656101226806, 0.004661248207092285, 0.004668416023254395, 0.004700160026550293, 0.004673535823822021, 0.004681727886199951, 0.004674560070037842, 0.004675583839416504, 0.004689919948577881, 0.004827136039733886, 0.0046929922103881834, 0.00471347188949585, 0.004687871932983399, 0.004709375858306885, 0.004681727886199951, 0.004710400104522705, 0.00471449613571167, 0.004707327842712402, 0.004710400104522705, 0.00470630407333374, 0.004709375858306885, 0.004718624114990235, 0.004692959785461426, 0.004702208042144776, 0.004671487808227539, 0.004753407955169678, 0.005144576072692871, 0.004781055927276612, 0.004680704116821289, 0.004670464038848877, 0.004774911880493164, 0.004696095943450928, 0.00467248010635376, 0.004684800148010254, 0.00468175983428955, 0.004671520233154297, 0.004665279865264892, 0.004729856014251709, 0.004680704116821289, 0.00466534423828125, 0.00468281602859497, 0.004673471927642822, 0.00466534423828125, 0.004711455821990966, 0.0046806721687316895, 0.010470399856567383, 0.0047216639518737795, 0.004689919948577881, 0.004668416023254395, 0.004679679870605469, 0.004743167877197266, 0.004773888111114502, 0.004682752132415771, 0.004799488067626953, 0.00469708776473999, 0.004684800148010254, 0.004685823917388916, 0.004802559852600098, 0.00467251205444336, 0.004689919948577881, 0.004661248207092285, 0.004671487808227539, 0.004679679870605469, 0.004696063995361328, 0.004698112010955811, 0.004673535823822021, 0.004678656101226806, 0.0046694397926330565, 0.004709375858306885, 0.0047226881980895995, 0.0046888961791992185, 0.0046919679641723635, 0.004719615936279297, 0.0046929922103881834, 0.004715519905090332, 0.00470630407333374, 0.004715519905090332, 0.004678656101226806, 0.004685823917388916, 0.004726784229278564, 0.004698112010955811, 0.004684800148010254, 0.004701183795928955, 0.00468281602859497, 0.004706240177154541, 0.004675583839416504, 0.004712448120117187, 0.004695040225982666, 0.00470630407333374, 0.004694047927856445, 0.004696032047271728, 0.004707327842712402, 0.004698112010955811, 0.004738048076629638, 0.004695040225982666, 0.004684800148010254, 0.004673535823822021, 0.004702208042144776, 0.004719615936279297, 0.004681727886199951, 0.0047216639518737795, 0.004682752132415771, 0.004703231811523437, 0.00470630407333374, 0.004748288154602051, 0.004687871932983399, 0.004680704116821289, 0.004689919948577881, 0.010488832473754883, 0.004701183795928955, 0.004684800148010254, 0.004707327842712402, 0.00469708776473999, 0.00470527982711792, 0.004704256057739258, 0.004676608085632325, 0.004677631855010986, 0.004682752132415771, 0.004683775901794434, 0.004631552219390869, 0.004677631855010986, 0.0050094079971313476, 0.00481279993057251, 0.004830207824707031, 0.0046919679641723635, 0.004666368007659912, 0.004670464038848877, 0.004661248207092285, 0.00469708776473999, 0.004679679870605469, 0.004683775901794434, 0.004667391777038574, 0.004676608085632325, 0.004679679870605469, 0.004686848163604736, 0.0047288317680358885, 0.004679679870605469, 0.004741119861602783, 0.004684800148010254, 0.004667391777038574, 0.0046888961791992185, 0.004668416023254395, 0.0048056321144104, 0.004679679870605469, 0.004676608085632325, 0.004701183795928955, 0.0046919679641723635, 0.004676608085632325, 0.004668416023254395, 0.0046694397926330565, 0.00467251205444336, 0.0047226881980895995, 0.004707327842712402, 0.004667391777038574, 0.004668416023254395, 0.004707327842712402, 0.0047216639518737795, 0.00470527982711792, 0.004677631855010986, 0.00469708776473999, 0.004670464038848877, 0.004680704116821289, 0.004634624004364014, 0.004743167877197266, 0.004644864082336426, 0.004630591869354248, 0.004636608123779297, 0.004620287895202637, 0.004633600234985351, 0.004653056144714355, 0.00466534423828125, 0.010447872161865235, 0.004687871932983399, 0.004710400104522705, 0.004690944194793701, 0.004686848163604736, 0.004684800148010254, 0.004719615936279297, 0.004687871932983399, 0.004704256057739258, 0.004694015979766846, 0.004676608085632325, 0.0046919679641723635, 0.004690944194793701, 0.004709375858306885, 0.0046919679641723635, 0.0046888961791992185, 0.0046919679641723635, 0.004673567771911621, 0.0046806721687316895, 0.004687871932983399, 0.004668416023254395, 0.004695040225982666, 0.004677631855010986, 0.004677631855010986, 0.004687871932983399, 0.004707327842712402, 0.004682784080505371, 0.004805600166320801, 0.004690944194793701, 0.004689919948577881, 0.004724736213684082, 0.004695040225982666, 0.004694015979766846, 0.004680704116821289, 0.004701183795928955, 0.004701183795928955, 0.004726784229278564, 0.004699168205261231, 0.004692959785461426, 0.00470630407333374, 0.004696063995361328, 0.004696063995361328, 0.004715519905090332, 0.0046929922103881834, 0.004723711967468262, 0.004709375858306885, 0.004686848163604736, 0.004685823917388916, 0.004759552001953125, 0.00470527982711792, 0.004698112010955811, 0.004686848163604736, 0.004676608085632325, 0.0046991357803344725, 0.004681727886199951, 0.004704256057739258, 0.004686848163604736, 0.004678656101226806, 0.0046929922103881834, 0.0047257599830627444, 0.004694015979766846, 0.004685823917388916, 0.0046592001914978025, 0.010380288124084473, 0.004684800148010254, 0.004671487808227539, 0.004687871932983399, 0.004676608085632325, 0.0046929922103881834, 0.0047216639518737795, 0.004666368007659912, 0.0046919679641723635, 0.004694015979766846, 0.004686848163604736, 0.004710400104522705, 0.004709375858306885, 0.004686848163604736, 0.004679679870605469, 0.004695040225982666, 0.004674560070037842, 0.0047185921669006346, 0.0046929922103881834, 0.004685823917388916, 0.004682784080505371, 0.0046919360160827634, 0.004707327842712402, 0.005293119907379151, 0.00481171178817749, 0.0048230400085449215, 0.004736000061035156, 0.004712448120117187, 0.004675583839416504, 0.00469708776473999, 0.0046919679641723635, 0.004701183795928955, 0.004671487808227539, 0.004701183795928955, 0.004685823917388916, 0.0046991357803344725, 0.004684800148010254, 0.004685823917388916, 0.004709375858306885, 0.004704256057739258, 0.004698112010955811, 0.004670464038848877, 0.004668416023254395, 0.004678688049316406, 0.004696032047271728, 0.004694015979766846, 0.00469708776473999, 0.004687871932983399, 0.004676608085632325, 0.0046991357803344725, 0.004696063995361328, 0.004687871932983399, 0.004670464038848877, 0.00466431999206543, 0.004668416023254395, 0.004667391777038574, 0.004685855865478515, 0.004664351940155029, 0.004673471927642822, 0.004702208042144776, 0.0046581759452819825, 0.004671487808227539, 0.0046929922103881834, 0.010424320220947265, 0.00469708776473999, 0.004712448120117187, 0.004683775901794434, 0.004677631855010986, 0.004677631855010986, 0.0046622719764709475, 0.004676608085632325, 0.004700160026550293, 0.0046694397926330565, 0.004685823917388916, 0.004682752132415771, 0.00468179178237915, 0.004681663990020752, 0.004701183795928955, 0.004695040225982666, 0.0048455681800842285, 0.0046888961791992185, 0.0046991357803344725, 0.004695040225982666, 0.004744192123413086, 0.00469708776473999, 0.004682752132415771, 0.00467251205444336, 0.004666431903839112, 0.0046908798217773435, 0.004679679870605469, 0.004685823917388916, 0.004710464000701904, 0.004664256095886231, 0.004675583839416504, 0.0046991357803344725, 0.004684800148010254, 0.0046694397926330565, 0.004921343803405762, 0.00470630407333374, 0.004670464038848877, 0.004712512016296386, 0.004696000099182129, 0.004700160026550293, 0.0046929922103881834, 0.004680799961090088, 0.004699039936065674, 0.004682752132415771, 0.004705376148223877, 0.00467142391204834, 0.004708320140838623, 0.004682784080505371, 0.0047144641876220705, 0.00471449613571167, 0.004674560070037842, 0.004690944194793701, 0.004676608085632325, 0.004675583839416504, 0.004682752132415771, 0.004695040225982666, 0.004683775901794434, 0.004689919948577881, 0.005478400230407715, 0.004779007911682129, 0.004680704116821289, 0.004671487808227539, 0.004679743766784668, 0.010496959686279297, 0.004704256057739258, 0.004688960075378418, 0.004685760021209717, 0.004684800148010254, 0.004696063995361328, 0.004687871932983399, 0.004676640033721924, 0.0046878399848937985, 0.004675583839416504, 0.004700160026550293, 0.004683775901794434, 0.004683775901794434, 0.00467251205444336, 0.004717567920684815, 0.004696063995361328, 0.00471449613571167, 0.004678656101226806, 0.004684800148010254, 0.004674560070037842, 0.004710464000701904, 0.004676544189453125, 0.0046929922103881834, 0.004666368007659912, 0.0046888961791992185, 0.004667391777038574, 0.004678656101226806, 0.004686848163604736, 0.004719615936279297, 0.004681727886199951, 0.004679679870605469, 0.004685823917388916, 0.004670464038848877, 0.004696063995361328, 0.004696063995361328, 0.004687871932983399, 0.004674560070037842, 0.004682752132415771, 0.0046704959869384765, 0.004683743953704834, 0.004695040225982666, 0.004708352088928222, 0.004690944194793701, 0.004667391777038574, 0.0047257599830627444, 0.004696063995361328, 0.004681727886199951, 0.004761600017547608, 0.004670464038848877, 0.004673535823822021, 0.004698112010955811, 0.004681727886199951, 0.004668416023254395, 0.004685823917388916, 0.004696063995361328, 0.004674560070037842, 0.004680704116821289, 0.004694015979766846, 0.0046929922103881834, 0.004681727886199951, 0.004682752132415771, 0.0046930241584777835, 0.004693984031677246, 0.010465279579162597, 0.004733952045440673, 0.004707327842712402, 0.004774911880493164, 0.004694015979766846, 0.004690944194793701, 0.004682752132415771, 0.004695040225982666, 0.00467251205444336, 0.0046991357803344725, 0.004707327842712402, 0.004690944194793701, 0.004707327842712402, 0.004690944194793701, 0.004717567920684815, 0.004801536083221435, 0.004723711967468262, 0.00470527982711792, 0.004757503986358643, 0.00470527982711792, 0.004747263908386231, 0.004707327842712402, 0.0046929922103881834, 0.004905983924865722, 0.004746240139007568, 0.004709375858306885, 0.004753407955169678, 0.004723711967468262, 0.004717567920684815, 0.004710400104522705, 0.004694015979766846, 0.004717567920684815, 0.004743167877197266, 0.004724736213684082, 0.004677631855010986, 0.0047513599395751956, 0.004712448120117187, 0.004685823917388916, 0.0046888961791992185, 0.004687871932983399, 0.004675583839416504, 0.0046991357803344725, 0.004764671802520752, 0.004690944194793701, 0.004684800148010254, 0.004677631855010986, 0.004673535823822021, 0.004680704116821289, 0.004857855796813965, 0.004773888111114502, 0.004676608085632325, 0.004673535823822021, 0.004682752132415771, 0.004703231811523437, 0.004671487808227539, 0.0047185921669006346, 0.004668416023254395, 0.00467251205444336, 0.004685823917388916, 0.0046991357803344725, 0.004684800148010254, 0.0047350401878356934, 0.004718527793884277]",tokens/s,207.68987718121548,,,main,False,False,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,google/gemma-2b,google/gemma-2b,cuda,0,42,,,True,,,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.223-212.873.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.0,,,,1.21.4,,,,0.12.0,,,gemma,MB,1540.591616,6797.39392,0.0,6150.946816,5419.87328,s,10,0.34272935104370117,0.034272935104370114,0.00016033990741588656,0.034217168807983395,0.03445612640380859,0.034516015625,0.034563927001953126,"[0.03417638397216797, 0.03443254470825195, 0.034114017486572265, 0.03444281768798828, 0.03413734436035156, 0.03425795364379883, 0.034132640838623045, 0.034360992431640626, 0.034575904846191406, 0.03409875106811523]",tokens/s,7469.450726073287,kWh,4.030529538656716e-07,2.2085292515515724e-07,2.0395788026894292e-06,2.6634846817102585e-06,tokens/kWh,96114688.30960914,MB,1619.746816,6797.39392,0.0,6150.946816,5419.87584,s,10,13.906595947265625,1.3906595947265625,0.012992237950615299,1.3940347290039061,1.4009793212890624,1.40474521484375,1.4077579296875,"[1.3932276611328125, 1.3954830322265626, 1.386512451171875, 1.4001424560546876, 1.38755517578125, 1.4085111083984374, 1.3565802001953124, 1.3873514404296876, 1.396390625, 1.394841796875]",tokens/s,45.302243797762266,kWh,1.643336603953662e-05,9.00535192017021e-06,5.3826799223709894e-05,7.926551718341672e-05,tokens/kWh,794797.0597885703,,s,630,13.90490514564516,0.022071278008960544,0.0004617176382976281,0.022253568649291993,0.022417715263366698,0.022550783443450927,0.02311234594345093,"[0.022221824645996095, 0.02234880065917969, 0.02226585578918457, 0.022230016708374024, 0.022253568649291993, 0.022353919982910156, 0.022183935165405275, 0.021407743453979493, 0.02145280075073242, 0.02186751937866211, 0.021811199188232423, 0.02207027244567871, 0.022014976501464844, 0.022412288665771486, 0.022412288665771486, 0.022189056396484375, 0.022034431457519533, 0.022622207641601562, 0.022561792373657227, 0.022362112045288086, 0.02222492790222168, 0.022249439239501952, 0.022000640869140626, 0.022173696517944336, 0.022167552947998048, 0.02206617546081543, 0.022244352340698242, 0.022260736465454102, 0.022412288665771486, 0.022320127487182616, 0.022331392288208008, 0.02311680030822754, 0.02251366424560547, 0.02230681610107422, 0.02234982490539551, 0.022330368041992187, 0.022196224212646484, 0.022205472946166992, 0.022229984283447267, 0.022195199966430663, 0.02149068832397461, 0.02225663948059082, 0.022236160278320313, 0.022445056915283205, 0.022375423431396483, 0.02206208038330078, 0.022296575546264647, 0.02230988883972168, 0.022131711959838866, 0.021275648117065428, 0.02146099281311035, 0.02141900825500488, 0.021424127578735352, 0.02147430419921875, 0.02147942352294922, 0.021445632934570313, 0.021485567092895508, 0.021321727752685548, 0.021403648376464843, 0.02272153663635254, 0.022961151123046874, 0.022350847244262697, 0.022252544403076172, 0.0225218563079834, 0.022303743362426756, 0.02224332809448242, 0.022331392288208008, 0.02211327934265137, 0.02208153533935547, 0.022392831802368163, 0.022328319549560546, 0.022321151733398437, 0.022361087799072265, 0.022328319549560546, 0.022330368041992187, 0.022280191421508787, 0.022364160537719727, 0.022297599792480468, 0.022443008422851563, 0.023173120498657225, 0.022562816619873048, 0.02229555130004883, 0.021489664077758788, 0.021448703765869142, 0.021745664596557617, 0.02227916717529297, 0.021410816192626952, 0.02146099281311035, 0.021563392639160156, 0.022187007904052734, 0.0234833927154541, 0.022611967086791994, 0.022600704193115235, 0.022429695129394533, 0.022378496170043945, 0.022322175979614257, 0.02226688003540039, 0.02150297546386719, 0.021444608688354492, 0.021380096435546874, 0.021396480560302734, 0.02145484733581543, 0.022614015579223632, 0.022163455963134765, 0.022280191421508787, 0.022286336898803712, 0.022176767349243166, 0.022410240173339844, 0.02231091117858887, 0.022238208770751954, 0.022358015060424806, 0.022232063293457033, 0.02229964828491211, 0.021204992294311522, 0.021485567092895508, 0.02145996856689453, 0.021437440872192383, 0.022012928009033202, 0.02229862403869629, 0.022235136032104492, 0.022245376586914063, 0.022296575546264647, 0.022392831802368163, 0.022303743362426756, 0.022361087799072265, 0.02228121566772461, 0.02147635269165039, 0.02149171257019043, 0.02127462387084961, 0.02149478340148926, 0.021497856140136717, 0.021513216018676756, 0.02164121627807617, 0.021440511703491212, 0.021407743453979493, 0.02147737693786621, 0.021415935516357423, 0.021501951217651367, 0.02152038383483887, 0.021348352432250976, 0.02145792007446289, 0.02145075225830078, 0.02145280075073242, 0.02143027114868164, 0.021535743713378908, 0.021436416625976562, 0.021929983139038087, 0.02230784034729004, 0.0223191032409668, 0.02266214370727539, 0.022385663986206054, 0.022468608856201173, 0.022207487106323243, 0.0220897274017334, 0.02126028823852539, 0.02126233673095703, 0.021979135513305666, 0.022335487365722655, 0.022344703674316405, 0.022235136032104492, 0.02205695915222168, 0.022376447677612304, 0.02239897537231445, 0.022467584609985353, 0.022320127487182616, 0.022322175979614257, 0.02234982490539551, 0.022255615234375, 0.022312959671020507, 0.022353919982910156, 0.02235699272155762, 0.022342655181884767, 0.022268928527832032, 0.02235699272155762, 0.022331392288208008, 0.02231603240966797, 0.022370304107666016, 0.022378496170043945, 0.02248294448852539, 0.022330368041992187, 0.02225971221923828, 0.022343679428100584, 0.02227916717529297, 0.022289407730102538, 0.02226380729675293, 0.02229350471496582, 0.022268928527832032, 0.022280191421508787, 0.022260736465454102, 0.022347776412963868, 0.022376447677612304, 0.022237184524536133, 0.0223242244720459, 0.022197248458862305, 0.022717439651489257, 0.022417407989501953, 0.022376447677612304, 0.02230271911621094, 0.02243071937561035, 0.022375423431396483, 0.022312959671020507, 0.022228992462158204, 0.02224844741821289, 0.022405120849609376, 0.022240255355834963, 0.02246246337890625, 0.022360063552856444, 0.022358015060424806, 0.022328319549560546, 0.02234982490539551, 0.022304767608642577, 0.02229350471496582, 0.02231705665588379, 0.022271999359130858, 0.022114303588867186, 0.02222489547729492, 0.022320127487182616, 0.02226585578918457, 0.02233344078063965, 0.02233241653442383, 0.02244915199279785, 0.022393856048583984, 0.02234880065917969, 0.022382591247558595, 0.022320127487182616, 0.022261760711669923, 0.022244352340698242, 0.02227507209777832, 0.022177791595458983, 0.02233241653442383, 0.022566911697387695, 0.022269952774047853, 0.02229964828491211, 0.021400575637817384, 0.02145996856689453, 0.021805055618286134, 0.02229043197631836, 0.02227712059020996, 0.02265088081359863, 0.02267238426208496, 0.022377471923828125, 0.022245376586914063, 0.022252544403076172, 0.022238208770751954, 0.022195199966430663, 0.022200319290161134, 0.021444608688354492, 0.021413888931274414, 0.021421056747436523, 0.021372928619384765, 0.02234060859680176, 0.02142207908630371, 0.021459903717041016, 0.02141900825500488, 0.021249055862426758, 0.021263328552246094, 0.02127462387084961, 0.02143436813354492, 0.021369855880737306, 0.021429248809814453, 0.021405696868896484, 0.02145996856689453, 0.02141798400878906, 0.02150297546386719, 0.022210559844970702, 0.0224399356842041, 0.02223411178588867, 0.021393407821655275, 0.02149068832397461, 0.021445632934570313, 0.02143539237976074, 0.02147430419921875, 0.021501951217651367, 0.02147430419921875, 0.021529600143432616, 0.021597183227539063, 0.02190028762817383, 0.022374399185180666, 0.022542335510253905, 0.023766016006469725, 0.022533119201660155, 0.022379520416259766, 0.022181888580322266, 0.0231014404296875, 0.022381568908691408, 0.02206515121459961, 0.02230681610107422, 0.022326271057128907, 0.021994495391845705, 0.02226688003540039, 0.02222489547729492, 0.0223191032409668, 0.02224127960205078, 0.02226380729675293, 0.022624256134033204, 0.02231091117858887, 0.022220800399780274, 0.02220134353637695, 0.022268928527832032, 0.022215679168701173, 0.02230169677734375, 0.02231603240966797, 0.02226585578918457, 0.02211840057373047, 0.02234060859680176, 0.02225049591064453, 0.02230169677734375, 0.022211584091186523, 0.02142207908630371, 0.02222489547729492, 0.02247372817993164, 0.022355968475341798, 0.02230886459350586, 0.02233344078063965, 0.022232063293457033, 0.02233241653442383, 0.022311935424804686, 0.022288383483886717, 0.02225049591064453, 0.022213632583618165, 0.022154239654541014, 0.022338560104370117, 0.022214656829833986, 0.02225663948059082, 0.02222489547729492, 0.022774784088134766, 0.02247987174987793, 0.022495231628417968, 0.021403648376464843, 0.021519359588623048, 0.02248192024230957, 0.022353919982910156, 0.022408191680908202, 0.022390783309936522, 0.022374399185180666, 0.02229043197631836, 0.02210508728027344, 0.0222873592376709, 0.02243891143798828, 0.022353919982910156, 0.022335519790649416, 0.0223723201751709, 0.022320127487182616, 0.022321151733398437, 0.022353919982910156, 0.022378496170043945, 0.022147071838378905, 0.02224127960205078, 0.02234163284301758, 0.022379520416259766, 0.022426624298095704, 0.022396928787231447, 0.022366207122802736, 0.022738943099975584, 0.022289407730102538, 0.022388736724853517, 0.02225868797302246, 0.02232729530334473, 0.022253568649291993, 0.022342655181884767, 0.022171648025512695, 0.02228121566772461, 0.0222873592376709, 0.022203392028808593, 0.02205286407470703, 0.02227609634399414, 0.0222873592376709, 0.02227097511291504, 0.02229452705383301, 0.022305791854858398, 0.024895488739013674, 0.02309939193725586, 0.022527999877929687, 0.022452224731445314, 0.022372352600097657, 0.022231039047241212, 0.022331392288208008, 0.02227609634399414, 0.02145382308959961, 0.0214466552734375, 0.02142310333251953, 0.02142720031738281, 0.021448703765869142, 0.02145587158203125, 0.021497856140136717, 0.02145280075073242, 0.021428224563598632, 0.02148659133911133, 0.021425151824951173, 0.021445632934570313, 0.021498880386352538, 0.02142617607116699, 0.021573631286621094, 0.021497856140136717, 0.021238784790039062, 0.0214517765045166, 0.021626880645751953, 0.021604352951049805, 0.02151628875732422, 0.02146918487548828, 0.02143948745727539, 0.02147225570678711, 0.02146201515197754, 0.02143846321105957, 0.021398527145385742, 0.021489664077758788, 0.02143129539489746, 0.021401599884033205, 0.02145382308959961, 0.021395456314086913, 0.021420032501220702, 0.021413888931274414, 0.021400575637817384, 0.021403648376464843, 0.021489728927612306, 0.021461952209472657, 0.02146611213684082, 0.021489664077758788, 0.02142617607116699, 0.0214517765045166, 0.021338111877441408, 0.02126540756225586, 0.02127872085571289, 0.02125312042236328, 0.021331968307495116, 0.02122854423522949, 0.02125721549987793, 0.021246976852416992, 0.021507072448730468, 0.0214466552734375, 0.0214517765045166, 0.022395904541015626, 0.02229145622253418, 0.02226483154296875, 0.022002687454223634, 0.022189056396484375, 0.02221772766113281, 0.02227712059020996, 0.021497856140136717, 0.021498880386352538, 0.022074367523193358, 0.022367231369018553, 0.02204876708984375, 0.021990400314331054, 0.022002687454223634, 0.02122547149658203, 0.02122444725036621, 0.02185932731628418, 0.02226380729675293, 0.02230681610107422, 0.022321151733398437, 0.022181888580322266, 0.02250649642944336, 0.022334527969360352, 0.022273984909057618, 0.022288383483886717, 0.02228428840637207, 0.022289407730102538, 0.022313983917236328, 0.02226278305053711, 0.022771711349487304, 0.022329343795776366, 0.02226688003540039, 0.022420480728149415, 0.022365184783935548, 0.022382591247558595, 0.02268880081176758, 0.022457311630249024, 0.022361087799072265, 0.02228326416015625, 0.02230988883972168, 0.022676479339599608, 0.022351871490478514, 0.02229350471496582, 0.02229145622253418, 0.02230067253112793, 0.02230988883972168, 0.02228326416015625, 0.02231603240966797, 0.022176767349243166, 0.0220579833984375, 0.022179840087890625, 0.022025215148925782, 0.02229350471496582, 0.021223424911499023, 0.02145587158203125, 0.02148454475402832, 0.021317632675170898, 0.021300224304199217, 0.021325824737548828, 0.02130534362792969, 0.02145280075073242, 0.02147430419921875, 0.02143129539489746, 0.021522432327270507, 0.02146406364440918, 0.021384191513061524, 0.02148249626159668, 0.021295103073120117, 0.02145792007446289, 0.02146303939819336, 0.021577728271484374, 0.022000640869140626, 0.02322329521179199, 0.02244812774658203, 0.02230271911621094, 0.02226380729675293, 0.022145023345947267, 0.022232063293457033, 0.02234163284301758, 0.022342655181884767, 0.022296575546264647, 0.02232729530334473, 0.02230681610107422, 0.022311935424804686, 0.02246451187133789, 0.022363136291503907, 0.02226380729675293, 0.02224947166442871, 0.02225868797302246, 0.022288383483886717, 0.022358015060424806, 0.0223191032409668, 0.02227916717529297, 0.022260736465454102, 0.02227609634399414, 0.02230681610107422, 0.022207487106323243, 0.022163455963134765, 0.022286336898803712, 0.0223191032409668, 0.022393856048583984, 0.022717439651489257, 0.02271232032775879, 0.022444032669067384, 0.022483968734741212, 0.02231603240966797, 0.02225049591064453, 0.022351871490478514, 0.02234163284301758, 0.022212608337402344, 0.021407743453979493, 0.021432319641113282, 0.02147942352294922, 0.02125721549987793, 0.02126131248474121, 0.021395456314086913, 0.021511167526245118, 0.021411840438842773, 0.021399551391601563, 0.021987327575683592, 0.02226278305053711, 0.02227097511291504, 0.02222591972351074, 0.02224947166442871, 0.022190080642700196, 0.022001663208007814, 0.02225152015686035, 0.022230016708374024, 0.02226585578918457, 0.022199296951293947, 0.022321151733398437, 0.022261760711669923, 0.022271999359130858, 0.022245376586914063, 0.022158336639404298, 0.02206617546081543, 0.021542911529541017, 0.02146713638305664, 0.02142310333251953, 0.021430335998535156, 0.02145478439331055, 0.021448703765869142, 0.021441535949707033, 0.021421056747436523, 0.02125823974609375, 0.025083904266357423, 0.022561792373657227, 0.022391807556152343, 0.02230681610107422, 0.02226483154296875, 0.02226790428161621, 0.022208511352539064, 0.02145484733581543, 0.021393407821655275, 0.022278144836425783, 0.022237184524536133, 0.022286336898803712, 0.022320127487182616, 0.022192127227783204, 0.022012928009033202, 0.022039552688598633, 0.022557695388793944, 0.022323200225830078, 0.022256704330444337, 0.022269887924194334, 0.02247270393371582, 0.022362112045288086, 0.02234060859680176, 0.022193151473999022, 0.02222697639465332, 0.022263776779174804, 0.022228992462158204, 0.02210918426513672, 0.022355968475341798, 0.02233344078063965, 0.022246400833129884, 0.02226380729675293, 0.02220134353637695, 0.022192127227783204, 0.022261760711669923, 0.02227916717529297, 0.02220953559875488, 0.02228121566772461, 0.02225868797302246, 0.022221824645996095, 0.022361087799072265, 0.022252544403076172, 0.022240255355834963, 0.02244095993041992, 0.022271999359130858, 0.02229350471496582, 0.0212490234375, 0.021947391510009767, 0.022393856048583984, 0.021617664337158202, 0.022054912567138672, 0.022171648025512695, 0.02226585578918457, 0.022211584091186523]",tokens/s,45.307752437082165,,,,,,MB,933.752832,6768.033792,0.0,6138.363904,6060.931072,s,1,6.96395263671875,6.96395263671875,0.0,6.96395263671875,6.96395263671875,6.96395263671875,6.96395263671875,[6.96395263671875],,kWh,1.9617644062527185e-06,1.0590836363984596e-06,3.841114184000016e-06,6.861962226651194e-06, float16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,s,s,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/s/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66947374-34b2c09d1124fa2d569010da;302b1a4b-f9e8-43a9-9a89-fc444943510f) Repository Not Found for url: https://huggingface.co/s/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: s is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-40b,tiiuae/falcon-40b,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ self.load_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 250, in load_model_with_no_weights self.pretrained_model = self.automodel_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 559, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3704, in from_pretrained config = cls._autoset_attn_implementation( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1481, in _autoset_attn_implementation cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1572, in _check_and_enable_flash_attn_2 raise ValueError( ValueError: FalconForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmp3zad2fvi/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,google/gemma-7b,google/gemma-7b,cuda,0,42,,,True,,,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.223-212.873.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.0,,,,1.21.4,,,,0.12.0,,,gemma,MB,1541.111808,19300.614144,0.0,18654.16704,17596.665856,s,10,0.9690501785278319,0.09690501785278319,0.0002441175764946301,0.09681404876708985,0.09704476852416992,0.09732038459777832,0.09754087745666504,"[0.09759600067138671, 0.09684259033203126, 0.09675174713134765, 0.0968153305053711, 0.0969835205078125, 0.09677030181884766, 0.09681276702880859, 0.09676316833496093, 0.09697042846679688, 0.09674432373046875]",tokens/s,2641.7620642608185,kWh,1.1453979304754358e-06,6.27622780079611e-07,6.548100858134662e-06,8.321121568689708e-06,tokens/kWh,30765083.51509534,MB,1603.137536,19302.711296,0.0,18656.264192,17596.668416,s,10,31.165118896484373,3.116511889648437,0.0028056664738443445,3.116374267578125,3.1198353759765625,3.1203074584960935,3.1206851245117186,"[3.11887109375, 3.112717529296875, 3.11574169921875, 3.1153173828125, 3.11294384765625, 3.113296630859375, 3.1170068359375, 3.120779541015625, 3.11973046875, 3.1187138671875]",tokens/s,20.21490763736724,kWh,3.674833278265007e-05,2.0139808391414205e-05,0.0001488593290232657,0.00020574747019732999,tokens/kWh,306200.60572106886,,s,630,31.163331577301047,0.04946560567825559,0.0003132609689793999,0.04935270309448242,0.04982671203613281,0.05023597679138184,0.050552095375061036,"[0.04953395080566406, 0.049168384552001954, 0.04925132751464844, 0.049122303009033204, 0.049192958831787106, 0.04914688110351562, 0.04918067169189453, 0.049235969543457034, 0.04930355072021484, 0.04919705581665039, 0.049247230529785156, 0.049468414306640625, 0.04978278350830078, 0.04964863967895508, 0.04955852890014648, 0.049522689819335934, 0.049258495330810545, 0.04939571380615235, 0.049635326385498044, 0.04963020706176758, 0.05062451171875, 0.04986470413208008, 0.049650688171386716, 0.0496517105102539, 0.049263614654541016, 0.04928204727172852, 0.04927385711669922, 0.04958720016479492, 0.04933631896972656, 0.049452030181884765, 0.04928204727172852, 0.04935782241821289, 0.049235969543457034, 0.049410049438476565, 0.04929228973388672, 0.04931891250610351, 0.04930047988891602, 0.049309696197509766, 0.049360897064208986, 0.04950630569458008, 0.05045657730102539, 0.05069004821777344, 0.04936294555664063, 0.049576961517333984, 0.04960870361328125, 0.04941823959350586, 0.04931686401367188, 0.04938649749755859, 0.04947763061523437, 0.049448959350585936, 0.04939263916015625, 0.04939571380615235, 0.04942233657836914, 0.04966092681884766, 0.04997017669677734, 0.04993331146240235, 0.04943360137939453, 0.04943564987182617, 0.04931174468994141, 0.049355777740478515, 0.04951039886474609, 0.050441215515136716, 0.04985036849975586, 0.04951039886474609, 0.04921753692626953, 0.04917657470703125, 0.04919500732421875, 0.04917248153686524, 0.04922060775756836, 0.04915507125854492, 0.04922982406616211, 0.04936294555664063, 0.04962723159790039, 0.04918159866333008, 0.049203201293945314, 0.04925439834594727, 0.04963840103149414, 0.049152000427246094, 0.04912639999389649, 0.049296382904052735, 0.050423809051513675, 0.04973158264160156, 0.04960358428955078, 0.04924518585205078, 0.049263614654541016, 0.0491960334777832, 0.04920115280151367, 0.04929024124145508, 0.049274879455566405, 0.049271808624267575, 0.049288192749023435, 0.04951347351074219, 0.049710079193115236, 0.04925030517578125, 0.0492410888671875, 0.04924518585205078, 0.04935270309448242, 0.049223678588867184, 0.04929228973388672, 0.049259521484375, 0.04935987091064453, 0.0504453125, 0.049809406280517575, 0.04964352035522461, 0.049277950286865234, 0.04931071853637695, 0.049247230529785156, 0.049271808624267575, 0.04934348678588867, 0.04926464080810547, 0.049339393615722656, 0.04929536056518555, 0.049448959350585936, 0.04938240051269531, 0.04934963226318359, 0.049263614654541016, 0.04972236633300781, 0.049301502227783206, 0.04935065460205078, 0.0493834228515625, 0.049355777740478515, 0.04933222579956055, 0.05056716918945312, 0.04984320068359375, 0.049726463317871096, 0.049306625366210936, 0.04947148895263672, 0.04921036911010742, 0.04916326522827148, 0.04926259231567383, 0.04943667221069336, 0.049223678588867184, 0.04928307342529297, 0.04913151931762695, 0.04921343994140625, 0.049165313720703124, 0.0495206413269043, 0.049844223022460936, 0.04974489593505859, 0.04959641647338867, 0.049631233215332034, 0.05045043182373047, 0.04969472122192383, 0.04921855926513672, 0.04926566314697266, 0.04917145538330078, 0.049258495330810545, 0.04933631896972656, 0.049544193267822265, 0.049185791015625, 0.04919398498535156, 0.04925235366821289, 0.04928921508789062, 0.04930355072021484, 0.04931686401367188, 0.049274879455566405, 0.04943974304199219, 0.049568767547607424, 0.04940083312988281, 0.0494284782409668, 0.049225727081298826, 0.049312767028808595, 0.050337791442871094, 0.0497151985168457, 0.049304576873779295, 0.049277950286865234, 0.049476608276367184, 0.04943462371826172, 0.049320960998535154, 0.04926054382324219, 0.04935168075561523, 0.04930355072021484, 0.04936703872680664, 0.049326080322265625, 0.049293312072753906, 0.04959641647338867, 0.04928102493286133, 0.04939571380615235, 0.04964352035522461, 0.04976332855224609, 0.04935270309448242, 0.04970905685424805, 0.049329151153564454, 0.0504637451171875, 0.049871871948242184, 0.04940902328491211, 0.049331199645996096, 0.05018624114990235, 0.049430526733398435, 0.04940697479248047, 0.049206272125244144, 0.049223678588867184, 0.04912844848632812, 0.04948070526123047, 0.049495040893554686, 0.04918272018432617, 0.04920115280151367, 0.04961177444458008, 0.04937625503540039, 0.049219585418701174, 0.049347583770751956, 0.049291263580322264, 0.050500606536865236, 0.049759231567382815, 0.04924620819091797, 0.049209342956542966, 0.049358848571777345, 0.0491776008605957, 0.04929843139648438, 0.04940083312988281, 0.04929536056518555, 0.04922675323486328, 0.04928102493286133, 0.049397758483886715, 0.04935987091064453, 0.04921343994140625, 0.04927078247070313, 0.049560577392578124, 0.04944486236572266, 0.049296382904052735, 0.049294334411621094, 0.04925747299194336, 0.049301502227783206, 0.050423809051513675, 0.04942745590209961, 0.049307647705078124, 0.04967833709716797, 0.04928307342529297, 0.049306625366210936, 0.049342464447021485, 0.049329151153564454, 0.04932505416870117, 0.04934656143188477, 0.04927590560913086, 0.04935065460205078, 0.04930252838134765, 0.049966079711914066, 0.04984832000732422, 0.04982476806640625, 0.049432575225830076, 0.04929228973388672, 0.04934860610961914, 0.04930867385864258, 0.04933529663085937, 0.05039513778686523, 0.0498073616027832, 0.04938137435913086, 0.04932198333740234, 0.049339393615722656, 0.049535999298095705, 0.050103294372558595, 0.04987596893310547, 0.04944076919555664, 0.049188865661621096, 0.04915097427368164, 0.04917248153686524, 0.04954214477539062, 0.049189888000488284, 0.04922163009643555, 0.04915097427368164, 0.04919807815551758, 0.04918783950805664, 0.04931584167480469, 0.05047296142578125, 0.04928716659545898, 0.04964147186279297, 0.04962713623046875, 0.049181697845458984, 0.04963225555419922, 0.04926259231567383, 0.04915814590454102, 0.0492492790222168, 0.04920729446411133, 0.049247230529785156, 0.049154048919677736, 0.04925337600708008, 0.049635326385498044, 0.0496814079284668, 0.04922777557373047, 0.04922675323486328, 0.049276927947998046, 0.0492410888671875, 0.049258495330810545, 0.049324031829833984, 0.05050163269042969, 0.049442817687988284, 0.04964556884765625, 0.04932710266113281, 0.04957183837890625, 0.049481727600097655, 0.04939059066772461, 0.04964761734008789, 0.04926265716552734, 0.04928307342529297, 0.04925433731079101, 0.049258495330810545, 0.049522689819335934, 0.04937625503540039, 0.049361919403076174, 0.04965785598754883, 0.049314815521240236, 0.049360897064208986, 0.04931891250610351, 0.04930559921264648, 0.04943462371826172, 0.050541568756103515, 0.04942233657836914, 0.04931174468994141, 0.04962918472290039, 0.04934963226318359, 0.0493834228515625, 0.04939571380615235, 0.0493383674621582, 0.049360897064208986, 0.04931071853637695, 0.04942540740966797, 0.049140735626220705, 0.04919705581665039, 0.049209342956542966, 0.04922060775756836, 0.04919398498535156, 0.049225727081298826, 0.04913356781005859, 0.04939980697631836, 0.050331649780273435, 0.04947353744506836, 0.04918783950805664, 0.0492308464050293, 0.04915609741210938, 0.04925747299194336, 0.04921139144897461, 0.04925235366821289, 0.04935987091064453, 0.050553886413574216, 0.049669086456298826, 0.049691646575927735, 0.04919910430908203, 0.04919705581665039, 0.049157119750976565, 0.04926259231567383, 0.04911206436157227, 0.04929024124145508, 0.049261566162109374, 0.0492308464050293, 0.049363967895507815, 0.05043404769897461, 0.04977459335327149, 0.049293312072753906, 0.049225727081298826, 0.04925337600708008, 0.04923699188232422, 0.04936601638793945, 0.04928307342529297, 0.049271808624267575, 0.04930355072021484, 0.0498155517578125, 0.04941823959350586, 0.0492410888671875, 0.04929024124145508, 0.0493199348449707, 0.049274879455566405, 0.049312767028808595, 0.049258495330810545, 0.04962201690673828, 0.04935270309448242, 0.04951551818847656, 0.05065216064453125, 0.049825790405273435, 0.04969574356079102, 0.04931584167480469, 0.04935065460205078, 0.049309696197509766, 0.049312767028808595, 0.049314815521240236, 0.049377281188964846, 0.04965683364868164, 0.04940697479248047, 0.04943462371826172, 0.04937625503540039, 0.049312767028808595, 0.049189888000488284, 0.049225727081298826, 0.049206272125244144, 0.04933324813842774, 0.049306625366210936, 0.050372608184814455, 0.04975513458251953, 0.04933529663085937, 0.049576961517333984, 0.049185791015625, 0.0492677116394043, 0.04929536056518555, 0.049258495330810545, 0.04919398498535156, 0.049721343994140625, 0.04919398498535156, 0.049255424499511716, 0.0492492790222168, 0.049486846923828126, 0.04926464080810547, 0.049517566680908204, 0.04919091033935547, 0.04956979370117188, 0.04928921508789062, 0.04919500732421875, 0.05019750213623047, 0.0505456657409668, 0.04976128005981445, 0.049271808624267575, 0.04920217514038086, 0.04948275375366211, 0.049288192749023435, 0.049291263580322264, 0.04929740905761719, 0.04984832000732422, 0.04938444900512695, 0.04935168075561523, 0.04923289489746094, 0.04928307342529297, 0.04936908721923828, 0.04921139144897461, 0.04933017730712891, 0.049503231048583986, 0.049375232696533204, 0.04933324813842774, 0.049478656768798826, 0.049691646575927735, 0.05046886444091797, 0.0497520637512207, 0.04968447875976562, 0.049353729248046874, 0.0493568000793457, 0.049870849609375, 0.04985036849975586, 0.049793022155761715, 0.049770496368408204, 0.04959539031982422, 0.04941619110107422, 0.049285118103027346, 0.049448959350585936, 0.04932710266113281, 0.04945305633544922, 0.04917555236816406, 0.04914790344238281, 0.04964044952392578, 0.050138111114501956, 0.05053849411010742, 0.04935782241821289, 0.04925337600708008, 0.04918067169189453, 0.04936601638793945, 0.0491673583984375, 0.04926566314697266, 0.049448959350585936, 0.04922265625, 0.04922163009643555, 0.04935475158691406, 0.04923392105102539, 0.04926976013183594, 0.0494202880859375, 0.04963840103149414, 0.049770496368408204, 0.04946944046020508, 0.049560577392578124, 0.04965785598754883, 0.04951039886474609, 0.04984832000732422, 0.050121726989746096, 0.049683456420898435, 0.04913049697875976, 0.04917145538330078, 0.049637374877929685, 0.049757183074951174, 0.0494284782409668, 0.04952678298950195, 0.04961894226074219, 0.04930559921264648, 0.04946636962890625, 0.04983910369873047, 0.04965580749511719, 0.049410049438476565, 0.04976844787597656, 0.04940595245361328, 0.049460224151611325, 0.04956979370117188, 0.04924620819091797, 0.04940697479248047, 0.05000806427001953, 0.05049856185913086, 0.04968960189819336, 0.04928204727172852, 0.049667072296142575, 0.049775615692138675, 0.04982271957397461, 0.04969779205322265, 0.04949708938598633, 0.04970700836181641, 0.049568767547607424, 0.04940492630004883, 0.04941823959350586, 0.04930867385864258, 0.049516544342041016, 0.049397758483886715, 0.049414142608642575, 0.0494571533203125, 0.049159168243408206, 0.050267135620117184, 0.04977664184570312, 0.049634304046630856, 0.04940390396118164, 0.04928716659545898, 0.049203201293945314, 0.04922675323486328, 0.049160190582275394, 0.04923494338989258, 0.049188865661621096, 0.049229888916015624, 0.04917446517944336, 0.04928409576416016, 0.04944793701171875, 0.049261566162109374, 0.04926873779296875, 0.049285118103027346, 0.04964863967895508, 0.049329151153564454, 0.04922470474243164, 0.04919807815551758, 0.050283519744873044, 0.049823806762695315, 0.050229183197021486, 0.04983910369873047, 0.04962713623046875, 0.04928204727172852, 0.04933324813842774, 0.0492410888671875, 0.0492564468383789, 0.049304576873779295, 0.0492308464050293, 0.04927385711669922, 0.049329151153564454, 0.049446910858154294, 0.04933222579956055, 0.04953497695922852, 0.049719295501708984, 0.04926668930053711, 0.049329151153564454, 0.04932505416870117, 0.04943564987182617, 0.050644992828369144, 0.04988620758056641, 0.04966502380371094, 0.04969267272949219, 0.049791999816894535, 0.04963942337036133, 0.0497367057800293, 0.04940185546875, 0.04962508773803711, 0.049360897064208986, 0.04976128005981445, 0.0497151985168457, 0.04988415908813477, 0.05024153518676758, 0.04986265563964844, 0.0496445426940918, 0.049539070129394534, 0.049360897064208986, 0.049307647705078124, 0.050476032257080077, 0.04972851181030274, 0.04951347351074219, 0.04957388687133789, 0.04918067169189453, 0.049261566162109374, 0.049159168243408206, 0.04920115280151367, 0.04927078247070313, 0.049393665313720705, 0.04933017730712891, 0.04925439834594727, 0.049255424499511716, 0.049242111206054685, 0.049304576873779295, 0.04929024124145508, 0.0492042236328125, 0.04923289489746094, 0.04922060775756836, 0.04920832061767578, 0.049263614654541016, 0.05063577651977539, 0.04983500671386719, 0.049312767028808595, 0.04926054382324219, 0.049667072296142575, 0.049274879455566405, 0.049410049438476565, 0.04973158264160156, 0.04996915054321289, 0.049448959350585936, 0.04955340957641602, 0.0495912971496582, 0.04938547134399414, 0.05008793640136719, 0.049748992919921874, 0.04967628860473633, 0.049670143127441405, 0.04957900619506836, 0.04936294555664063, 0.049408000946044923, 0.04936601638793945, 0.050547710418701174, 0.05000601577758789, 0.04958720016479492, 0.04928102493286133, 0.049294334411621094, 0.04930252838134765, 0.04935987091064453, 0.04948275375366211, 0.04966400146484375, 0.04946432113647461, 0.04937932968139649, 0.04936601638793945, 0.04927590560913086, 0.049616897583007816, 0.04933222579956055, 0.04937625503540039, 0.049304576873779295, 0.04934143829345703, 0.04930976104736328, 0.04956051254272461, 0.05012377548217774]",tokens/s,20.21606702856135,,,,,,MB,935.800832,19279.642624,0.0,18649.972736,18648.241152,s,1,7.0336591796875,7.0336591796875,0.0,7.0336591796875,7.0336591796875,7.0336591796875,7.0336591796875,[7.0336591796875],,kWh,2.7511501743036765e-06,1.491781122347987e-06,6.234727210008284e-06,1.0477658506659947e-05, float16-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-125m,,cuda,0,42,,,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,d35829e539df8480b726c647eeabf91e41eae047,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,,MB,1177.94816,1021.837312,0.0,375.390208,302.523392,s,27,0.1810455346107483,0.006705390170768455,0.0002705648861503602,0.006606912136077881,0.006904780769348144,0.007322272014617919,0.007635545606613159,"[0.007686048030853271, 0.00749180793762207, 0.006639999866485596, 0.0066425600051879885, 0.006556575775146485, 0.006516511917114258, 0.006568863868713379, 0.006573984146118164, 0.006536672115325928, 0.006594751834869385, 0.006672639846801758, 0.006588064193725586, 0.00667468786239624, 0.0065548157691955565, 0.00652182388305664, 0.006765567779541016, 0.006757887840270996, 0.006926688194274902, 0.006606912136077881, 0.006890175819396972, 0.006575039863586426, 0.006581727981567383, 0.0065723519325256346, 0.006719679832458496, 0.006611487865447998, 0.006559840202331543, 0.0066583681106567386]",tokens/s,38178.240710884944,kWh,7.981352664686897e-08,4.373296065590989e-08,1.8992207385944711e-07,3.1346856116222596e-07,tokens/kWh,816668820.1548707,MB,1178.27584,1021.837312,0.0,375.390208,312.411136,s,27,9.993893798828124,0.37014421477141196,0.008688703950292843,0.36762567138671876,0.3743817565917969,0.3830754180908203,0.40156628479003903,"[0.40693777465820313, 0.3862781982421875, 0.37118072509765626, 0.3681455993652344, 0.36609967041015623, 0.368382568359375, 0.37356808471679687, 0.36386138916015626, 0.3661558837890625, 0.36410232543945314, 0.36762567138671876, 0.37187216186523436, 0.3638875427246094, 0.3638258361816406, 0.36992922973632814, 0.3756022644042969, 0.3724678955078125, 0.37205621337890626, 0.36720703125, 0.3734965515136719, 0.3655494384765625, 0.3629034423828125, 0.3689253234863281, 0.366518310546875, 0.36670758056640623, 0.36427529907226563, 0.366331787109375]",tokens/s,170.203929943648,kWh,4.353792633628862e-06,2.38562450471523e-06,7.727738581252736e-06,1.4467155719596826e-05,tokens/kWh,4354691.497144934,,s,1700,10.161002492427814,0.005977060289663427,0.0008917497430451113,0.005799935817718506,0.006005760192871094,0.0063869442224502564,0.012781751899719238,"[0.006756351947784424, 0.0065413122177124024, 0.0065771517753601075, 0.006540287971496582, 0.006673408031463623, 0.006632448196411133, 0.006620160102844238, 0.006489088058471679, 0.0064737281799316405, 0.006434815883636475, 0.0064880638122558594, 0.006303743839263916, 0.006191103935241699, 0.00643174409866333, 0.006391808032989502, 0.00658739185333252, 0.0065177597999572755, 0.006487040042877197, 0.006523903846740723, 0.006391808032989502, 0.006398975849151611, 0.006603775978088379, 0.0063805441856384275, 0.006527999877929688, 0.00658844804763794, 0.006414303779602051, 0.006433792114257812, 0.006386688232421875, 0.0065382399559021, 0.006508543968200684, 0.0062904319763183595, 0.006294528007507324, 0.006310912132263183, 0.006364160060882569, 0.0063324799537658695, 0.006403007984161377, 0.006440959930419922, 0.006394879817962646, 0.006460415840148926, 0.006561791896820069, 0.006505472183227539, 0.006615039825439453, 0.00643993616104126, 0.006404096126556396, 0.0064778242111206055, 0.006268928050994873, 0.006344704151153564, 0.00643174409866333, 0.006330368041992188, 0.006396927833557129, 0.00638156795501709, 0.006419456005096436, 0.006404096126556396, 0.006555647850036621, 0.006520832061767578, 0.00632422399520874, 0.006418432235717773, 0.00637337589263916, 0.006235136032104492, 0.006408192157745361, 0.006316031932830811, 0.006489088058471679, 0.014823424339294433, 0.006553599834442139, 0.006559743881225586, 0.006453248023986816, 0.006436863899230957, 0.006376448154449463, 0.006372352123260498, 0.006342656135559082, 0.006261760234832763, 0.006203455924987793, 0.006192063808441162, 0.006294528007507324, 0.006367231845855713, 0.006062079906463623, 0.005992447853088379, 0.005984255790710449, 0.006060031890869141, 0.0059688959121704105, 0.006247424125671387, 0.006178815841674804, 0.0061972479820251464, 0.0061931519508361815, 0.006392831802368164, 0.007174143791198731, 0.007126016139984131, 0.007066688060760498, 0.006812607765197754, 0.00607539176940918, 0.006021120071411133, 0.006086656093597412, 0.0060026879310607914, 0.006016064167022705, 0.005983232021331787, 0.005988287925720215, 0.005957632064819336, 0.005982207775115967, 0.005934080123901367, 0.005952511787414551, 0.005937151908874512, 0.005913599967956543, 0.005929984092712402, 0.005929984092712402, 0.005891071796417236, 0.005909503936767578, 0.005914624214172363, 0.005932032108306885, 0.005976064205169678, 0.0059054079055786135, 0.005907455921173096, 0.005903359889984131, 0.005925888061523437, 0.00591974401473999, 0.0059433279037475585, 0.005891039848327637, 0.005925888061523437, 0.00591161584854126, 0.005922751903533936, 0.005850111961364746, 0.005773344039916992, 0.005794784069061279, 0.005763072013854981, 0.005750783920288086, 0.005792768001556397, 0.01286451244354248, 0.005829631805419922, 0.005792768001556397, 0.005792768001556397, 0.005740543842315674, 0.005810175895690918, 0.005792768001556397, 0.005796864032745362, 0.005818367958068848, 0.005798912048339844, 0.005783552169799804, 0.0058091521263122555, 0.005835775852203369, 0.005796864032745362, 0.005805056095123291, 0.005873663902282715, 0.0057825279235839844, 0.005768191814422607, 0.005766143798828125, 0.005781504154205322, 0.0058716158866882326, 0.005792768001556397, 0.005786623954772949, 0.005773312091827393, 0.00576204776763916, 0.005775360107421875, 0.005746687889099121, 0.005775360107421875, 0.005732351779937744, 0.005846015930175781, 0.005792768001556397, 0.005915647983551026, 0.00781824016571045, 0.006100992202758789, 0.005971968173980713, 0.005933055877685547, 0.006015999794006348, 0.005962751865386963, 0.006073344230651856, 0.006005760192871094, 0.006012928009033203, 0.005983232021331787, 0.006097919940948486, 0.005934080123901367, 0.006020095825195313, 0.005963776111602783, 0.005997632026672363, 0.005983168125152588, 0.005925888061523437, 0.005946368217468262, 0.005883903980255127, 0.005981184005737304, 0.005851136207580566, 0.005804096221923828, 0.005790656089782715, 0.0057784318923950195, 0.005762112140655517, 0.005783487796783447, 0.005788671970367432, 0.005757952213287353, 0.005786623954772949, 0.005760000228881836, 0.005791744232177734, 0.01284505558013916, 0.005831679821014404, 0.005777408123016357, 0.005739520072937012, 0.005797887802124023, 0.005786623954772949, 0.005810239791870117, 0.005738431930541992, 0.005801983833312988, 0.005807104110717773, 0.005790719985961914, 0.0058009600639343266, 0.005758975982666016, 0.005822463989257813, 0.005760000228881836, 0.005892096042633056, 0.005805088043212891, 0.005775328159332275, 0.0058419198989868165, 0.005982207775115967, 0.005979135990142822, 0.005952511787414551, 0.005913599967956543, 0.005980160236358643, 0.005901311874389649, 0.0059351038932800295, 0.005946368217468262, 0.005961728096008301, 0.005976064205169678, 0.005976064205169678, 0.005952511787414551, 0.005938176155090332, 0.005967872142791748, 0.005976064205169678, 0.005925888061523437, 0.0059955201148986816, 0.005916672229766846, 0.0058757119178771975, 0.005787648200988769, 0.005794816017150879, 0.0058388481140136715, 0.005739520072937012, 0.0057835841178894044, 0.005751776218414307, 0.0058009600639343266, 0.005755904197692871, 0.005775360107421875, 0.005791744232177734, 0.005781504154205322, 0.005808127880096436, 0.005730303764343261, 0.005806079864501953, 0.00573747205734253, 0.005794816017150879, 0.0057497601509094234, 0.005822463989257813, 0.005825535774230957, 0.005752831935882568, 0.005824512004852295, 0.005764095783233642, 0.005833727836608887, 0.005765120029449463, 0.005772287845611572, 0.012836864471435547, 0.005822463989257813, 0.005826560020446778, 0.0058419198989868165, 0.005988351821899414, 0.00592793607711792, 0.005993472099304199, 0.005941247940063477, 0.005972991943359375, 0.0059402561187744144, 0.005883872032165527, 0.005907455921173096, 0.005893119812011719, 0.0059269118309021, 0.005921792030334472, 0.00582041597366333, 0.005780479907989502, 0.005764095783233642, 0.005754879951477051, 0.005797887802124023, 0.005752831935882568, 0.005720064163208007, 0.005752831935882568, 0.005775360107421875, 0.005766143798828125, 0.005763072013854981, 0.005781504154205322, 0.0057794561386108395, 0.005780479907989502, 0.00577023983001709, 0.005753856182098388, 0.005776383876800537, 0.005758975982666016, 0.0057794561386108395, 0.005765120029449463, 0.005776383876800537, 0.005787648200988769, 0.00582041597366333, 0.005745664119720459, 0.0057794561386108395, 0.005753888130187988, 0.005755871772766113, 0.005765120029449463, 0.0057712640762329105, 0.005839871883392334, 0.005799935817718506, 0.005860352039337159, 0.005849088191986084, 0.005830656051635743, 0.005763072013854981, 0.00576204776763916, 0.005765120029449463, 0.005726208209991455, 0.005784575939178467, 0.005735424041748047, 0.005775360107421875, 0.005715968132019043, 0.005785600185394287, 0.005744639873504639, 0.005728256225585937, 0.005768191814422607, 0.005716991901397705, 0.005766143798828125, 0.012689408302307128, 0.005765120029449463, 0.005760000228881836, 0.005750783920288086, 0.00576204776763916, 0.005765120029449463, 0.005746687889099121, 0.005752831935882568, 0.005731328010559082, 0.005780479907989502, 0.005780479907989502, 0.00573747205734253, 0.005785600185394287, 0.00573747205734253, 0.005766143798828125, 0.005769216060638428, 0.005746687889099121, 0.005751808166503906, 0.005746687889099121, 0.00576204776763916, 0.00577023983001709, 0.005774335861206055, 0.005720064163208007, 0.005795839786529541, 0.005763072013854981, 0.00576204776763916, 0.005793791770935058, 0.005818367958068848, 0.005791744232177734, 0.005777408123016357, 0.005735424041748047, 0.005831679821014404, 0.005753856182098388, 0.005791744232177734, 0.0057497601509094234, 0.005763072013854981, 0.005781504154205322, 0.005769216060638428, 0.0058726401329040525, 0.005884928226470947, 0.006020095825195313, 0.005880864143371582, 0.006007775783538818, 0.005954559803009033, 0.005931007862091065, 0.005985280036926269, 0.005938176155090332, 0.0059985918998718265, 0.0059361281394958495, 0.005920767784118652, 0.005948416233062744, 0.005900288105010986, 0.0059770879745483394, 0.005903359889984131, 0.005966847896575928, 0.006073344230651856, 0.005907455921173096, 0.006072319984436035, 0.005931007862091065, 0.006048768043518066, 0.005917695999145508, 0.005940224170684814, 0.005967872142791748, 0.013011967658996582, 0.005960703849792481, 0.005950463771820068, 0.005959680080413818, 0.005952511787414551, 0.0059688959121704105, 0.005925951957702636, 0.005946303844451904, 0.0060026879310607914, 0.005913599967956543, 0.00597811222076416, 0.005884928226470947, 0.005979135990142822, 0.005938176155090332, 0.005942272186279297, 0.005986303806304932, 0.005910528182983398, 0.005948416233062744, 0.00591974401473999, 0.005982207775115967, 0.005922815799713135, 0.0059023361206054685, 0.005929984092712402, 0.005932032108306885, 0.005957632064819336, 0.0059391999244689945, 0.00592796802520752, 0.005978079795837402, 0.005884928226470947, 0.006008831977844238, 0.005870656013488769, 0.006001599788665771, 0.005917695999145508, 0.005911551952362061, 0.005938176155090332, 0.005950496196746826, 0.005970911979675293, 0.005945343971252442, 0.005902400016784668, 0.006002624034881592, 0.005874688148498535, 0.0060067839622497555, 0.005881855964660645, 0.005986303806304932, 0.005914624214172363, 0.005885983943939209, 0.005924831867218017, 0.00590444803237915, 0.005918655872344971, 0.005886975765228271, 0.005929984092712402, 0.0059023361206054685, 0.005941247940063477, 0.005915647983551026, 0.005910528182983398, 0.005879807949066162, 0.005922815799713135, 0.005897247791290283, 0.005857247829437256, 0.005733376026153565, 0.005766143798828125, 0.00570470380783081, 0.005774335861206055, 0.012761088371276855, 0.005774335861206055, 0.005752831935882568, 0.005773312091827393, 0.005743616104125977, 0.005784575939178467, 0.005766143798828125, 0.005751808166503906, 0.005788671970367432, 0.005793791770935058, 0.005803008079528809, 0.005733376026153565, 0.005798912048339844, 0.0057497601509094234, 0.005810175895690918, 0.005769216060638428, 0.005799935817718506, 0.005731328010559082, 0.0057784318923950195, 0.00576204776763916, 0.005777408123016357, 0.005777408123016357, 0.005780479907989502, 0.005788671970367432, 0.0057487359046936035, 0.005788671970367432, 0.005775360107421875, 0.005804031848907471, 0.00573747205734253, 0.005787648200988769, 0.005740543842315674, 0.005773312091827393, 0.0057712640762329105, 0.005789696216583252, 0.005786623954772949, 0.005740543842315674, 0.005773312091827393, 0.005744639873504639, 0.0058009600639343266, 0.005751808166503906, 0.0057825279235839844, 0.005787648200988769, 0.005745664119720459, 0.0057794561386108395, 0.0058122239112854005, 0.005796864032745362, 0.0057497601509094234, 0.005791744232177734, 0.005732351779937744, 0.005764095783233642, 0.0057712640762329105, 0.005734399795532226, 0.005786623954772949, 0.005731328010559082, 0.005757952213287353, 0.005752863883972168, 0.005753824234008789, 0.005764095783233642, 0.00573747205734253, 0.00581324815750122, 0.005728256225585937, 0.005816319942474365, 0.005710847854614258, 0.012739583969116211, 0.00576204776763916, 0.00582041597366333, 0.005751808166503906, 0.005810175895690918, 0.005774335861206055, 0.005738495826721191, 0.00576204776763916, 0.00577023983001709, 0.005760000228881836, 0.005745664119720459, 0.005758975982666016, 0.005735424041748047, 0.005760000228881836, 0.005776383876800537, 0.005731328010559082, 0.005751808166503906, 0.005787648200988769, 0.0057497601509094234, 0.005750783920288086, 0.005751808166503906, 0.005744639873504639, 0.0057292799949646, 0.00576204776763916, 0.005760000228881836, 0.005747712135314941, 0.005726208209991455, 0.005756927967071533, 0.005786623954772949, 0.005763072013854981, 0.005777408123016357, 0.005734399795532226, 0.005738495826721191, 0.005774335861206055, 0.005752831935882568, 0.005764095783233642, 0.005738495826721191, 0.005776383876800537, 0.005756927967071533, 0.0057825279235839844, 0.005786623954772949, 0.005780479907989502, 0.005814271926879883, 0.005736447811126709, 0.005856256008148194, 0.005817344188690185, 0.0059064321517944334, 0.005798912048339844, 0.005822463989257813, 0.005824512004852295, 0.005991424083709717, 0.005988351821899414, 0.005996543884277344, 0.0059391999244689945, 0.005892096042633056, 0.005952511787414551, 0.005950463771820068, 0.005931007862091065, 0.005944320201873779, 0.005921792030334472, 0.005886975765228271, 0.005923840045928955, 0.005883903980255127, 0.012781567573547363, 0.005780543804168701, 0.00573740816116333, 0.005740543842315674, 0.005742591857910156, 0.005762080192565918, 0.0057743039131164554, 0.005715968132019043, 0.005774335861206055, 0.005744639873504639, 0.005799935817718506, 0.005798975944519043, 0.005781439781188965, 0.005750783920288086, 0.00576204776763916, 0.00577023983001709, 0.005745664119720459, 0.005730303764343261, 0.0057487359046936035, 0.005757952213287353, 0.00582860803604126, 0.005769216060638428, 0.005755904197692871, 0.005795839786529541, 0.005736447811126709, 0.005791744232177734, 0.005849088191986084, 0.005799935817718506, 0.005767168045043946, 0.005795839786529541, 0.005754879951477051, 0.005755904197692871, 0.005787648200988769, 0.005756927967071533, 0.005792768001556397, 0.005818367958068848, 0.0057487359046936035, 0.005792768001556397, 0.005777408123016357, 0.005819392204284668, 0.005757952213287353, 0.005803008079528809, 0.005765120029449463, 0.005773344039916992, 0.005749728202819824, 0.005763072013854981, 0.005827583789825439, 0.005753856182098388, 0.005799935817718506, 0.005747712135314941, 0.005794816017150879, 0.005764095783233642, 0.0057825279235839844, 0.005790719985961914, 0.005730303764343261, 0.005819392204284668, 0.005716991901397705, 0.005789696216583252, 0.005747712135314941, 0.005819392204284668, 0.005746687889099121, 0.005766143798828125, 0.005818367958068848, 0.012986368179321289, 0.005958655834197998, 0.0059658241271972655, 0.006091775894165039, 0.005938176155090332, 0.006050816059112549, 0.005937151908874512, 0.006033408164978027, 0.005961728096008301, 0.006024191856384278, 0.005986303806304932, 0.006005760192871094, 0.005996543884277344, 0.006003776073455811, 0.005966784000396729, 0.006008831977844238, 0.005909503936767578, 0.005944320201873779, 0.005756927967071533, 0.005743616104125977, 0.005776383876800537, 0.005751808166503906, 0.005816319942474365, 0.005761023998260498, 0.005795839786529541, 0.005739520072937012, 0.005814271926879883, 0.005763072013854981, 0.005755904197692871, 0.005796864032745362, 0.005726208209991455, 0.005816319942474365, 0.005706751823425293, 0.005787648200988769, 0.005755904197692871, 0.005787648200988769, 0.005781504154205322, 0.005755936145782471, 0.005772255897521973, 0.005744639873504639, 0.0057794561386108395, 0.005765120029449463, 0.00576204776763916, 0.005808127880096436, 0.005917695999145508, 0.005757952213287353, 0.005744639873504639, 0.005754879951477051, 0.005763072013854981, 0.005757952213287353, 0.005752831935882568, 0.0057487359046936035, 0.005774335861206055, 0.00576204776763916, 0.005761023998260498, 0.005755904197692871, 0.005734399795532226, 0.005767168045043946, 0.0057794561386108395, 0.005743616104125977, 0.005773312091827393, 0.005746687889099121, 0.005735424041748047, 0.012906496047973632, 0.005769216060638428, 0.005791744232177734, 0.005724160194396972, 0.005755904197692871, 0.005787648200988769, 0.005745664119720459, 0.005801023960113526, 0.0058858880996704105, 0.005860352039337159, 0.005799935817718506, 0.005768191814422607, 0.005786623954772949, 0.005763072013854981, 0.005785600185394287, 0.005793791770935058, 0.005908480167388916, 0.00586137580871582, 0.005949440002441406, 0.0059391999244689945, 0.005960703849792481, 0.005948416233062744, 0.0059054079055786135, 0.005943295955657959, 0.0059351038932800295, 0.0059054079055786135, 0.005975039958953857, 0.005946368217468262, 0.005953536033630371, 0.005940224170684814, 0.005955584049224853, 0.005931007862091065, 0.005956607818603516, 0.0059351038932800295, 0.0059361600875854495, 0.005913568019866943, 0.005970943927764893, 0.0059391999244689945, 0.005951488018035888, 0.005909503936767578, 0.005961728096008301, 0.005925888061523437, 0.005895167827606201, 0.005947391986846923, 0.005908480167388916, 0.005944320201873779, 0.005985280036926269, 0.0059269118309021, 0.005923840045928955, 0.005916672229766846, 0.005911551952362061, 0.00592902421951294, 0.005916607856750488, 0.005996543884277344, 0.005860352039337159, 0.005954559803009033, 0.005874688148498535, 0.005967872142791748, 0.005880832195281982, 0.005983232021331787, 0.00588595199584961, 0.005924863815307617, 0.005886975765228271, 0.013011967658996582, 0.0059688959121704105, 0.005894144058227539, 0.005882880210876465, 0.005730303764343261, 0.005739520072937012, 0.005682176113128662, 0.0057149438858032225, 0.00572108793258667, 0.005783552169799804, 0.005768191814422607, 0.005761023998260498, 0.005808127880096436, 0.005741600036621094, 0.005817311763763428, 0.005772287845611572, 0.0057712640762329105, 0.005758975982666016, 0.00572108793258667, 0.005805056095123291, 0.005734399795532226, 0.005776383876800537, 0.005754879951477051, 0.005769216060638428, 0.005757952213287353, 0.005739520072937012, 0.005788671970367432, 0.005747712135314941, 0.005783552169799804, 0.005746687889099121, 0.005731328010559082, 0.005774335861206055, 0.005738495826721191, 0.005768191814422607, 0.005772287845611572, 0.005794816017150879, 0.005769216060638428, 0.005760000228881836, 0.005777408123016357, 0.0058419198989868165, 0.0057487359046936035, 0.005786623954772949, 0.0057712640762329105, 0.005764095783233642, 0.005739520072937012, 0.0057487359046936035, 0.005781504154205322, 0.0057497601509094234, 0.005796864032745362, 0.005752831935882568, 0.0057118721008300784, 0.005747712135314941, 0.005801983833312988, 0.005744671821594239, 0.005743584156036377, 0.005774335861206055, 0.005762080192565918, 0.005764063835144043, 0.005740543842315674, 0.005772287845611572, 0.00573747205734253, 0.00573747205734253, 0.005744639873504639, 0.01273043155670166, 0.005759935855865478, 0.005774335861206055, 0.005733376026153565, 0.005797887802124023, 0.005788671970367432, 0.005790719985961914, 0.005718016147613526, 0.005757952213287353, 0.005787648200988769, 0.005752831935882568, 0.005784575939178467, 0.005747712135314941, 0.005783552169799804, 0.005746687889099121, 0.005790719985961914, 0.005775360107421875, 0.005723135948181152, 0.005781568050384521, 0.005748672008514404, 0.005801983833312988, 0.0057712640762329105, 0.005789696216583252, 0.005788671970367432, 0.00577023983001709, 0.005769216060638428, 0.005776415824890137, 0.005810143947601318, 0.005750783920288086, 0.00582041597366333, 0.005743616104125977, 0.005754879951477051, 0.005777472019195556, 0.005790656089782715, 0.005776383876800537, 0.005772287845611572, 0.0057415680885314945, 0.0057497601509094234, 0.005776383876800537, 0.005783552169799804, 0.005784575939178467, 0.005777408123016357, 0.005791744232177734, 0.005769216060638428, 0.005783616065979004, 0.005732287883758545, 0.005764095783233642, 0.005760000228881836, 0.0057487359046936035, 0.005777408123016357, 0.005732351779937744, 0.0057415680885314945, 0.00575596809387207, 0.005800896167755127, 0.00577023983001709, 0.005769216060638428, 0.005745664119720459, 0.005753856182098388, 0.005772287845611572, 0.005734399795532226, 0.00581324815750122, 0.0057497601509094234, 0.005769216060638428, 0.012906496047973632, 0.005928959846496582, 0.005910528182983398, 0.005776383876800537, 0.005752831935882568, 0.0057712640762329105, 0.005735424041748047, 0.005853184223175049, 0.005775360107421875, 0.0057825279235839844, 0.005785600185394287, 0.0057712640762329105, 0.005810175895690918, 0.005783552169799804, 0.005942272186279297, 0.005889023780822754, 0.005947391986846923, 0.005896192073822021, 0.00591871976852417, 0.005911551952362061, 0.005938176155090332, 0.005880832195281982, 0.005938176155090332, 0.005882880210876465, 0.006044672012329101, 0.005957632064819336, 0.005940224170684814, 0.005912576198577881, 0.005952511787414551, 0.005948416233062744, 0.005958655834197998, 0.005915647983551026, 0.005903359889984131, 0.005882880210876465, 0.005954624176025391, 0.005882815837860107, 0.005951488018035888, 0.005756927967071533, 0.0057784318923950195, 0.005781504154205322, 0.005731328010559082, 0.005801983833312988, 0.005751808166503906, 0.005806111812591553, 0.005775328159332275, 0.005787648200988769, 0.005788671970367432, 0.005752831935882568, 0.00577235221862793, 0.0057292160987854005, 0.005839871883392334, 0.005776383876800537, 0.005852159976959229, 0.005788671970367432, 0.0059770879745483394, 0.005990399837493897, 0.005974016189575195, 0.006005760192871094, 0.0059985918998718265, 0.00591974401473999, 0.005980160236358643, 0.005742591857910156, 0.005925888061523437, 0.013165568351745606, 0.0059699201583862304, 0.005921792030334472, 0.005985280036926269, 0.0059699201583862304, 0.005949440002441406, 0.0060364799499511715, 0.005884928226470947, 0.005903359889984131, 0.005831679821014404, 0.005956607818603516, 0.005846015930175781, 0.005992447853088379, 0.005967872142791748, 0.005910528182983398, 0.005976064205169678, 0.005952511787414551, 0.005933055877685547, 0.005987328052520752, 0.0060293121337890625, 0.006017024040222168, 0.0059054079055786135, 0.0059985918998718265, 0.005994495868682862, 0.005890048027038574, 0.005988351821899414, 0.005910528182983398, 0.005997568130493164, 0.005980160236358643, 0.005942272186279297, 0.005979135990142822, 0.005895167827606201, 0.005987328052520752, 0.005957632064819336, 0.006004735946655273, 0.005956607818603516, 0.006082560062408447, 0.005972991943359375, 0.005899263858795166, 0.0060026879310607914, 0.005937151908874512, 0.005993472099304199, 0.0059269118309021, 0.005804031848907471, 0.006107135772705078, 0.0059269118309021, 0.005975039958953857, 0.005972991943359375, 0.005856256008148194, 0.006025216102600098, 0.00588595199584961, 0.006004735946655273, 0.0058787841796875, 0.005947391986846923, 0.005945343971252442, 0.005925920009613037, 0.005943264007568359, 0.005909503936767578, 0.005961728096008301, 0.006013951778411865, 0.005910528182983398, 0.006153215885162353, 0.00587775993347168, 0.013138943672180176, 0.005970943927764893, 0.005970943927764893, 0.005925888061523437, 0.005949440002441406, 0.005980224132537842, 0.005898176193237304, 0.006024191856384278, 0.005922815799713135, 0.006007808208465576, 0.005894144058227539, 0.006026336193084717, 0.0067439680099487306, 0.006091775894165039, 0.006106112003326416, 0.005950463771820068, 0.005949440002441406, 0.0059361281394958495, 0.005940224170684814, 0.0059361281394958495, 0.005833727836608887, 0.005788671970367432, 0.005799935817718506, 0.005810175895690918, 0.005806079864501953, 0.005932032108306885, 0.005988351821899414, 0.005942272186279297, 0.005932032108306885, 0.005858304023742676, 0.00573747205734253, 0.00581324815750122, 0.0057784318923950195, 0.005805056095123291, 0.0057487359046936035, 0.005794816017150879, 0.0058091521263122555, 0.0057794561386108395, 0.005815296173095703, 0.005766143798828125, 0.005798912048339844, 0.005769216060638428, 0.005785600185394287, 0.005804031848907471, 0.005735424041748047, 0.005811200141906738, 0.005760000228881836, 0.005876736164093017, 0.005970943927764893, 0.005961728096008301, 0.005934080123901367, 0.005911551952362061, 0.005970943927764893, 0.005896192073822021, 0.005941247940063477, 0.005925888061523437, 0.005990399837493897, 0.005964799880981446, 0.005886975765228271, 0.005937151908874512, 0.005889023780822754, 0.005915647983551026, 0.005904384136199951, 0.013314047813415527, 0.005986303806304932, 0.005927999973297119, 0.005994431972503662, 0.00586956787109375, 0.005994495868682862, 0.005883903980255127, 0.0059699201583862304, 0.005924863815307617, 0.005929984092712402, 0.005966847896575928, 0.0059054079055786135, 0.006054912090301514, 0.0060631041526794435, 0.005950463771820068, 0.005996543884277344, 0.005910528182983398, 0.005931007862091065, 0.005931007862091065, 0.005933055877685547, 0.005925888061523437, 0.005895167827606201, 0.0059269118309021, 0.006012928009033203, 0.0059136319160461425, 0.0058941121101379395, 0.005916672229766846, 0.00592793607711792, 0.005822463989257813, 0.005892096042633056, 0.006001664161682129, 0.005908480167388916, 0.0059023361206054685, 0.005889023780822754, 0.005867519855499268, 0.00571289587020874, 0.005716991901397705, 0.00567193603515625, 0.005713920116424561, 0.00566374397277832, 0.005696512222290039, 0.005789696216583252, 0.005706751823425293, 0.005784575939178467, 0.0057487359046936035, 0.006770688056945801, 0.005921792030334472, 0.005976064205169678, 0.005850111961364746, 0.00586240005493164, 0.00588595199584961, 0.005892096042633056, 0.006211584091186524, 0.005955584049224853, 0.005855231761932373, 0.005940224170684814, 0.005854207992553711, 0.005860352039337159, 0.005801023960113526, 0.0057363839149475095, 0.005781504154205322, 0.005724160194396972, 0.005799935817718506, 0.012800000190734863, 0.005772287845611572, 0.0057794561386108395, 0.00576204776763916, 0.005763072013854981, 0.005774335861206055, 0.005797887802124023, 0.0057487359046936035, 0.005799935817718506, 0.005772287845611572, 0.005768191814422607, 0.005798912048339844, 0.005754879951477051, 0.00577023983001709, 0.005758975982666016, 0.005822463989257813, 0.005730303764343261, 0.005758975982666016, 0.00581324815750122, 0.00581324815750122, 0.005750783920288086, 0.005740543842315674, 0.005856256008148194, 0.005756927967071533, 0.005785600185394287, 0.0057415680885314945, 0.005761023998260498, 0.005755904197692871, 0.005765120029449463, 0.005798912048339844, 0.0057190399169921875, 0.005823488235473633, 0.00572108793258667, 0.0057712640762329105, 0.005738495826721191, 0.00577023983001709, 0.005803008079528809, 0.005738495826721191, 0.005816319942474365, 0.0057415680885314945, 0.005797887802124023, 0.005750783920288086, 0.00577023983001709, 0.005730303764343261, 0.005742656230926514, 0.00580294418334961, 0.005727231979370117, 0.0057794561386108395, 0.0057292799949646, 0.005777408123016357, 0.005826560020446778, 0.00573747205734253, 0.005780479907989502, 0.006281216144561768, 0.00607539176940918, 0.006024191856384278, 0.006394879817962646, 0.006689792156219483, 0.005982207775115967, 0.0059770879745483394, 0.005915647983551026, 0.0058419198989868165, 0.005790719985961914, 0.013295616149902344, 0.005962751865386963, 0.005949440002441406, 0.005883903980255127, 0.005921823978424072, 0.0060210881233215335, 0.005847040176391601, 0.0059658241271972655, 0.005933055877685547, 0.005944320201873779, 0.005903359889984131, 0.00619212818145752, 0.005908480167388916, 0.005925888061523437, 0.005879807949066162, 0.005975103855133056, 0.005879744052886963, 0.0059391999244689945, 0.005920767784118652, 0.0059391999244689945, 0.0059391999244689945, 0.005946368217468262, 0.005938240051269531, 0.0059678077697753905, 0.005898240089416504, 0.005972032070159912, 0.005898176193237304, 0.005972991943359375, 0.005904384136199951, 0.005957632064819336, 0.00592793607711792, 0.005901311874389649, 0.005943295955657959, 0.005965888023376465, 0.005955520153045654, 0.00592793607711792, 0.005941247940063477, 0.005948448181152343, 0.005947360038757324, 0.0059688959121704105, 0.005959680080413818, 0.005932032108306885, 0.005943295955657959, 0.005913599967956543, 0.00597811222076416, 0.00586956787109375, 0.005986303806304932, 0.005894144058227539, 0.005982207775115967, 0.005932032108306885, 0.0059054079055786135, 0.005912576198577881, 0.005931007862091065, 0.005910528182983398, 0.0058787841796875, 0.005912576198577881, 0.005929984092712402, 0.005889023780822754, 0.005817344188690185, 0.005730303764343261, 0.005730368137359619, 0.005784512042999268, 0.005702655792236328, 0.0127457275390625, 0.005792799949645996, 0.005795807838439942, 0.005763072013854981, 0.005805056095123291, 0.005767168045043946, 0.005755904197692871, 0.005756927967071533, 0.005776383876800537, 0.005758975982666016, 0.00576204776763916, 0.005752831935882568, 0.005768191814422607, 0.005756927967071533, 0.005751808166503906, 0.005796864032745362, 0.005727231979370117, 0.005827583789825439, 0.005981184005737304, 0.005914624214172363, 0.005929984092712402, 0.005889023780822754, 0.005903359889984131, 0.005920767784118652, 0.005868544101715088, 0.006069248199462891, 0.005882880210876465, 0.005957664012908936, 0.005826528072357178, 0.005795839786529541, 0.005755904197692871, 0.0058009600639343266, 0.0057794561386108395, 0.005767168045043946, 0.005786623954772949, 0.005774335861206055, 0.005805056095123291, 0.005824512004852295, 0.005784575939178467, 0.005781504154205322, 0.005776383876800537, 0.005777408123016357, 0.00573747205734253, 0.005754879951477051, 0.005764095783233642, 0.005807104110717773, 0.005772287845611572, 0.00573747205734253, 0.00576204776763916, 0.005730303764343261, 0.005923840045928955, 0.005743616104125977, 0.0057712640762329105, 0.005740543842315674, 0.005757952213287353, 0.005730303764343261, 0.005743616104125977, 0.00573747205734253, 0.005757952213287353, 0.005758975982666016, 0.005754879951477051, 0.005726208209991455, 0.005718016147613526, 0.012769280433654785, 0.005763072013854981, 0.005708799839019775, 0.005767168045043946, 0.005732351779937744, 0.0057784318923950195, 0.005742591857910156, 0.005733376026153565, 0.005792768001556397, 0.005757952213287353, 0.005777408123016357, 0.005743616104125977, 0.005785600185394287, 0.005739552021026611, 0.005781472206115723, 0.005747712135314941, 0.005754879951477051, 0.005725183963775635, 0.005803008079528809, 0.005763072013854981, 0.005772287845611572, 0.005780479907989502, 0.005734399795532226, 0.005776383876800537, 0.005764095783233642, 0.005754879951477051, 0.0057415680885314945, 0.005744639873504639, 0.005756927967071533, 0.005758975982666016, 0.0058091521263122555, 0.0057415680885314945, 0.00577023983001709, 0.0057784318923950195, 0.005792799949645996, 0.005776351928710938, 0.005723135948181152, 0.005757952213287353, 0.005734399795532226, 0.005773312091827393, 0.005738495826721191, 0.005747712135314941, 0.005718016147613526, 0.005751808166503906, 0.0057712640762329105, 0.005718016147613526, 0.005763072013854981, 0.005733376026153565, 0.00573747205734253, 0.00572108793258667, 0.00571289587020874, 0.005733376026153565, 0.00572108793258667, 0.0057794561386108395, 0.005745664119720459, 0.005787648200988769, 0.005700607776641845, 0.005739552021026611, 0.005824480056762695, 0.005736447811126709, 0.005751808166503906, 0.005730303764343261, 0.005746687889099121, 0.012750847816467285, 0.005755904197692871, 0.005814271926879883, 0.005780479907989502, 0.005761023998260498, 0.005724160194396972, 0.005742591857910156, 0.005725183963775635, 0.005744639873504639, 0.005799935817718506, 0.005746687889099121, 0.005753856182098388, 0.005734399795532226, 0.005761023998260498, 0.0057487359046936035, 0.005768191814422607, 0.005758975982666016, 0.005709824085235596, 0.0057497601509094234, 0.005745664119720459, 0.005758975982666016, 0.00573747205734253, 0.005736447811126709, 0.005742591857910156, 0.005724160194396972, 0.005777408123016357, 0.005728256225585937, 0.005971968173980713, 0.005884928226470947, 0.005913599967956543, 0.0058787841796875, 0.005979135990142822, 0.00588595199584961, 0.005915711879730225, 0.005879744052886963, 0.005907455921173096, 0.005915647983551026, 0.0059269118309021, 0.005891071796417236, 0.005909503936767578, 0.005876736164093017, 0.005825535774230957, 0.0059269118309021, 0.005896192073822021, 0.005946368217468262, 0.005956607818603516, 0.00602726411819458, 0.005993472099304199, 0.006072319984436035, 0.0060067839622497555, 0.005928959846496582, 0.005882880210876465, 0.005945343971252442, 0.005874688148498535, 0.0059054079055786135, 0.005881855964660645, 0.005921792030334472, 0.005997568130493164, 0.005953536033630371, 0.005892096042633056, 0.005933055877685547, 0.005854207992553711, 0.0058757119178771975, 0.013130751609802246, 0.005893119812011719, 0.005895167827606201, 0.005856256008148194, 0.005925888061523437, 0.00591871976852417, 0.005921792030334472, 0.005856256008148194, 0.005917695999145508, 0.00591974401473999, 0.006102015972137451, 0.0058122239112854005, 0.005824512004852295, 0.005742591857910156, 0.0057487359046936035, 0.005742591857910156, 0.005768191814422607, 0.005728256225585937, 0.005750783920288086, 0.0057487359046936035, 0.005753856182098388, 0.00576204776763916, 0.005736447811126709, 0.005752831935882568, 0.0057712640762329105, 0.005727231979370117, 0.005761023998260498, 0.005774335861206055, 0.0057497601509094234, 0.005799935817718506, 0.005754879951477051, 0.005767168045043946, 0.005761023998260498, 0.005799935817718506, 0.005750783920288086, 0.005765120029449463, 0.005734399795532226, 0.005767168045043946, 0.005754879951477051, 0.00576204776763916, 0.005787648200988769, 0.0057825279235839844, 0.005806079864501953, 0.0057415680885314945, 0.005856256008148194, 0.005835775852203369, 0.005775360107421875, 0.005743616104125977, 0.005932032108306885, 0.0058788161277771, 0.005862368106842041, 0.005996543884277344, 0.0059023361206054685, 0.006011903762817383, 0.005892096042633056, 0.0058429441452026365, 0.005720064163208007, 0.005767168045043946, 0.005764095783233642, 0.0057415680885314945, 0.005739520072937012, 0.005739520072937012, 0.005750783920288086, 0.012777471542358398, 0.005768191814422607, 0.005728256225585937, 0.005753856182098388, 0.005766143798828125, 0.005764128208160401, 0.005780447959899902, 0.005723135948181152, 0.005791744232177734, 0.0057487359046936035, 0.005765120029449463, 0.0057292799949646, 0.005982207775115967, 0.0059996161460876465, 0.005895167827606201, 0.005964799880981446, 0.005898240089416504, 0.005934080123901367, 0.005732351779937744, 0.005817344188690185, 0.005769216060638428, 0.005758975982666016, 0.005766143798828125, 0.00576204776763916, 0.0057825279235839844, 0.005733376026153565, 0.005764095783233642, 0.005757952213287353, 0.005760000228881836, 0.005856256008148194, 0.005764095783233642, 0.00576204776763916, 0.00572108793258667, 0.005761023998260498, 0.0057569918632507325, 0.005774271965026855, 0.005758975982666016, 0.005766143798828125, 0.005760000228881836, 0.005735424041748047, 0.005757952213287353, 0.005744639873504639, 0.005787648200988769, 0.0057497601509094234, 0.005758975982666016, 0.0057415680885314945, 0.005766143798828125, 0.005772319793701172, 0.006099936008453369, 0.0059658241271972655, 0.005901311874389649, 0.005900288105010986, 0.005992447853088379, 0.005878848075866699, 0.005979072093963623, 0.0059361281394958495, 0.005946368217468262, 0.005957632064819336, 0.00591974401473999, 0.005938176155090332, 0.005730303764343261, 0.005777408123016357, 0.005755904197692871, 0.012800000190734863, 0.005804096221923828, 0.005726143836975097, 0.005792768001556397, 0.005727231979370117, 0.005804031848907471, 0.00577023983001709, 0.005755904197692871, 0.005745664119720459, 0.005772287845611572, 0.005857279777526855, 0.005735424041748047, 0.005829631805419922, 0.005734399795532226, 0.005819392204284668, 0.005744639873504639, 0.0057794561386108395, 0.0057487359046936035, 0.005735424041748047, 0.005780479907989502, 0.005752831935882568, 0.005794816017150879, 0.005734399795532226, 0.005790719985961914, 0.005763072013854981, 0.005772287845611572, 0.005796864032745362, 0.005761023998260498, 0.005886975765228271, 0.005777440071105957, 0.005785568237304688, 0.005776383876800537, 0.005776383876800537, 0.005794816017150879, 0.005760000228881836, 0.005789696216583252, 0.005745664119720459, 0.005797887802124023, 0.005775360107421875, 0.005755904197692871, 0.0057825279235839844, 0.005760000228881836, 0.005765120029449463, 0.005764095783233642, 0.0057712640762329105, 0.005740608215332031, 0.00579475212097168, 0.00576204776763916, 0.00573747205734253, 0.005764095783233642, 0.005746687889099121, 0.005781504154205322, 0.005780479907989502, 0.00577023983001709, 0.005775360107421875, 0.005899263858795166, 0.005807104110717773, 0.005816319942474365, 0.005773312091827393, 0.005774335861206055, 0.0057487359046936035, 0.0057712640762329105, 0.005725183963775635, 0.012851200103759765, 0.00582041597366333, 0.005773312091827393, 0.005750783920288086, 0.005769216060638428, 0.005747712135314941, 0.005793791770935058, 0.005773312091827393, 0.005785600185394287, 0.005780479907989502, 0.00576204776763916, 0.0057712640762329105, 0.005791744232177734, 0.005795839786529541, 0.005742591857910156, 0.005819392204284668, 0.005777408123016357, 0.0057415680885314945, 0.005755904197692871, 0.005774335861206055, 0.005767168045043946, 0.005753920078277588, 0.005764031887054443, 0.005751808166503906, 0.005785600185394287, 0.005758975982666016, 0.005801983833312988, 0.005848063945770264, 0.0058388481140136715, 0.005791744232177734, 0.005767231941223145, 0.005747647762298584, 0.006312960147857666, 0.0058726401329040525, 0.0057712640762329105, 0.00576313591003418, 0.005959616184234619, 0.005955584049224853, 0.00591974401473999, 0.005994527816772461, 0.005947360038757324, 0.005980160236358643, 0.005901311874389649, 0.0059658241271972655, 0.005849088191986084, 0.005836800098419189, 0.005766143798828125, 0.005752831935882568, 0.005751808166503906, 0.005766143798828125, 0.0057190399169921875, 0.005752831935882568, 0.005769216060638428, 0.005763072013854981, 0.005774335861206055, 0.005859327793121338, 0.005818367958068848, 0.005773312091827393, 0.005792768001556397, 0.005735424041748047, 0.005754879951477051, 0.0057497601509094234, 0.0057415680885314945]",tokens/s,167.30632644435158,,,main,False,False,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,databricks/dbrx-base,databricks/dbrx-base,cuda,0,42,,,True,,,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",dbrx,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-350m,,cuda,0,42,,,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,d35829e539df8480b726c647eeabf91e41eae047,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,,MB,1175.089152,1455.947776,0.0,809.500672,723.533824,s,15,0.17660559940338133,0.011773706626892089,0.0005060809352526505,0.01188105583190918,0.011992940521240234,0.012405353450775146,0.013093141155242918,"[0.013265088081359863, 0.011219743728637696, 0.011262335777282716, 0.011187264442443848, 0.011892576217651368, 0.01188105583190918, 0.011240384101867677, 0.012036895751953125, 0.011927007675170898, 0.011889760017395019, 0.011867008209228515, 0.011883199691772461, 0.011868415832519531, 0.011281344413757324, 0.01190351963043213]",tokens/s,21743.364949766583,kWh,1.3865128598706948e-07,7.597172755684907e-08,3.956406399342899e-07,6.102636534782083e-07,tokens/kWh,419490819.32198244,MB,1175.089152,1455.947776,0.0,809.500672,749.216256,s,15,10.484396118164064,0.6989597412109376,0.010546266198019412,0.6962911987304687,0.7025856811523438,0.7139740173339844,0.7317457800292969,"[0.736188720703125, 0.6892533569335938, 0.69624951171875, 0.6921040649414063, 0.6986005249023437, 0.698226806640625, 0.6946048583984376, 0.6997840576171875, 0.6971741943359375, 0.6962911987304687, 0.694924072265625, 0.6990907592773438, 0.6946989135742188, 0.6927516479492187, 0.7044534301757812]",tokens/s,90.13394661451233,kWh,7.764596984365612e-06,4.254614382124614e-06,1.4961130904734307e-05,2.6980342271224535e-05,tokens/kWh,2335033.3871483784,,s,944,10.635618295669536,0.011266544804734699,0.0014591324737404984,0.01107148838043213,0.011258982372283935,0.011755068683624267,0.02223625762939453,"[0.012138496398925782, 0.012057600021362304, 0.011867136001586913, 0.01165721607208252, 0.011790335655212402, 0.011788288116455077, 0.011667455673217773, 0.011686911582946777, 0.012061696052551269, 0.011962368011474609, 0.011684864044189454, 0.011777024269104003, 0.011752511978149414, 0.01197868824005127, 0.011803647994995118, 0.012000255584716797, 0.011875328063964843, 0.011687935829162598, 0.011875328063964843, 0.01174937629699707, 0.011942912101745605, 0.011983872413635254, 0.011774016380310058, 0.0118721923828125, 0.01233407974243164, 0.011925503730773926, 0.011716608047485352, 0.011866111755371094, 0.011828224182128906, 0.011709440231323242, 0.011905023574829102, 0.01174835205078125, 0.011743231773376465, 0.012090368270874024, 0.012113920211791992, 0.01176576042175293, 0.011701312065124512, 0.011774911880493164, 0.011737088203430175, 0.011653120040893555, 0.011552767753601074, 0.011234304428100587, 0.011653120040893555, 0.011614208221435546, 0.011580415725708008, 0.01183027172088623, 0.011583488464355468, 0.011491328239440919, 0.011707391738891602, 0.011467776298522948, 0.011268095970153809, 0.01115340805053711, 0.011082752227783203, 0.011146240234375, 0.011140095710754394, 0.011147263526916504, 0.011131903648376466, 0.01115340805053711, 0.011222016334533692, 0.011147263526916504, 0.011168800354003906, 0.01108784008026123, 0.02211942481994629, 0.010397695541381835, 0.010449919700622558, 0.010497023582458496, 0.010392576217651368, 0.010460160255432128, 0.010432512283325195, 0.01045094394683838, 0.010445823669433594, 0.010406911849975586, 0.010437631607055664, 0.010470399856567383, 0.01042636775970459, 0.010480640411376953, 0.010445823669433594, 0.010461183547973632, 0.010458111763000488, 0.010431488037109376, 0.010424320220947265, 0.01043455982208252, 0.010455039978027344, 0.01093017578125, 0.011171839714050292, 0.011189248085021973, 0.011116543769836425, 0.011207679748535156, 0.011160575866699218, 0.011148287773132324, 0.011110400199890137, 0.011133952140808106, 0.011155455589294434, 0.011123711585998536, 0.011090944290161133, 0.011088895797729491, 0.011189248085021973, 0.011101183891296386, 0.011117568016052246, 0.011094016075134277, 0.011119615554809571, 0.011193344116210938, 0.01113804817199707, 0.011147263526916504, 0.011320320129394532, 0.011206656455993653, 0.01112166404724121, 0.011126784324645997, 0.011096063613891602, 0.01107148838043213, 0.011132927894592285, 0.011144191741943359, 0.011135999679565429, 0.011383808135986329, 0.01175551986694336, 0.01153331184387207, 0.011256832122802735, 0.011196415901184082, 0.0111595516204834, 0.011420672416687011, 0.011188223838806152, 0.011104255676269532, 0.01116262435913086, 0.01116262435913086, 0.011131903648376466, 0.02213580894470215, 0.010478591918945313, 0.010484736442565918, 0.011142144203186035, 0.011163647651672364, 0.011189248085021973, 0.011179007530212403, 0.011090944290161133, 0.011057151794433593, 0.01115443229675293, 0.011074560165405273, 0.011171839714050292, 0.01111244773864746, 0.011139072418212891, 0.011106304168701172, 0.011200511932373047, 0.011057151794433593, 0.010911744117736816, 0.011289600372314454, 0.010860544204711914, 0.01103052806854248, 0.011191295623779298, 0.011142144203186035, 0.011233280181884766, 0.01112377643585205, 0.011103167533874511, 0.011147263526916504, 0.011070528030395509, 0.011084735870361328, 0.011485183715820312, 0.011078656196594238, 0.011074560165405273, 0.011172863960266113, 0.011123711585998536, 0.01103667163848877, 0.011100159645080567, 0.011033599853515624, 0.011095040321350098, 0.01107148838043213, 0.010899456024169921, 0.010878975868225099, 0.010903552055358886, 0.010842111587524414, 0.010879008293151856, 0.010818528175354005, 0.01083084774017334, 0.010789888381958008, 0.01083903980255127, 0.010860544204711914, 0.011130880355834961, 0.011092991828918456, 0.011107328414916993, 0.01112166404724121, 0.011177984237670899, 0.011076607704162598, 0.011125760078430176, 0.011097087860107421, 0.011111424446105958, 0.0111278076171875, 0.01112883186340332, 0.011160575866699218, 0.011094016075134277, 0.0110448637008667, 0.02204979133605957, 0.010391551971435547, 0.01081651210784912, 0.011019264221191406, 0.01102847957611084, 0.011114496231079102, 0.011080703735351562, 0.011111424446105958, 0.011027520179748534, 0.01104684829711914, 0.011084799766540527, 0.011072544097900391, 0.011072480201721191, 0.011060223579406739, 0.011164671897888183, 0.011094016075134277, 0.011081727981567382, 0.011073535919189453, 0.011040767669677735, 0.011068415641784669, 0.011085824012756347, 0.011061247825622558, 0.011024383544921875, 0.011081727981567382, 0.010852352142333984, 0.010819583892822266, 0.010806271553039551, 0.010788864135742187, 0.010807295799255372, 0.010853407859802246, 0.011113439559936523, 0.01116262435913086, 0.011104255676269532, 0.011017215728759766, 0.011003904342651367, 0.011041791915893554, 0.011025407791137695, 0.011066368103027344, 0.0110448637008667, 0.01103769588470459, 0.01102950382232666, 0.011049983978271484, 0.011041791915893554, 0.01104691219329834, 0.011077631950378418, 0.011032575607299805, 0.011010047912597656, 0.010844160079956054, 0.010813440322875977, 0.010787839889526368, 0.010875904083251953, 0.011091967582702637, 0.011053055763244628, 0.011053055763244628, 0.011111424446105958, 0.01101414394378662, 0.010912768363952637, 0.010781696319580078, 0.010753024101257324, 0.010797056198120117, 0.01083187198638916, 0.011004927635192872, 0.011039744377136231, 0.023410688400268553, 0.011052032470703126, 0.01105510425567627, 0.011003904342651367, 0.011023360252380371, 0.011087871551513672, 0.011041791915893554, 0.011075584411621094, 0.011066399574279785, 0.011058143615722656, 0.011043840408325196, 0.011074560165405273, 0.011161600112915039, 0.011259903907775879, 0.01107968044281006, 0.011074560165405273, 0.011040767669677735, 0.011058176040649414, 0.011141119956970215, 0.011090944290161133, 0.011088895797729491, 0.011089920043945312, 0.011066368103027344, 0.01115443229675293, 0.01107968044281006, 0.01114521598815918, 0.011061247825622558, 0.011072511672973634, 0.011019264221191406, 0.01107049560546875, 0.011074527740478516, 0.011058176040649414, 0.011052032470703126, 0.011066368103027344, 0.010970111846923827, 0.011059200286865235, 0.01102847957611084, 0.011056127548217774, 0.011078656196594238, 0.01115135955810547, 0.011070464134216309, 0.0111278076171875, 0.011064319610595704, 0.01104691219329834, 0.011062272071838379, 0.011057151794433593, 0.011061247825622558, 0.011083776473999024, 0.011062272071838379, 0.011045887947082519, 0.011051008224487305, 0.011097087860107421, 0.011019264221191406, 0.011114496231079102, 0.011066368103027344, 0.011139072418212891, 0.011181056022644043, 0.011150336265563965, 0.011089920043945312, 0.011372544288635255, 0.011087871551513672, 0.011062272071838379, 0.011053055763244628, 0.023417856216430662, 0.011049983978271484, 0.011092991828918456, 0.01105510425567627, 0.011049983978271484, 0.011139072418212891, 0.011018239974975585, 0.011054080009460449, 0.011051008224487305, 0.01103052806854248, 0.011158592224121093, 0.011132863998413086, 0.011143168449401856, 0.011181056022644043, 0.01115135955810547, 0.011072511672973634, 0.011059200286865235, 0.011190272331237794, 0.01112063980102539, 0.0110448637008667, 0.011105279922485351, 0.011070464134216309, 0.011113471984863281, 0.011107328414916993, 0.011034624099731445, 0.011074560165405273, 0.011105279922485351, 0.011081727981567382, 0.011089920043945312, 0.011096063613891602, 0.011076607704162598, 0.011095040321350098, 0.011051008224487305, 0.011147263526916504, 0.011059200286865235, 0.011070464134216309, 0.01103667163848877, 0.011065343856811523, 0.011130880355834961, 0.011054080009460449, 0.011077631950378418, 0.011048959732055665, 0.011065343856811523, 0.011052032470703126, 0.011113471984863281, 0.011072511672973634, 0.011115519523620606, 0.011100159645080567, 0.011015168190002441, 0.011144191741943359, 0.0110448637008667, 0.011076607704162598, 0.011132927894592285, 0.011057151794433593, 0.011021311759948731, 0.010954751968383789, 0.010874879837036134, 0.01084620761871338, 0.011087871551513672, 0.011068415641784669, 0.011119615554809571, 0.01111244773864746, 0.011054080009460449, 0.022202367782592772, 0.010402815818786621, 0.010449919700622558, 0.011032575607299805, 0.011088895797729491, 0.01104691219329834, 0.011006976127624512, 0.011085824012756347, 0.011072511672973634, 0.011025407791137695, 0.011027456283569336, 0.01103667163848877, 0.011070464134216309, 0.011086848258972168, 0.010819583892822266, 0.010833984375, 0.01081439971923828, 0.01095372772216797, 0.01083187198638916, 0.011095104217529296, 0.011084735870361328, 0.0110448637008667, 0.01112985610961914, 0.011302911758422851, 0.011109375953674316, 0.01101414394378662, 0.011039744377136231, 0.011113471984863281, 0.011100159645080567, 0.011259903907775879, 0.011041791915893554, 0.011059200286865235, 0.011110400199890137, 0.011118592262268067, 0.011064319610595704, 0.011016223907470703, 0.011090911865234376, 0.011254783630371093, 0.011076607704162598, 0.011010047912597656, 0.01116262435913086, 0.011091967582702637, 0.011072511672973634, 0.011025407791137695, 0.0110448637008667, 0.01107148838043213, 0.011069439888000488, 0.011080767631530762, 0.011090880393981933, 0.011061247825622558, 0.01092403221130371, 0.010891263961791992, 0.010870783805847169, 0.010827775955200195, 0.01084108829498291, 0.011094016075134277, 0.011086879730224609, 0.011106271743774414, 0.011158528327941895, 0.011142144203186035, 0.011083776473999024, 0.011052032470703126, 0.01102847957611084, 0.023323648452758788, 0.011080767631530762, 0.010983455657958984, 0.011079584121704102, 0.011096063613891602, 0.011982848167419433, 0.011134976387023926, 0.010999808311462403, 0.011023360252380371, 0.011230208396911622, 0.011049983978271484, 0.011078656196594238, 0.010974207878112792, 0.011049983978271484, 0.011017215728759766, 0.01213644790649414, 0.011421695709228515, 0.01214566421508789, 0.011182080268859864, 0.011249664306640626, 0.011106304168701172, 0.011099136352539063, 0.011016192436218262, 0.010991616249084473, 0.011004927635192872, 0.010983424186706543, 0.01100595188140869, 0.011000831604003907, 0.011051008224487305, 0.011083776473999024, 0.01101414394378662, 0.010982399940490722, 0.011083776473999024, 0.011109375953674316, 0.01105510425567627, 0.011027456283569336, 0.010991616249084473, 0.010999808311462403, 0.01102233600616455, 0.011200511932373047, 0.011125760078430176, 0.011000831604003907, 0.011080703735351562, 0.011008000373840332, 0.011061247825622558, 0.010998784065246582, 0.010796031951904296, 0.010761216163635253, 0.010984448432922364, 0.011100159645080567, 0.0111278076171875, 0.01108176040649414, 0.011106271743774414, 0.011104255676269532, 0.011049983978271484, 0.010997759819030761, 0.011039744377136231, 0.011043840408325196, 0.011081727981567382, 0.011049983978271484, 0.011004927635192872, 0.011086848258972168, 0.011081727981567382, 0.02349772834777832, 0.010971136093139648, 0.011004927635192872, 0.011001855850219726, 0.011040767669677735, 0.011041791915893554, 0.010982399940490722, 0.011059200286865235, 0.011111424446105958, 0.011010047912597656, 0.011004927635192872, 0.011035648345947266, 0.011043904304504394, 0.011024319648742675, 0.010991616249084473, 0.011039744377136231, 0.011060223579406739, 0.011073535919189453, 0.011016192436218262, 0.011066399574279785, 0.011066335678100586, 0.011020288467407227, 0.010982399940490722, 0.011085824012756347, 0.011023360252380371, 0.011041791915893554, 0.01103052806854248, 0.011102208137512207, 0.011045887947082519, 0.011048959732055665, 0.011117568016052246, 0.011016192436218262, 0.011064319610595704, 0.011072511672973634, 0.01103769588470459, 0.011317248344421387, 0.011150336265563965, 0.011147263526916504, 0.011049983978271484, 0.011048959732055665, 0.010994688034057617, 0.011033599853515624, 0.011058176040649414, 0.011073535919189453, 0.011021311759948731, 0.011062272071838379, 0.011096063613891602, 0.011086848258972168, 0.01101414394378662, 0.01102233600616455, 0.01104793643951416, 0.01103667163848877, 0.01103872013092041, 0.011034624099731445, 0.01122815990447998, 0.01126912021636963, 0.011086848258972168, 0.011060223579406739, 0.011107328414916993, 0.011076607704162598, 0.01103052806854248, 0.011092991828918456, 0.011011072158813476, 0.023358463287353515, 0.010990592002868652, 0.011009023666381837, 0.011004927635192872, 0.011003904342651367, 0.010980352401733399, 0.010906623840332032, 0.011048959732055665, 0.011114496231079102, 0.011043840408325196, 0.010972160339355469, 0.010981375694274903, 0.011039744377136231, 0.011023360252380371, 0.010992639541625977, 0.010978303909301757, 0.010966015815734862, 0.011045887947082519, 0.011019264221191406, 0.01096396827697754, 0.011002880096435547, 0.011033599853515624, 0.011007040023803712, 0.011027392387390136, 0.011003904342651367, 0.01103052806854248, 0.01103052806854248, 0.011103232383728028, 0.011019264221191406, 0.011002880096435547, 0.010986559867858887, 0.010992575645446777, 0.011026432037353515, 0.011149312019348144, 0.011064319610595704, 0.01103769588470459, 0.011196415901184082, 0.011034624099731445, 0.01100595188140869, 0.01100595188140869, 0.011006976127624512, 0.01102950382232666, 0.01103667163848877, 0.011004927635192872, 0.011056127548217774, 0.011084799766540527, 0.011019264221191406, 0.011095040321350098, 0.011042816162109375, 0.011066368103027344, 0.011100159645080567, 0.011077664375305175, 0.011090911865234376, 0.011150336265563965, 0.01111244773864746, 0.011092991828918456, 0.011107328414916993, 0.011163647651672364, 0.011117568016052246, 0.011105279922485351, 0.011048959732055665, 0.011288576126098633, 0.01112166404724121, 0.023423999786376954, 0.011000831604003907, 0.010979328155517578, 0.011001855850219726, 0.011026432037353515, 0.011033599853515624, 0.010976256370544434, 0.010976256370544434, 0.01102847957611084, 0.011041791915893554, 0.011056127548217774, 0.01105510425567627, 0.010995712280273438, 0.010989567756652833, 0.01100595188140869, 0.010988544464111329, 0.010991616249084473, 0.011126784324645997, 0.011016192436218262, 0.010951711654663085, 0.011007967948913575, 0.011049983978271484, 0.010992639541625977, 0.010916864395141602, 0.010827775955200195, 0.010815520286560059, 0.01099465560913086, 0.011035648345947266, 0.010984448432922364, 0.011013119697570802, 0.01095680046081543, 0.010994688034057617, 0.01102847957611084, 0.011065376281738282, 0.011089887619018555, 0.011064319610595704, 0.011024383544921875, 0.011042816162109375, 0.011025407791137695, 0.010976256370544434, 0.011019264221191406, 0.01104691219329834, 0.011012096405029297, 0.011039744377136231, 0.011031552314758301, 0.011027456283569336, 0.011076640129089356, 0.01110422420501709, 0.011056127548217774, 0.01102847957611084, 0.011065343856811523, 0.011042816162109375, 0.011115519523620606, 0.011028544425964355, 0.011113408088684083, 0.011076607704162598, 0.01105510425567627, 0.011061280250549317, 0.01109603214263916, 0.011019264221191406, 0.011023360252380371, 0.011065343856811523, 0.011039744377136231, 0.023407615661621094, 0.010978303909301757, 0.010981375694274903, 0.011058176040649414, 0.011073535919189453, 0.011033599853515624, 0.0109486083984375, 0.011002911567687988, 0.010992608070373534, 0.010993663787841796, 0.011011072158813476, 0.0110632963180542, 0.011070464134216309, 0.011092991828918456, 0.011156479835510253, 0.01144217586517334, 0.011598848342895507, 0.011254783630371093, 0.011147263526916504, 0.011066368103027344, 0.011118592262268067, 0.011103232383728028, 0.01112063980102539, 0.011068415641784669, 0.01103052806854248, 0.011039744377136231, 0.011169792175292969, 0.011075615882873535, 0.011022303581237793, 0.011060223579406739, 0.011160575866699218, 0.011082752227783203, 0.011065343856811523, 0.011053055763244628, 0.011033599853515624, 0.011076607704162598, 0.011049983978271484, 0.011026432037353515, 0.011016223907470703, 0.011038687705993653, 0.011024383544921875, 0.011085824012756347, 0.01111251163482666, 0.011107263565063476, 0.011247615814208984, 0.011100159645080567, 0.011042816162109375, 0.011062335968017579, 0.011111359596252441, 0.011152383804321288, 0.011080703735351562, 0.011119615554809571, 0.011057151794433593, 0.011068415641784669, 0.011021311759948731, 0.01104793643951416, 0.011161600112915039, 0.011126784324645997, 0.011076607704162598, 0.011054080009460449, 0.011202560424804688, 0.011080703735351562, 0.011062272071838379, 0.023365631103515624, 0.010972160339355469, 0.01102950382232666, 0.011017215728759766, 0.011016192436218262, 0.01102847957611084, 0.011011072158813476, 0.011003904342651367, 0.011016192436218262, 0.011031552314758301, 0.011053088188171387, 0.011050975799560547, 0.011039744377136231, 0.011001855850219726, 0.011066368103027344, 0.01112985610961914, 0.010855423927307128, 0.01092300796508789, 0.010872832298278809, 0.010899456024169921, 0.010778623580932617, 0.010829824447631836, 0.010994688034057617, 0.011018239974975585, 0.011035648345947266, 0.011034624099731445, 0.011143168449401856, 0.011077664375305175, 0.011027423858642579, 0.01105510425567627, 0.011064319610595704, 0.01108790397644043, 0.011046879768371582, 0.01101414394378662, 0.011060223579406739, 0.011064319610595704, 0.011098112106323242, 0.011034624099731445, 0.011096063613891602, 0.01104793643951416, 0.011021311759948731, 0.011065343856811523, 0.011182080268859864, 0.01112166404724121, 0.011107328414916993, 0.011080703735351562, 0.011080703735351562, 0.01113702392578125, 0.011103232383728028, 0.011082752227783203, 0.011018303871154786, 0.011081664085388184, 0.011163647651672364, 0.011080767631530762, 0.011104191780090332, 0.011074560165405273, 0.010855423927307128, 0.010910719871520995, 0.01083903980255127, 0.010877951622009278, 0.010870783805847169, 0.01080832004547119, 0.010946559906005859, 0.022261823654174805, 0.010372032165527343, 0.01041305637359619, 0.010391551971435547, 0.010406911849975586, 0.0104017915725708, 0.010338303565979003, 0.010470399856567383, 0.010395648002624512, 0.010359807968139649, 0.010357760429382324, 0.010380288124084473, 0.010528767585754394, 0.011116543769836425, 0.011110400199890137, 0.011115519523620606, 0.011082752227783203, 0.01115443229675293, 0.011058176040649414, 0.011031552314758301, 0.011048959732055665, 0.011095040321350098, 0.011101183891296386, 0.0110632963180542, 0.011089920043945312, 0.01107148838043213, 0.011062272071838379, 0.011088895797729491, 0.01105510425567627, 0.011023360252380371, 0.011111424446105958, 0.011047967910766602, 0.011032544136047363, 0.011105279922485351, 0.011359231948852539, 0.011072511672973634, 0.01105510425567627, 0.011084799766540527, 0.011080703735351562, 0.011078656196594238, 0.012894207954406739, 0.01135206413269043, 0.011204607963562012, 0.011106304168701172, 0.011083776473999024, 0.011175935745239257, 0.011152383804321288, 0.01111244773864746, 0.0110448637008667, 0.011191295623779298, 0.011116543769836425, 0.011083776473999024, 0.01107968044281006, 0.011076640129089356, 0.011076576232910156, 0.011108351707458495, 0.01112063980102539, 0.011068415641784669, 0.011074560165405273, 0.011169792175292969, 0.011084832191467286, 0.011101152420043945, 0.011115519523620606, 0.023419904708862304, 0.011096063613891602, 0.011101183891296386, 0.011091967582702637, 0.011140095710754394, 0.01112063980102539, 0.011058176040649414, 0.011149312019348144, 0.011003904342651367, 0.011060223579406739, 0.011060223579406739, 0.011188223838806152, 0.011062272071838379, 0.01140121555328369, 0.01126912021636963, 0.011206720352172852, 0.011110336303710937, 0.011169792175292969, 0.011081727981567382, 0.011107328414916993, 0.011073535919189453, 0.011057151794433593, 0.011122688293457032, 0.011110400199890137, 0.011061247825622558, 0.011032575607299805, 0.011084799766540527, 0.011122688293457032, 0.011084863662719727, 0.01113593578338623, 0.011147263526916504, 0.01306214427947998, 0.012184576034545898, 0.011279359817504882, 0.011052032470703126, 0.011067392349243164, 0.011080703735351562, 0.011072511672973634, 0.011158592224121093, 0.011073472023010254, 0.011085824012756347, 0.011311103820800781, 0.011406335830688476, 0.011143168449401856, 0.0111278076171875, 0.011105279922485351, 0.011146240234375, 0.011101183891296386, 0.011067392349243164, 0.01120358371734619, 0.011150336265563965, 0.011087871551513672, 0.011042816162109375, 0.011045887947082519, 0.011133952140808106, 0.011110400199890137, 0.011117568016052246, 0.011411456108093262, 0.011166720390319825, 0.011073535919189453, 0.011205632209777832, 0.011075584411621094, 0.011076607704162598]",tokens/s,88.75835647320693,,,main,False,False,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,-,-,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 106, in _inner_fn validate_repo_id(arg_value) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 160, in validate_repo_id raise HFValidationError( huggingface_hub.errors.HFValidationError: Repo id must use alphanumeric chars or '-', '_', '.', '--' and '..' are forbidden, '-' and '.' cannot start or end the name, max length is 96: '-'. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 466, in cached_file raise EnvironmentError( OSError: Incorrect path_or_model_id: '-'. Please provide either the path to a local folder or the repo_id of a model on the Hub. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,stabilityai/stablelm-2-12b,stabilityai/stablelm-2-12b,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ self.load_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 250, in load_model_with_no_weights self.pretrained_model = self.automodel_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3710, in from_pretrained model = cls(config, *model_args, **model_kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/stablelm/modeling_stablelm.py"", line 1188, in __init__ self.model = StableLmModel(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/stablelm/modeling_stablelm.py"", line 966, in __init__ [StableLmDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/stablelm/modeling_stablelm.py"", line 966, in [StableLmDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/stablelm/modeling_stablelm.py"", line 748, in __init__ self.mlp = StableLmMLP(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/stablelm/modeling_stablelm.py"", line 206, in __init__ self.down_proj = nn.Linear(self.intermediate_size, self.hidden_size, bias=False) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/linear.py"", line 98, in __init__ self.weight = Parameter(torch.empty((out_features, in_features), **factory_kwargs)) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_device.py"", line 78, in __torch_function__ return func(*args, **kwargs) torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 136.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,v,v,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/v/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6693257b-69b429466361c54263047998;432f1032-ff4a-4ab5-bb62-6c1b1fcae31a) Repository Not Found for url: https://huggingface.co/v/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: v is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,openai-community/gpt2,openai-community/gpt2,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1241.489408,1032.323072,0.0,385.875968,318.907904,s,22,0.1727903051376343,0.007854104778983377,0.00021828714121974632,0.007849168062210084,0.008014716815948487,0.008073207807540892,0.008506698875427245,"[0.008621120452880859, 0.00801529598236084, 0.007841887950897217, 0.00789247989654541, 0.007945151805877685, 0.007904096126556397, 0.007792160034179687, 0.00763804817199707, 0.007616352081298828, 0.007600927829742431, 0.007618944168090821, 0.007842016220092774, 0.008009504318237305, 0.00767625617980957, 0.0076583361625671385, 0.007916768074035645, 0.008076255798339843, 0.007917215824127197, 0.007856319904327393, 0.007921696186065674, 0.007704671859741211, 0.007724800109863281]",tokens/s,32594.421287200632,kWh,9.232484657938282e-08,5.058973639364694e-08,2.2594636131248934e-07,3.688609442855191e-07,tokens/kWh,694028478.6611662,MB,1241.817088,1032.323072,0.0,385.875968,328.799232,s,22,10.20196258544922,0.4637255720658736,0.007148621931380267,0.46636041259765626,0.4706716918945313,0.47408568267822265,0.4757752774047852,"[0.4761787109375, 0.4681607666015625, 0.4679932861328125, 0.46901544189453126, 0.4742575988769531, 0.4637789001464844, 0.46674624633789064, 0.46934344482421875, 0.4509463806152344, 0.45601358032226563, 0.45271380615234375, 0.4674017639160156, 0.4605509033203125, 0.4532318115234375, 0.4636131896972656, 0.47081927490234377, 0.4678555908203125, 0.4659745788574219, 0.46723831176757813, 0.45941766357421876, 0.45926031494140623, 0.4514510192871094]",tokens/s,135.8562128013304,kWh,5.298799510899404e-06,2.9032766492392166e-06,9.577154915506028e-06,1.777923107564465e-05,tokens/kWh,3543460.3291872516,,s,1385,10.358330366611499,0.007478938892860275,0.0010302901439165216,0.007405600070953369,0.007520870304107666,0.007575955390930176,0.01531293697357178,"[0.008210432052612305, 0.008208383560180664, 0.007930912017822266, 0.007940063953399659, 0.007887872219085693, 0.008407039642333984, 0.007456768035888672, 0.007417856216430664, 0.007498752117156982, 0.0075345921516418455, 0.007508992195129394, 0.007489535808563232, 0.00748748779296875, 0.00749567985534668, 0.007513088226318359, 0.007448703765869141, 0.007488383769989014, 0.007420928001403809, 0.0074967041015625, 0.007464960098266602, 0.007458816051483155, 0.007475200176239013, 0.007497727870941162, 0.007476223945617676, 0.007479296207427978, 0.007468031883239746, 0.007554048061370849, 0.007468031883239746, 0.007485439777374267, 0.007532544136047363, 0.0074711041450500485, 0.00748748779296875, 0.007620607852935791, 0.007480319976806641, 0.007358463764190673, 0.007480319976806641, 0.007523327827453613, 0.007514111995697022, 0.007489535808563232, 0.007486464023590088, 0.007554048061370849, 0.0074670081138610836, 0.007485439777374267, 0.00744755220413208, 0.00749567985534668, 0.007479296207427978, 0.0074700798988342285, 0.007458816051483155, 0.007442431926727295, 0.007481344223022461, 0.00748748779296875, 0.00744652795791626, 0.007486464023590088, 0.007556096076965332, 0.0075008001327514645, 0.0074107198715209965, 0.007475168228149414, 0.007476223945617676, 0.007453695774078369, 0.007512063980102539, 0.007519231796264648, 0.007463935852050781, 0.016294912338256837, 0.0075008001327514645, 0.007459839820861816, 0.00744755220413208, 0.0074414081573486324, 0.007476223945617676, 0.007369887828826904, 0.0074934720993041996, 0.007459871768951416, 0.007434207916259766, 0.007356416225433349, 0.007384064197540283, 0.007391232013702393, 0.007457791805267334, 0.007592959880828858, 0.007480415821075439, 0.007444384098052978, 0.007438464164733887, 0.007566207885742187, 0.0075304961204528805, 0.007488512039184571, 0.007444672107696533, 0.00742790412902832, 0.007450623989105225, 0.007372928142547607, 0.007389056205749512, 0.007421120166778565, 0.00741161584854126, 0.0075437121391296385, 0.007527423858642578, 0.007377920150756836, 0.007331840038299561, 0.00738918399810791, 0.0073820161819458, 0.007414783954620361, 0.007562240123748779, 0.007416831970214844, 0.007407616138458252, 0.007349247932434082, 0.007367680072784424, 0.007329792022705078, 0.007396351814270019, 0.007366655826568603, 0.007359488010406494, 0.0073400321006774905, 0.007351295948028564, 0.007497727870941162, 0.007391232013702393, 0.007414783954620361, 0.007383039951324463, 0.007391295909881592, 0.007329728126525879, 0.007346176147460938, 0.007391232013702393, 0.007398399829864502, 0.007377920150756836, 0.007377920150756836, 0.007333888053894043, 0.007356544017791748, 0.007405439853668213, 0.0073994240760803225, 0.007408639907836914, 0.007370783805847168, 0.01571836757659912, 0.007368703842163086, 0.007349247932434082, 0.007390207767486572, 0.007467135906219482, 0.007430016040802002, 0.00742195177078247, 0.007599103927612305, 0.007445631980895996, 0.007404416084289551, 0.007460864067077637, 0.007384064197540283, 0.0073994240760803225, 0.007478335857391358, 0.007400383949279785, 0.007415808200836181, 0.00739737606048584, 0.0074065918922424315, 0.007400544166564942, 0.007381919860839844, 0.007377920150756836, 0.007443456172943115, 0.007334911823272705, 0.007452672004699707, 0.007458816051483155, 0.007394303798675537, 0.007386112213134765, 0.007352352142333984, 0.007349215984344483, 0.007325695991516113, 0.007451648235321045, 0.007362559795379638, 0.007358463764190673, 0.007388160228729248, 0.007391232013702393, 0.0073400321006774905, 0.007334911823272705, 0.007332863807678222, 0.007448575973510742, 0.007460864067077637, 0.007465983867645264, 0.0074291200637817386, 0.007432191848754883, 0.007584767818450928, 0.007450623989105225, 0.0074301438331604, 0.007502848148345947, 0.007518208026885987, 0.007435264110565186, 0.007336959838867187, 0.0074332160949707035, 0.00748748779296875, 0.0074629120826721195, 0.007432191848754883, 0.0074670081138610836, 0.007468031883239746, 0.0074332160949707035, 0.007467040061950684, 0.0074137282371521, 0.00742195177078247, 0.007442431926727295, 0.0074629120826721195, 0.007420928001403809, 0.01585459232330322, 0.007437312126159668, 0.007432191848754883, 0.00744652795791626, 0.0074414081573486324, 0.007483391761779785, 0.00753769588470459, 0.007520224094390869, 0.00749567985534668, 0.00748748779296875, 0.00743833589553833, 0.007208960056304932, 0.007267327785491944, 0.007501823902130127, 0.007581696033477783, 0.0074629120826721195, 0.0074301438331604, 0.007436287879943848, 0.007547967910766602, 0.007482304096221924, 0.007477248191833496, 0.007486591815948487, 0.007415679931640625, 0.007415808200836181, 0.0074629120826721195, 0.007437344074249268, 0.007532512187957764, 0.007426047801971435, 0.007444479942321777, 0.0074670081138610836, 0.0073768959045410155, 0.007427072048187256, 0.007468031883239746, 0.007187456130981445, 0.007189504146575928, 0.007227392196655274, 0.007185408115386963, 0.007252992153167725, 0.007164927959442138, 0.007333888053894043, 0.007463935852050781, 0.007476223945617676, 0.00743833589553833, 0.007443456172943115, 0.007535647869110107, 0.007463903903961182, 0.0074997758865356446, 0.007502848148345947, 0.00744755220413208, 0.007511040210723877, 0.007482367992401123, 0.007450623989105225, 0.007505919933319092, 0.00744652795791626, 0.007491583824157715, 0.0075304961204528805, 0.007411712169647216, 0.00748748779296875, 0.007504896163940429, 0.007476223945617676, 0.007478271961212158, 0.007477248191833496, 0.007541759967803955, 0.01589350414276123, 0.007507967948913574, 0.007458816051483155, 0.007450623989105225, 0.00754585599899292, 0.007468031883239746, 0.007540736198425293, 0.00748748779296875, 0.007451648235321045, 0.007493631839752197, 0.007493631839752197, 0.007513088226318359, 0.007535615921020508, 0.007514111995697022, 0.007552000045776367, 0.007485439777374267, 0.007580671787261963, 0.007615488052368164, 0.007436351776123047, 0.007497663974761963, 0.007489535808563232, 0.0074711041450500485, 0.007458816051483155, 0.007512063980102539, 0.007577600002288819, 0.007601151943206787, 0.007461887836456299, 0.007513088226318359, 0.007467040061950684, 0.007530464172363281, 0.007557184219360352, 0.007482304096221924, 0.007552000045776367, 0.007508992195129394, 0.007507967948913574, 0.007540736198425293, 0.007513088226318359, 0.007576576232910156, 0.007542784214019775, 0.0075038719177246095, 0.007544832229614258, 0.007511040210723877, 0.007528448104858398, 0.007498752117156982, 0.00748748779296875, 0.007511040210723877, 0.007561215877532959, 0.007485439777374267, 0.007562240123748779, 0.007491583824157715, 0.0075008001327514645, 0.007570432186126709, 0.007520256042480469, 0.007790592193603516, 0.007592959880828858, 0.007544832229614258, 0.007542784214019775, 0.007519231796264648, 0.007540768146514893, 0.007540703773498535, 0.007481344223022461, 0.007524352073669433, 0.007445504188537597, 0.01599897575378418, 0.007501887798309326, 0.00753657579421997, 0.007548927783966064, 0.007497727870941162, 0.007534656047821045, 0.007485375881195068, 0.007522304058074952, 0.007523327827453613, 0.007511040210723877, 0.007588863849639893, 0.0074967041015625, 0.007577600002288819, 0.007610367774963379, 0.007482367992401123, 0.007552000045776367, 0.007548927783966064, 0.007455743789672851, 0.007541759967803955, 0.007581696033477783, 0.008041472434997558, 0.007274496078491211, 0.007349247932434082, 0.0071792640686035155, 0.007160831928253173, 0.007165952205657959, 0.007175168037414551, 0.007184383869171143, 0.007147520065307617, 0.007364607810974121, 0.007542784214019775, 0.0074700798988342285, 0.0074403839111328125, 0.007475200176239013, 0.007451648235321045, 0.007463935852050781, 0.007449600219726562, 0.0074332160949707035, 0.007201791763305664, 0.007162879943847656, 0.007156735897064209, 0.007132160186767578, 0.0071485438346862796, 0.007173120021820068, 0.007165952205657959, 0.007174143791198731, 0.007136256217956543, 0.007163904190063477, 0.007150591850280762, 0.007146495819091797, 0.007163904190063477, 0.007136256217956543, 0.0073471999168395995, 0.007452672004699707, 0.007175168037414551, 0.007268352031707764, 0.007169023990631103, 0.00714137601852417, 0.007153664112091064, 0.007168000221252442, 0.007178271770477295, 0.007303135871887207, 0.007264256000518799, 0.01561292839050293, 0.007322624206542969, 0.007942143917083741, 0.0076605439186096195, 0.007515135765075683, 0.007536640167236328, 0.007544832229614258, 0.00820736026763916, 0.007490560054779053, 0.0074301438331604, 0.007490560054779053, 0.007459839820861816, 0.007418879985809326, 0.00753766393661499, 0.0074629120826721195, 0.007408639907836914, 0.007420928001403809, 0.0074403839111328125, 0.007428095817565918, 0.007746560096740723, 0.007481344223022461, 0.007475200176239013, 0.007463935852050781, 0.00753766393661499, 0.007357439994812012, 0.00742195177078247, 0.007453695774078369, 0.007432191848754883, 0.00742195177078247, 0.007431168079376221, 0.007469056129455566, 0.007422976016998291, 0.007422976016998291, 0.007436287879943848, 0.0074301438331604, 0.0074332160949707035, 0.007449600219726562, 0.007392255783081054, 0.0073820161819458, 0.0074414081573486324, 0.007422976016998291, 0.007425024032592774, 0.00743833589553833, 0.0074291200637817386, 0.007456768035888672, 0.007422976016998291, 0.007155712127685547, 0.007252992153167725, 0.007176191806793213, 0.007137279987335205, 0.007202847957611084, 0.007159776210784912, 0.00728985595703125, 0.007156735897064209, 0.007168000221252442, 0.007294976234436036, 0.007268352031707764, 0.0071792640686035155, 0.007160831928253173, 0.007174143791198731, 0.00713318395614624, 0.007146495819091797, 0.007159808158874512, 0.01528217601776123, 0.007168000221252442, 0.007355391979217529, 0.0074301438331604, 0.007398399829864502, 0.007449600219726562, 0.007452672004699707, 0.0074670081138610836, 0.007404543876647949, 0.007529471874237061, 0.007434239864349365, 0.007435264110565186, 0.007449600219726562, 0.007459839820861816, 0.007475200176239013, 0.00744652795791626, 0.007449600219726562, 0.007437312126159668, 0.007425024032592774, 0.007443456172943115, 0.007513088226318359, 0.007446591854095459, 0.007465919971466064, 0.007417888164520264, 0.007414752006530761, 0.0074106879234313965, 0.007535615921020508, 0.0074824318885803225, 0.007451648235321045, 0.007465919971466064, 0.00744755220413208, 0.007473152160644531, 0.007444479942321777, 0.007449600219726562, 0.007460864067077637, 0.007434239864349365, 0.007445504188537597, 0.007427072048187256, 0.007489535808563232, 0.0074199681282043455, 0.007444447994232177, 0.007472095966339111, 0.007437312126159668, 0.007444511890411377, 0.007437280178070068, 0.00748748779296875, 0.007414783954620361, 0.007450623989105225, 0.00745472002029419, 0.007428095817565918, 0.007515168190002442, 0.0074496641159057615, 0.007479231834411621, 0.007433184146881103, 0.0074414081573486324, 0.0074670081138610836, 0.007434239864349365, 0.007452672004699707, 0.00744652795791626, 0.007584767818450928, 0.007465983867645264, 0.0074414081573486324, 0.0074291200637817386, 0.015262720108032226, 0.007154687881469727, 0.007160831928253173, 0.007168000221252442, 0.007130112171173096, 0.007163904190063477, 0.007173120021820068, 0.007131135940551757, 0.007137279987335205, 0.007158783912658692, 0.0071526398658752445, 0.007127039909362793, 0.007150591850280762, 0.007139328002929687, 0.007144447803497315, 0.007156735897064209, 0.0071495680809020995, 0.007179327964782715, 0.00719046401977539, 0.007214079856872559, 0.007142399787902832, 0.007142399787902832, 0.0071188478469848635, 0.007137279987335205, 0.007160831928253173, 0.007150591850280762, 0.007131135940551757, 0.00709939193725586, 0.0071567678451538086, 0.007144415855407715, 0.007112703800201416, 0.007077919960021973, 0.00720688009262085, 0.007130112171173096, 0.007197696208953858, 0.0071157760620117185, 0.0071157760620117185, 0.007161856174468994, 0.00714035177230835, 0.0071485438346862796, 0.00714137601852417, 0.007108607769012451, 0.007146495819091797, 0.007159808158874512, 0.007163904190063477, 0.007178239822387696, 0.007155712127685547, 0.007174143791198731, 0.007151616096496582, 0.007142399787902832, 0.0071198720932006835, 0.007197696208953858, 0.007143424034118652, 0.007151616096496582, 0.007164927959442138, 0.007200767993927002, 0.007153664112091064, 0.007157760143280029, 0.007159808158874512, 0.007172095775604248, 0.007113728046417236, 0.007192575931549072, 0.007132160186767578, 0.015258624076843261, 0.007193600177764893, 0.007174143791198731, 0.007139328002929687, 0.007161856174468994, 0.007203839778900147, 0.0071567678451538086, 0.007160799980163574, 0.007165952205657959, 0.007178239822387696, 0.007171072006225586, 0.007120895862579346, 0.007138304233551026, 0.007154687881469727, 0.007175168037414551, 0.007162879943847656, 0.007217152118682861, 0.007143424034118652, 0.007160831928253173, 0.0071495680809020995, 0.007166975975036621, 0.007123968124389648, 0.007155712127685547, 0.00719155216217041, 0.0071495680809020995, 0.007143424034118652, 0.007201791763305664, 0.007166975975036621, 0.007154687881469727, 0.007423999786376953, 0.007489535808563232, 0.0073820161819458, 0.007490560054779053, 0.007434239864349365, 0.007419904232025146, 0.0074403839111328125, 0.0074997758865356446, 0.007415808200836181, 0.007418879985809326, 0.007449600219726562, 0.007482367992401123, 0.007423999786376953, 0.007417856216430664, 0.007455743789672851, 0.007194623947143554, 0.007138304233551026, 0.0071824960708618164, 0.007138144016265869, 0.007160831928253173, 0.007163904190063477, 0.007138304233551026, 0.007170048236846924, 0.007197696208953858, 0.007184415817260742, 0.007290847778320312, 0.007129087924957276, 0.007158783912658692, 0.007156735897064209, 0.007132160186767578, 0.007158783912658692, 0.00713424015045166, 0.007152607917785644, 0.007174143791198731, 0.015292415618896485, 0.007132160186767578, 0.007153728008270264, 0.007152575969696045, 0.0071792640686035155, 0.007151616096496582, 0.007087103843688965, 0.007165952205657959, 0.007159808158874512, 0.007155712127685547, 0.0071536960601806645, 0.007155680179595948, 0.007168000221252442, 0.007170048236846924, 0.007205887794494629, 0.007153664112091064, 0.007154687881469727, 0.007175168037414551, 0.007157760143280029, 0.007169023990631103, 0.007145472049713135, 0.007163904190063477, 0.007173120021820068, 0.007123968124389648, 0.007144447803497315, 0.0072130560874938965, 0.007150591850280762, 0.007172095775604248, 0.007145631790161133, 0.007125855922698975, 0.007130112171173096, 0.007162879943847656, 0.007170048236846924, 0.007142399787902832, 0.0071157760620117185, 0.007163904190063477, 0.007231488227844239, 0.007165952205657959, 0.007577600002288819, 0.007476223945617676, 0.00742195177078247, 0.007419904232025146, 0.0074926080703735356, 0.007291903972625732, 0.007142399787902832, 0.00714035177230835, 0.007131135940551757, 0.00714137601852417, 0.007147520065307617, 0.007158783912658692, 0.007131135940551757, 0.007101439952850342, 0.007150591850280762, 0.00714035177230835, 0.007139328002929687, 0.007109632015228271, 0.007126016139984131, 0.007114751815795899, 0.00713318395614624, 0.007142399787902832, 0.007117824077606201, 0.007142399787902832, 0.007143424034118652, 0.015757311820983888, 0.007415808200836181, 0.007457791805267334, 0.007425024032592774, 0.0075008001327514645, 0.007174272060394287, 0.007165823936462402, 0.007198719978332519, 0.007137279987335205, 0.007896063804626464, 0.007502848148345947, 0.007459839820861816, 0.007450623989105225, 0.007473152160644531, 0.0074997758865356446, 0.007553023815155029, 0.007523327827453613, 0.007459839820861816, 0.00748748779296875, 0.007477248191833496, 0.007442431926727295, 0.00748748779296875, 0.00748854398727417, 0.007440351963043213, 0.0074035201072692874, 0.007501823902130127, 0.00744758415222168, 0.0075294399261474606, 0.00742195177078247, 0.0074997758865356446, 0.007485439777374267, 0.007481344223022461, 0.007442431926727295, 0.007458816051483155, 0.007194623947143554, 0.007308288097381592, 0.007109632015228271, 0.0071495680809020995, 0.007144447803497315, 0.007774208068847656, 0.007444479942321777, 0.007391232013702393, 0.007383039951324463, 0.007365632057189942, 0.007394303798675537, 0.007357439994812012, 0.00733081579208374, 0.007365632057189942, 0.007350272178649903, 0.007386112213134765, 0.007304192066192627, 0.007332863807678222, 0.007362559795379638, 0.007435264110565186, 0.0074629120826721195, 0.00745472002029419, 0.007425024032592774, 0.007391232013702393, 0.007158815860748291, 0.007548895835876465, 0.007488512039184571, 0.0074291200637817386, 0.007465983867645264, 0.015963135719299318, 0.007461887836456299, 0.007464960098266602, 0.007448575973510742, 0.007435264110565186, 0.007584767818450928, 0.00743833589553833, 0.007450623989105225, 0.007451648235321045, 0.007464960098266602, 0.0074301438331604, 0.007464960098266602, 0.007457791805267334, 0.00744755220413208, 0.007658495903015137, 0.007955455780029297, 0.0075642881393432615, 0.007437312126159668, 0.007422976016998291, 0.007357439994812012, 0.007432191848754883, 0.00743833589553833, 0.007460864067077637, 0.0074711041450500485, 0.00731660795211792, 0.0073521919250488284, 0.007426047801971435, 0.007357439994812012, 0.007337984085083008, 0.007346208095550537, 0.007343071937561035, 0.007285759925842285, 0.007301119804382325, 0.007078911781311035, 0.007054336071014404, 0.0070594558715820314, 0.007054336071014404, 0.007147520065307617, 0.0071495680809020995, 0.007150591850280762, 0.007122943878173828, 0.0071485438346862796, 0.007145472049713135, 0.007145472049713135, 0.007114751815795899, 0.007096320152282715, 0.007150591850280762, 0.007178239822387696, 0.007323647975921631, 0.007162879943847656, 0.007145472049713135, 0.007159808158874512, 0.007134208202362061, 0.007159808158874512, 0.00713321590423584, 0.007135200023651123, 0.007132160186767578, 0.007245823860168457, 0.007180287837982178, 0.00714137601852417, 0.007163904190063477, 0.007186431884765625, 0.007170048236846924, 0.015305727958679198, 0.007160831928253173, 0.0071823358535766605, 0.007180287837982178, 0.007174143791198731, 0.007189504146575928, 0.00713318395614624, 0.007150591850280762, 0.007160831928253173, 0.007157760143280029, 0.007139328002929687, 0.007172095775604248, 0.007166975975036621, 0.007208960056304932, 0.007238656044006348, 0.007162879943847656, 0.007128064155578613, 0.007115839958190918, 0.007164864063262939, 0.007171072006225586, 0.007136256217956543, 0.007134208202362061, 0.0071157760620117185, 0.007174143791198731, 0.0071905279159545895, 0.007157760143280029, 0.007175168037414551, 0.007176191806793213, 0.00719155216217041, 0.0071495680809020995, 0.00714035177230835, 0.0071526398658752445, 0.007159808158874512, 0.0071792640686035155, 0.007160831928253173, 0.007178239822387696, 0.007161856174468994, 0.007143424034118652, 0.007122943878173828, 0.0073134078979492185, 0.007147520065307617, 0.0071823358535766605, 0.007168000221252442, 0.007184383869171143, 0.007151616096496582, 0.007226367950439453, 0.007192575931549072, 0.007194623947143554, 0.007497727870941162, 0.007504896163940429, 0.007473152160644531, 0.007425024032592774, 0.007143424034118652, 0.007155712127685547, 0.007176191806793213, 0.007165952205657959, 0.007171072006225586, 0.007131135940551757, 0.007169023990631103, 0.00722431993484497, 0.007181312084197998, 0.007128064155578613, 0.007171072006225586, 0.015277055740356446, 0.007169023990631103, 0.007165952205657959, 0.007116799831390381, 0.007164927959442138, 0.0071485438346862796, 0.007178239822387696, 0.007138304233551026, 0.007139328002929687, 0.007172095775604248, 0.007160831928253173, 0.007143424034118652, 0.007135231971740722, 0.00719974422454834, 0.007161856174468994, 0.007187456130981445, 0.007181312084197998, 0.007164927959442138, 0.007169023990631103, 0.007176191806793213, 0.007170048236846924, 0.007169023990631103, 0.007120895862579346, 0.0071495680809020995, 0.007203904151916504, 0.00738809585571289, 0.008572928428649903, 0.007788544178009033, 0.0075345921516418455, 0.007489535808563232, 0.007436287879943848, 0.007522304058074952, 0.007449600219726562, 0.007518208026885987, 0.007465983867645264, 0.007459839820861816, 0.007473152160644531, 0.0074629120826721195, 0.007443456172943115, 0.007185408115386963, 0.007147520065307617, 0.0071526398658752445, 0.007334911823272705, 0.007408639907836914, 0.007482367992401123, 0.0074414081573486324, 0.00745472002029419, 0.007488512039184571, 0.0076687359809875484, 0.0074711041450500485, 0.00745472002029419, 0.007392255783081054, 0.007426047801971435, 0.0074670081138610836, 0.007459839820861816, 0.007480319976806641, 0.007521279811859131, 0.0074332160949707035, 0.007400447845458984, 0.007457791805267334, 0.007477248191833496, 0.0074700798988342285, 0.007449600219726562, 0.015819775581359864, 0.007419904232025146, 0.007414783954620361, 0.007416831970214844, 0.007422976016998291, 0.007672832012176513, 0.007556096076965332, 0.00748748779296875, 0.007484416007995606, 0.0075008001327514645, 0.007456768035888672, 0.007418879985809326, 0.007438399791717529, 0.0074208641052246095, 0.0074414401054382325, 0.0074700479507446285, 0.007481344223022461, 0.007428095817565918, 0.0074670081138610836, 0.0074035201072692874, 0.007323647975921631, 0.007304192066192627, 0.007352320194244385, 0.007461887836456299, 0.007426047801971435, 0.007371776103973389, 0.007154687881469727, 0.007442431926727295, 0.007464960098266602, 0.007502848148345947, 0.0074414081573486324, 0.007437312126159668, 0.007444479942321777, 0.007463935852050781, 0.007464960098266602, 0.00749567985534668, 0.0074700798988342285, 0.007453695774078369, 0.007387135982513428, 0.007437312126159668, 0.0074332160949707035, 0.007390207767486572, 0.0074291200637817386, 0.007666687965393066, 0.007510047912597656, 0.007450592041015625, 0.007508992195129394, 0.007475200176239013, 0.00758073616027832, 0.007771071910858155, 0.007611392021179199, 0.007663616180419922, 0.007512063980102539, 0.007532544136047363, 0.0074700798988342285, 0.007486464023590088, 0.007480319976806641, 0.00748748779296875, 0.007456768035888672, 0.007506944179534912, 0.0074752321243286134, 0.007463903903961182, 0.007483391761779785, 0.01617100715637207, 0.0074403839111328125, 0.007444479942321777, 0.007416831970214844, 0.007428095817565918, 0.00744652795791626, 0.007422976016998291, 0.007456768035888672, 0.007476223945617676, 0.007484416007995606, 0.00745472002029419, 0.0074629120826721195, 0.007478271961212158, 0.007623680114746094, 0.007460864067077637, 0.007472127914428711, 0.00744755220413208, 0.007405600070953369, 0.007157728195190429, 0.007187456130981445, 0.007204864025115967, 0.007163904190063477, 0.007168000221252442, 0.007146527767181396, 0.007193568229675293, 0.007188479900360107, 0.007455743789672851, 0.007558144092559814, 0.007425024032592774, 0.007480319976806641, 0.007469056129455566, 0.007425024032592774, 0.0074414081573486324, 0.007448575973510742, 0.0074291200637817386, 0.0074403839111328125, 0.0074291200637817386, 0.007202847957611084, 0.007511007785797119, 0.007506944179534912, 0.007482367992401123, 0.007444479942321777, 0.007400447845458984, 0.007426047801971435, 0.007452672004699707, 0.007456768035888672, 0.007494656085968018, 0.007442463874816894, 0.007464928150177002, 0.007525375843048096, 0.007437344074249268, 0.007433184146881103, 0.007461887836456299, 0.007436287879943848, 0.007527423858642578, 0.007434239864349365, 0.007422976016998291, 0.007425024032592774, 0.007420959949493408, 0.007427040100097656, 0.007415808200836181, 0.007476223945617676, 0.007411712169647216, 0.01586892795562744, 0.007416831970214844, 0.00748748779296875, 0.0074700798988342285, 0.007464960098266602, 0.007346176147460938, 0.00738918399810791, 0.007398399829864502, 0.007353343963623047, 0.007370751857757568, 0.007355391979217529, 0.007390207767486572, 0.0073820161819458, 0.007350272178649903, 0.00733081579208374, 0.007420928001403809, 0.007471231937408447, 0.0073818879127502445, 0.007346176147460938, 0.007392255783081054, 0.007375872135162354, 0.007350272178649903, 0.007380032062530518, 0.007322559833526612, 0.007345151901245117, 0.0074106879234313965, 0.007357439994812012, 0.007398399829864502, 0.007374847888946533, 0.007407616138458252, 0.00758784008026123, 0.007337984085083008, 0.007380095958709717, 0.007363455772399902, 0.007366655826568603, 0.007354368209838867, 0.007355391979217529, 0.007331840038299561, 0.007359519958496093, 0.007412703990936279, 0.0074629120826721195, 0.007387135982513428, 0.007392416000366211, 0.007442368030548096, 0.007395232200622558, 0.0073820161819458, 0.007401472091674805, 0.007365632057189942, 0.0073175039291381834, 0.007455743789672851, 0.007364607810974121, 0.007398399829864502, 0.007432191848754883, 0.00733900785446167, 0.007359488010406494, 0.007341119766235351, 0.007389120101928711, 0.007318687915802002, 0.007402336120605469, 0.007372799873352051, 0.0073471999168395995, 0.0073431038856506346, 0.0074486079216003415, 0.0158822078704834, 0.0073471999168395995, 0.007395328044891358, 0.007385087966918945, 0.007372799873352051, 0.007350272178649903, 0.007370751857757568, 0.007379968166351319, 0.0073697280883789065, 0.007351295948028564, 0.007329792022705078, 0.007345151901245117, 0.007356416225433349, 0.007361536026000977, 0.007365632057189942, 0.0073820161819458, 0.0074700798988342285, 0.007355391979217529, 0.0073431038856506346, 0.007388160228729248, 0.007379968166351319, 0.007299232006072998, 0.007335775852203369, 0.00729804801940918, 0.007366655826568603, 0.007385087966918945, 0.007357439994812012, 0.007310336112976074, 0.007357439994812012, 0.007379968166351319, 0.007463935852050781, 0.007383039951324463, 0.007388160228729248, 0.007372799873352051, 0.007386112213134765, 0.00742195177078247, 0.007442431926727295, 0.007436351776123047, 0.007443391799926758, 0.007464960098266602, 0.00748748779296875, 0.007473152160644531, 0.007477248191833496, 0.00758784008026123, 0.00744755220413208, 0.007452672004699707, 0.007445504188537597, 0.007486464023590088, 0.007479296207427978, 0.007455743789672851, 0.007428095817565918, 0.0074403839111328125, 0.00748748779296875, 0.0074711041450500485, 0.007472127914428711, 0.007459839820861816, 0.007452672004699707, 0.0074414081573486324, 0.007414783954620361, 0.007413760185241699, 0.007484416007995606, 0.007435264110565186, 0.007479296207427978, 0.015879167556762695, 0.00748748779296875, 0.007461887836456299, 0.007425024032592774, 0.007428095817565918, 0.0075304961204528805, 0.007524352073669433, 0.007517183780670166, 0.007689216136932373, 0.007659520149230957, 0.007589888095855713, 0.007494656085968018, 0.007443456172943115, 0.007489535808563232, 0.0075038719177246095, 0.007479296207427978, 0.007532544136047363, 0.0075038719177246095, 0.007488512039184571, 0.007480319976806641, 0.007444479942321777, 0.007169023990631103, 0.0071792640686035155, 0.007176191806793213, 0.007173120021820068, 0.007144447803497315, 0.007110688209533691, 0.0071751360893249514, 0.007161856174468994, 0.007139328002929687, 0.007170048236846924, 0.007129119873046875, 0.007152607917785644, 0.007193600177764893, 0.007146495819091797, 0.007173120021820068, 0.007159808158874512, 0.007193600177764893, 0.007338016033172607, 0.007167967796325683, 0.007160831928253173, 0.007155712127685547, 0.007192575931549072, 0.007143424034118652, 0.007157760143280029, 0.007160831928253173, 0.0071792640686035155, 0.007235583782196045, 0.0071792640686035155, 0.0071833600997924804, 0.007216127872467041, 0.0071823358535766605, 0.00719974422454834, 0.007173120021820068, 0.007165952205657959, 0.007163904190063477, 0.007159808158874512, 0.007218175888061523, 0.007175168037414551, 0.00719974422454834, 0.007155712127685547, 0.007164927959442138, 0.007154687881469727, 0.015350784301757812, 0.007101439952850342, 0.007189504146575928, 0.007165952205657959, 0.0071905279159545895, 0.007163904190063477, 0.007127039909362793, 0.007174272060394287, 0.007178112030029297, 0.007170048236846924, 0.007054336071014404, 0.007154816150665283, 0.007121791839599609, 0.007259136199951172, 0.007485439777374267, 0.007422976016998291, 0.007407616138458252, 0.00733900785446167, 0.007329792022705078, 0.007387135982513428, 0.007349311828613281, 0.007333824157714844, 0.007331840038299561, 0.0073400321006774905, 0.007370751857757568, 0.007365791797637939, 0.007350111961364746, 0.007352320194244385, 0.007363584041595459, 0.0074301438331604, 0.00744755220413208, 0.00739737606048584, 0.007443488121032715, 0.007573472023010254, 0.007488512039184571, 0.007443456172943115, 0.007412735939025879, 0.007451648235321045, 0.007460864067077637, 0.007485439777374267, 0.0074414081573486324, 0.007395328044891358, 0.007554048061370849, 0.00749567985534668, 0.007180287837982178, 0.007203839778900147, 0.00714137601852417, 0.0071485438346862796, 0.007137279987335205, 0.007158783912658692, 0.007156735897064209, 0.007229440212249756, 0.007186431884765625, 0.007121920108795166, 0.007150591850280762, 0.007251967906951904, 0.007158783912658692, 0.0071823358535766605, 0.00713318395614624, 0.007165984153747558, 0.007143392086029053, 0.007136288166046142, 0.007136256217956543, 0.015279071807861328, 0.007146495819091797, 0.007097343921661377, 0.0072540159225463864, 0.007161856174468994, 0.007134208202362061, 0.00706774377822876, 0.007062431812286377, 0.007058432102203369, 0.007113728046417236, 0.007071743965148926, 0.007108607769012451, 0.007122943878173828, 0.007067647933959961, 0.00710041618347168, 0.0070553598403930665, 0.00709222412109375, 0.007175168037414551, 0.007101439952850342, 0.007160895824432373, 0.00714847993850708, 0.007164927959442138, 0.007153664112091064, 0.007095295906066895, 0.007169023990631103, 0.007166975975036621, 0.007173120021820068, 0.007116799831390381, 0.007285759925842285, 0.00722431993484497, 0.007184383869171143, 0.007161856174468994, 0.00714035177230835, 0.007205887794494629, 0.007195648193359375, 0.0071823358535766605, 0.007176224231719971, 0.007157728195190429, 0.007175168037414551, 0.007195648193359375, 0.007163904190063477, 0.007176191806793213, 0.007154687881469727, 0.007206912040710449, 0.007164927959442138, 0.007180319786071777, 0.007135200023651123, 0.007160831928253173, 0.00719974422454834, 0.007170048236846924, 0.0072427520751953125, 0.007162879943847656, 0.007175168037414551, 0.007186431884765625, 0.007336959838867187, 0.0072325758934020995, 0.007157695770263672, 0.00719155216217041, 0.007177216053009033, 0.007215104103088379, 0.007207935810089112, 0.007134208202362061, 0.007181312084197998]",tokens/s,133.70880740243032,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Salesforce/codegen-6B-nl,Salesforce/codegen-6B-nl,cuda,0,42,,,True,,,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",codegen,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-410m,,cuda,0,42,,,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,217063f5c507ed7cc255df7e1f64c4333a0b4dfe,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,,MB,1300.873216,1713.897472,0.0,1067.450368,1018.226688,s,10,0.16997203254699708,0.016997203254699708,0.00041507905486039146,0.016764400482177734,0.017378297042846682,0.017702124214172364,0.017961185951232913,"[0.01802595138549805, 0.01674083137512207, 0.016743040084838866, 0.017091136932373047, 0.01730633544921875, 0.016676704406738282, 0.017260576248168947, 0.016785760879516602, 0.01666521644592285, 0.01667647933959961]",tokens/s,15061.30133080666,kWh,2.0177843848050524e-07,1.1056505171908154e-07,5.087115370409696e-07,8.210550272405565e-07,tokens/kWh,311793962.0446364,MB,1300.873216,1713.897472,0.0,1067.450368,1059.39456,s,10,10.173291931152344,1.0173291931152344,0.011496006145209937,1.0139341125488281,1.0373599243164062,1.0384082946777344,1.0392469909667967,"[1.0083255615234374, 1.0059702758789062, 1.0048641357421875, 1.0131695556640625, 1.037126953125, 1.0214266357421875, 1.0394566650390624, 1.0166776733398437, 1.0146986694335938, 1.0115758056640625]",tokens/s,61.9268575268968,kWh,1.2068471352134998e-05,6.613018416964417e-06,2.1189972488558242e-05,3.987146225765766e-05,tokens/kWh,1580077.489831723,,s,629,10.308060156822195,0.016388012967920834,0.0020900941940995166,0.015986687660217287,0.01648599052429199,0.01668833236694336,0.03302572143554688,"[0.01744486427307129, 0.016336896896362304, 0.015946751594543458, 0.015913984298706055, 0.016058368682861326, 0.016199680328369142, 0.016114688873291014, 0.015892479896545412, 0.015848447799682617, 0.015821824073791502, 0.01581984043121338, 0.015839167594909668, 0.016021503448486327, 0.016135168075561524, 0.016210943222045898, 0.01589145565032959, 0.0158341121673584, 0.015820799827575685, 0.015887359619140624, 0.015841376304626464, 0.015886240005493164, 0.015796223640441896, 0.015760448455810545, 0.015928256034851075, 0.01587718391418457, 0.015871935844421387, 0.015812607765197755, 0.015836159706115724, 0.015807488441467284, 0.015812607765197755, 0.015892479896545412, 0.01581670379638672, 0.015814656257629393, 0.015789055824279787, 0.01578700828552246, 0.016380928039550782, 0.016539648056030275, 0.01658367919921875, 0.017051647186279297, 0.016524288177490236, 0.01601535987854004, 0.01586796760559082, 0.015919103622436523, 0.015910847663879395, 0.01580134391784668, 0.015847423553466796, 0.015970303535461427, 0.015848511695861815, 0.015943615913391114, 0.015871999740600586, 0.01589145565032959, 0.01584537601470947, 0.015911935806274414, 0.015930368423461915, 0.015904767990112305, 0.015805503845214845, 0.015861696243286132, 0.01581670379638672, 0.015857664108276368, 0.01582592010498047, 0.01593446445465088, 0.015844351768493654, 0.03305472183227539, 0.01598361587524414, 0.015930368423461915, 0.015981568336486816, 0.015898624420166017, 0.01592627239227295, 0.015904767990112305, 0.0158853120803833, 0.015904767990112305, 0.01590681552886963, 0.01589350414276123, 0.01586892795562744, 0.015882240295410157, 0.015831040382385253, 0.015969280242919923, 0.015846400260925292, 0.015888383865356445, 0.01587712001800537, 0.015851519584655763, 0.015840255737304687, 0.015890432357788087, 0.01590169620513916, 0.015863807678222656, 0.015861824035644533, 0.015911871910095216, 0.01587097644805908, 0.015922176361083985, 0.015895551681518554, 0.015884287834167482, 0.015824895858764648, 0.015873023986816406, 0.015949824333190917, 0.015952896118164063, 0.015876095771789552, 0.015887359619140624, 0.016089088439941408, 0.01638297653198242, 0.01621811294555664, 0.015873023986816406, 0.015981568336486816, 0.015871999740600586, 0.016227327346801757, 0.016313343048095702, 0.016475135803222657, 0.015997952461242675, 0.015935487747192383, 0.016076799392700195, 0.01590681552886963, 0.015913984298706055, 0.015866880416870118, 0.015882240295410157, 0.015873023986816406, 0.01587814426422119, 0.01590681552886963, 0.015932415962219237, 0.015930368423461915, 0.0159334716796875, 0.016533472061157226, 0.01647315216064453, 0.016004032135009765, 0.01597542381286621, 0.01590783977508545, 0.015903743743896484, 0.033363967895507815, 0.0163768310546875, 0.015852543830871584, 0.015987711906433104, 0.01642803192138672, 0.016021503448486327, 0.01583513641357422, 0.015953920364379884, 0.015805439949035646, 0.015904767990112305, 0.015859711647033693, 0.015904767990112305, 0.016062463760375977, 0.01621708869934082, 0.0158341121673584, 0.015898624420166017, 0.01588326358795166, 0.01589350414276123, 0.015833087921142578, 0.015866880416870118, 0.015824895858764648, 0.015881216049194336, 0.01590169620513916, 0.015905792236328126, 0.015905792236328126, 0.015827967643737794, 0.015866880416870118, 0.01589964771270752, 0.015977472305297852, 0.015865856170654297, 0.01586278438568115, 0.015864831924438477, 0.015900671958923338, 0.01590681552886963, 0.01618124771118164, 0.01636966323852539, 0.016339967727661133, 0.015910911560058593, 0.015905792236328126, 0.015904767990112305, 0.015900671958923338, 0.015924223899841307, 0.01594777584075928, 0.01594268798828125, 0.016163808822631836, 0.016119808197021485, 0.015917056083679198, 0.015946751594543458, 0.01590176010131836, 0.015846336364746094, 0.015889408111572266, 0.015865856170654297, 0.015879167556762695, 0.015838208198547362, 0.01585868835449219, 0.015836159706115724, 0.015851519584655763, 0.015856639862060547, 0.01586995220184326, 0.015879167556762695, 0.015859711647033693, 0.01586995220184326, 0.01588633632659912, 0.03381043243408203, 0.015919103622436523, 0.015846400260925292, 0.015859711647033693, 0.015852543830871584, 0.015849472045898438, 0.01581875228881836, 0.01585868835449219, 0.015823871612548827, 0.015781951904296876, 0.015829952239990234, 0.015831040382385253, 0.01594777584075928, 0.01582592010498047, 0.015856639862060547, 0.01582694435119629, 0.01582592010498047, 0.01581158447265625, 0.0158341121673584, 0.0158023681640625, 0.015873023986816406, 0.015848447799682617, 0.015857664108276368, 0.015849504470825195, 0.01609622383117676, 0.01661747169494629, 0.01682636833190918, 0.0162857608795166, 0.01607468795776367, 0.016063488006591797, 0.016004095077514647, 0.015944704055786133, 0.016004095077514647, 0.01597644805908203, 0.015856639862060547, 0.015807488441467284, 0.015742976188659667, 0.01567743968963623, 0.015900671958923338, 0.01599078369140625, 0.015953920364379884, 0.015949824333190917, 0.015921152114868165, 0.016261119842529297, 0.016463872909545898, 0.016734207153320312, 0.016343040466308592, 0.016358400344848634, 0.016352256774902343, 0.01640959930419922, 0.01637478446960449, 0.01639116859436035, 0.01647724723815918, 0.016352191925048828, 0.01637785530090332, 0.01636147117614746, 0.016453632354736326, 0.016390144348144533, 0.016257024765014647, 0.015841279983520508, 0.015847423553466796, 0.01684377670288086, 0.016446464538574217, 0.034802688598632815, 0.01747148895263672, 0.017386495590209963, 0.016472063064575194, 0.016440319061279296, 0.016278528213500978, 0.016363519668579102, 0.016508928298950197, 0.016909311294555664, 0.016881664276123046, 0.016772096633911132, 0.016570367813110352, 0.01643110466003418, 0.01637171173095703, 0.01660927963256836, 0.01641881561279297, 0.01638400077819824, 0.016381952285766603, 0.01640345573425293, 0.016406591415405274, 0.016439231872558593, 0.016451583862304688, 0.016689151763916017, 0.016525312423706053, 0.016455680847167968, 0.016348159790039063, 0.01639423942565918, 0.016338943481445312, 0.01637068748474121, 0.016360448837280273, 0.016442367553710938, 0.01640959930419922, 0.01640243148803711, 0.01638400077819824, 0.01640140724182129, 0.01641574478149414, 0.016481279373168945, 0.016356351852416993, 0.016336896896362304, 0.016670719146728515, 0.01656524848937988, 0.016373760223388673, 0.016347135543823242, 0.01637785530090332, 0.0164136962890625, 0.016408575057983397, 0.016492544174194337, 0.016446464538574217, 0.016372735977172852, 0.016473087310791015, 0.0163768310546875, 0.016366592407226564, 0.016350208282470705, 0.01637068748474121, 0.016326656341552736, 0.016408575057983397, 0.016105472564697267, 0.015841279983520508, 0.01703424072265625, 0.0163768310546875, 0.01599897575378418, 0.015867903709411622, 0.01584332847595215, 0.033031169891357424, 0.015859711647033693, 0.015847423553466796, 0.01587712001800537, 0.015896575927734375, 0.016067583084106444, 0.016114688873291014, 0.016885759353637696, 0.017666048049926757, 0.016687103271484375, 0.01639423942565918, 0.01591808032989502, 0.015927295684814453, 0.015881216049194336, 0.01590988826751709, 0.015823871612548827, 0.015864831924438477, 0.01584537601470947, 0.015863807678222656, 0.01581158447265625, 0.01639833641052246, 0.016430080413818358, 0.01615667152404785, 0.01642803192138672, 0.01642905616760254, 0.016336896896362304, 0.01636966323852539, 0.016368640899658202, 0.01636966323852539, 0.01637887954711914, 0.016314367294311523, 0.016397312164306642, 0.01638604736328125, 0.016434175491333008, 0.01640550422668457, 0.01641472053527832, 0.01637478446960449, 0.01660927963256836, 0.01644339179992676, 0.0164003849029541, 0.01640447998046875, 0.01641574478149414, 0.0158853120803833, 0.01583513641357422, 0.01588326358795166, 0.01599180793762207, 0.015815679550170898, 0.015814656257629393, 0.01581056022644043, 0.015908864021301268, 0.01590783977508545, 0.015796223640441896, 0.015884287834167482, 0.015856639862060547, 0.0158535680770874, 0.015968255996704102, 0.017472511291503907, 0.016491519927978517, 0.01640959930419922, 0.01640755271911621, 0.01641983985900879, 0.016455680847167968, 0.01642291259765625, 0.03405516815185547, 0.016364543914794923, 0.016424959182739257, 0.016668672561645507, 0.016679935455322266, 0.016359424591064452, 0.016392223358154295, 0.016350175857543944, 0.016372735977172852, 0.01659699249267578, 0.016424959182739257, 0.017356800079345702, 0.01657651138305664, 0.01641983985900879, 0.01640959930419922, 0.016387136459350585, 0.01634911918640137, 0.0164136962890625, 0.016464895248413085, 0.01636761665344238, 0.016377920150756835, 0.016424896240234375, 0.01653555107116699, 0.01640959930419922, 0.016472095489501952, 0.016417760848999024, 0.01639833641052246, 0.016373760223388673, 0.01637478446960449, 0.016448511123657226, 0.016454687118530275, 0.016370655059814453, 0.01645260810852051, 0.016501760482788085, 0.016488447189331054, 0.01638502311706543, 0.016447488784790038, 0.016438272476196288, 0.016342016220092775, 0.016408575057983397, 0.016338943481445312, 0.01641881561279297, 0.01643622398376465, 0.01643212890625, 0.016480255126953124, 0.01642291259765625, 0.01641062355041504, 0.016442367553710938, 0.01640652847290039, 0.016451583862304688, 0.016447488784790038, 0.016373760223388673, 0.016373760223388673, 0.017163263320922852, 0.016485376358032225, 0.016517120361328123, 0.01640550422668457, 0.016358400344848634, 0.01760972785949707, 0.01721139144897461, 0.016556032180786134, 0.016546815872192384, 0.016497663497924805, 0.033555454254150394, 0.016479232788085937, 0.0164136962890625, 0.016376895904541014, 0.016451520919799803, 0.016368640899658202, 0.01642393684387207, 0.016442367553710938, 0.0163656005859375, 0.016312288284301757, 0.016305152893066405, 0.01640345573425293, 0.01640243148803711, 0.01643929672241211, 0.01641062355041504, 0.01639833641052246, 0.01643622398376465, 0.01641164779663086, 0.01640345573425293, 0.016373760223388673, 0.01638502311706543, 0.01638707160949707, 0.01643622398376465, 0.01623859214782715, 0.015879167556762695, 0.015871999740600586, 0.01587814426422119, 0.015910911560058593, 0.01587814426422119, 0.015871999740600586, 0.01581670379638672, 0.01624678421020508, 0.016442367553710938, 0.016408575057983397, 0.015884287834167482, 0.015859711647033693, 0.015874048233032227, 0.015831040382385253, 0.015827967643737794, 0.01583619213104248, 0.015856608390808106, 0.015917087554931642, 0.015835103988647462, 0.015765503883361818, 0.015857664108276368, 0.015846400260925292, 0.015857664108276368, 0.01586899185180664, 0.01583199977874756, 0.015874048233032227, 0.016705535888671876, 0.016446464538574217, 0.01639321517944336, 0.01639219284057617, 0.016257024765014647, 0.01580339241027832, 0.015876095771789552, 0.015844351768493654, 0.01581772804260254, 0.01578598403930664, 0.015866880416870118, 0.015904767990112305, 0.01596723175048828, 0.03291648101806641, 0.015790143966674806, 0.015886272430419922, 0.01589452838897705, 0.015684608459472657, 0.015681568145751952, 0.015838175773620605, 0.01581670379638672, 0.01580851173400879, 0.015831040382385253, 0.015847423553466796, 0.015782912254333495, 0.015822848320007323, 0.015846400260925292, 0.01586892795562744, 0.01586176013946533, 0.015824895858764648, 0.015857664108276368, 0.01591808032989502, 0.015889408111572266, 0.015807488441467284, 0.015851519584655763, 0.015840288162231445, 0.01582588768005371, 0.016066560745239256, 0.016304128646850585, 0.0165980167388916, 0.01637171173095703, 0.016334848403930666, 0.016306175231933593, 0.01659187126159668, 0.016390144348144533, 0.016359487533569337, 0.016424896240234375, 0.016331775665283203, 0.016330751419067382, 0.01640345573425293, 0.016343040466308592, 0.016322559356689453, 0.016335872650146483, 0.016306175231933593, 0.016330751419067382, 0.016333856582641602, 0.01640547180175781, 0.01639936065673828, 0.016360448837280273, 0.01588633632659912, 0.01585459232330322, 0.015822848320007323, 0.017108991622924806, 0.01638707160949707, 0.016350208282470705, 0.0164003849029541, 0.016365568161010743, 0.016330751419067382, 0.015881216049194336, 0.01591603183746338, 0.015773695945739748, 0.015775744438171386, 0.01582694435119629, 0.015986687660217287, 0.016442367553710938, 0.016323583602905273, 0.03301171112060547, 0.015797247886657716, 0.015806464195251464, 0.015830016136169432, 0.015855615615844726, 0.015827967643737794, 0.015945728302001954, 0.01584537601470947, 0.015781888008117674, 0.015758336067199705, 0.01580953598022461, 0.016357376098632814, 0.01631545639038086, 0.01631430435180664, 0.016352256774902343, 0.016355327606201172, 0.01638502311706543, 0.016334848403930666, 0.016349184036254884, 0.016318464279174806, 0.016357376098632814, 0.015880191802978515, 0.015863807678222656, 0.01584332847595215, 0.01582592010498047, 0.015839232444763183, 0.01581670379638672, 0.015863807678222656, 0.015779840469360353, 0.015789055824279787, 0.01576140785217285, 0.015783935546875, 0.015783935546875, 0.015852543830871584, 0.015755264282226563, 0.016309247970581055, 0.016299007415771484, 0.0158023681640625, 0.015807488441467284, 0.015817791938781738, 0.01579206371307373, 0.015832063674926757, 0.01579520034790039, 0.015775744438171386, 0.01588633632659912, 0.015814656257629393, 0.01580851173400879, 0.016760831832885743, 0.016493568420410155, 0.01589145565032959, 0.0158341121673584, 0.017081344604492187, 0.016508928298950197, 0.016363519668579102, 0.016300031661987305, 0.016297983169555663, 0.016391231536865235, 0.01630611228942871, 0.016351264953613283, 0.01651606369018555, 0.016479232788085937, 0.01593958377838135, 0.01586892795562744]",tokens/s,61.02021044024541,,,main,False,False,,,,,,,,,,,,,,,,,,,,,,, float32-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,.,.,cuda,0,42,,,True,True,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 373, in cached_file raise EnvironmentError( OSError: . does not appear to have a file named config.json. Checkout 'https://huggingface.co/./tree/None' for available files. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,a,a,cuda,0,42,,,True,True,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/a/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66947470-4fb3414119f7c5871a2b7787;20e4f0a6-29a7-4b33-9bee-9b2268573547) Repository Not Found for url: https://huggingface.co/a/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: a is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-7B,Qwen/Qwen-7B,cuda,0,42,,,True,,,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",qwen,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-66b,facebook/opt-66b,cuda,0,42,,,True,,,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",opt,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,stabilityai/stablelm-base-alpha-7b,stabilityai/stablelm-base-alpha-7b,cuda,0,42,,,True,True,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ self.load_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 250, in load_model_with_no_weights self.pretrained_model = self.automodel_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3710, in from_pretrained model = cls(config, *model_args, **model_kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 1028, in __init__ self.gpt_neox = GPTNeoXModel(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 853, in __init__ self.layers = nn.ModuleList([GPTNeoXLayer(config) for _ in range(config.num_hidden_layers)]) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 853, in self.layers = nn.ModuleList([GPTNeoXLayer(config) for _ in range(config.num_hidden_layers)]) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 740, in __init__ self.attention = GPT_NEOX_ATTENTION_CLASSES[config._attn_implementation](config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 118, in __init__ self.query_key_value = nn.Linear(config.hidden_size, 3 * config.hidden_size, bias=config.attention_bias) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/linear.py"", line 98, in __init__ self.weight = Parameter(torch.empty((out_features, in_features), **factory_kwargs)) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_device.py"", line 78, in __torch_function__ return func(*args, **kwargs) torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 432.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/gpt-neo-2.7B,EleutherAI/gpt-neo-2.7B,cuda,0,42,,,True,True,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1084.882944,12046.565376,0.0,11406.409728,11107.918848,s,10,0.9638016738891602,0.09638016738891603,0.000964740454284909,0.09606313705444336,0.09666563415527343,0.09794406509399413,0.0989668098449707,"[0.09922249603271484, 0.09618201446533203, 0.09638153839111328, 0.09596109008789062, 0.09600742340087891, 0.09578521728515625, 0.09588813018798828, 0.09592435455322265, 0.09611885070800781, 0.09633055877685547]",tokens/s,2656.1481156904556,kWh,1.1512115324456588e-06,6.305885890416674e-07,8.048873753009786e-06,9.830673874497111e-06,tokens/kWh,26040941.167229567,MB,1084.882944,12046.565376,0.0,11406.409728,11304.03328,s,10,18.439808593749998,1.8439808593749998,0.005488303031145108,1.8441043701171875,1.8510441040039063,1.8513414489746094,1.8515793249511718,"[1.8488201904296875, 1.8379088134765624, 1.8442047119140625, 1.8442847900390624, 1.8325855712890624, 1.8440040283203125, 1.8426549072265626, 1.8516387939453125, 1.85097802734375, 1.842728759765625]",tokens/s,34.165213635326594,kWh,2.1696660993007907e-05,1.1890458622952635e-05,8.50219507035883e-05,0.00011860907031954884,tokens/kWh,531156.6799256541,,s,629,19.274700796127348,0.030643403491458377,0.011496102641603485,0.029314048767089845,0.029487110900878907,0.029598540496826174,0.1257538327026367,"[0.029079551696777343, 0.028883968353271484, 0.028891199111938475, 0.02885830307006836, 0.028874752044677734, 0.028882944107055664, 0.028914688110351562, 0.028851200103759765, 0.028919807434082033, 0.028880895614624022, 0.028875776290893555, 0.028858367919921874, 0.029283327102661134, 0.029282304763793947, 0.02976358413696289, 0.029931552886962892, 0.029721567153930664, 0.029702144622802733, 0.029672447204589843, 0.029425695419311525, 0.029335519790649415, 0.029336576461791993, 0.029343744277954102, 0.029428735733032226, 0.02936832046508789, 0.0294748477935791, 0.029379552841186523, 0.029355007171630858, 0.02937548828125, 0.029434879302978514, 0.02934272003173828, 0.029344768524169923, 0.02936832046508789, 0.02940108871459961, 0.029352960586547853, 0.02934988784790039, 0.029429759979248047, 0.02938470458984375, 0.029472768783569334, 0.029543424606323244, 0.029380607604980468, 0.02937446403503418, 0.02941644859313965, 0.02938572883605957, 0.02937548828125, 0.029453311920166016, 0.029449216842651366, 0.02940108871459961, 0.029486080169677735, 0.029512704849243163, 0.029417472839355467, 0.029436927795410156, 0.029372415542602538, 0.029491199493408202, 0.02948204803466797, 0.02944812774658203, 0.02954751968383789, 0.029500415802001953, 0.029404159545898437, 0.0295229434967041, 0.029478912353515626, 0.029091840744018556, 0.12595814514160156, 0.02889625549316406, 0.028879871368408205, 0.028852224349975586, 0.028856319427490236, 0.028785663604736327, 0.028833824157714842, 0.02898224067687988, 0.02892185592651367, 0.028878847122192384, 0.028943359375, 0.028825599670410155, 0.028847103118896485, 0.028889087677001952, 0.028867584228515625, 0.028867584228515625, 0.028828672409057617, 0.028829696655273438, 0.028827648162841796, 0.029294591903686523, 0.029271039962768555, 0.029361152648925783, 0.02933145523071289, 0.02889625549316406, 0.028922880172729492, 0.02899558448791504, 0.029054975509643553, 0.02894438362121582, 0.029281280517578126, 0.02919628715515137, 0.02942464065551758, 0.029418495178222655, 0.029405183792114258, 0.02933145523071289, 0.029318143844604492, 0.02934681510925293, 0.02939289665222168, 0.02938265609741211, 0.029379583358764647, 0.02937446403503418, 0.02942361640930176, 0.029404159545898437, 0.0294072322845459, 0.029472768783569334, 0.029465599060058592, 0.029414400100708008, 0.02944819259643555, 0.029394943237304686, 0.029443071365356444, 0.02931711959838867, 0.029433855056762694, 0.029492223739624023, 0.029456384658813478, 0.02941542434692383, 0.029465599060058592, 0.029120512008666992, 0.02913689613342285, 0.029076480865478517, 0.029088768005371093, 0.029066240310668946, 0.029115392684936525, 0.029090816497802735, 0.02911846351623535, 0.12704153442382812, 0.02934681510925293, 0.029295616149902344, 0.02929151916503906, 0.029320192337036134, 0.029247488021850586, 0.028860416412353516, 0.029174783706665038, 0.029017087936401367, 0.029262847900390625, 0.02932633590698242, 0.02910310363769531, 0.029290496826171877, 0.029256704330444337, 0.029336576461791993, 0.02893926429748535, 0.029105152130126953, 0.02911948776245117, 0.028815359115600587, 0.028819456100463867, 0.028822559356689453, 0.029133792877197265, 0.02915839958190918, 0.029043712615966798, 0.029181951522827147, 0.02934681510925293, 0.02937651252746582, 0.02936524772644043, 0.0294021110534668, 0.029313087463378906, 0.029383615493774416, 0.02936627197265625, 0.02936524772644043, 0.029473791122436522, 0.0293621768951416, 0.029343744277954102, 0.02938572883605957, 0.029221887588500976, 0.029206527709960937, 0.0293703670501709, 0.02935807991027832, 0.029246463775634765, 0.02896895980834961, 0.029014015197753908, 0.02909187126159668, 0.029362144470214843, 0.029430784225463868, 0.029411327362060546, 0.029282304763793947, 0.02940928077697754, 0.0294021110534668, 0.029444095611572265, 0.029428735733032226, 0.029388799667358398, 0.02933452796936035, 0.029149183273315428, 0.029128704071044922, 0.029048831939697265, 0.02918809509277344, 0.029487136840820313, 0.029467615127563476, 0.029486080169677735, 0.029496320724487303, 0.1257011184692383, 0.028887039184570314, 0.029181951522827147, 0.029211648941040037, 0.028831743240356447, 0.028753919601440428, 0.02914406394958496, 0.029444095611572265, 0.029718528747558592, 0.029338623046875, 0.029286399841308593, 0.029248512268066407, 0.029312000274658204, 0.029257728576660157, 0.029258752822875978, 0.029261823654174804, 0.029293567657470702, 0.029282304763793947, 0.029276159286499022, 0.029298688888549803, 0.029295616149902344, 0.02938572883605957, 0.02931711959838867, 0.029387775421142577, 0.029283327102661134, 0.029289472579956056, 0.02935091209411621, 0.029314048767089845, 0.029303808212280274, 0.02929151916503906, 0.029487104415893556, 0.029293567657470702, 0.02937548828125, 0.029361152648925783, 0.029438976287841798, 0.02943283271789551, 0.02934988784790039, 0.02934272003173828, 0.029372415542602538, 0.029404159545898437, 0.029344768524169923, 0.029363264083862306, 0.029383615493774416, 0.02933247947692871, 0.02931711959838867, 0.029349920272827148, 0.029444063186645508, 0.029457408905029295, 0.029024255752563476, 0.028969024658203123, 0.029035455703735353, 0.02902835273742676, 0.029069311141967775, 0.02925779151916504, 0.029386720657348632, 0.02943280029296875, 0.029328384399414063, 0.029244415283203123, 0.029064191818237304, 0.029043712615966798, 0.029088768005371093, 0.02919424057006836, 0.02905292892456055, 0.12588441467285155, 0.02898841667175293, 0.028856319427490236, 0.028876800537109375, 0.02910310363769531, 0.029024255752563476, 0.02878054428100586, 0.028837888717651368, 0.028875776290893555, 0.028818431854248046, 0.028791807174682618, 0.028826623916625976, 0.02879897689819336, 0.028812288284301758, 0.028816383361816408, 0.028931072235107422, 0.028767232894897462, 0.028819456100463867, 0.029068288803100587, 0.029329408645629884, 0.029199359893798828, 0.028914688110351562, 0.02919219207763672, 0.029273088455200196, 0.02903654479980469, 0.02920140838623047, 0.029288448333740235, 0.02896895980834961, 0.028985343933105468, 0.029165567398071288, 0.028852224349975586, 0.028852224349975586, 0.028919807434082033, 0.02899046325683594, 0.02899558448791504, 0.02915020751953125, 0.029344768524169923, 0.029481983184814452, 0.029261823654174804, 0.029177888870239258, 0.02898224067687988, 0.029025279998779296, 0.029049856185913086, 0.029037567138671876, 0.029029375076293946, 0.029000703811645507, 0.029061119079589845, 0.029063167572021483, 0.02901203155517578, 0.02906515121459961, 0.029067264556884766, 0.02913382339477539, 0.02934681510925293, 0.029065216064453125, 0.029051904678344728, 0.029076480865478517, 0.02940620803833008, 0.02996326446533203, 0.02974208068847656, 0.029569023132324217, 0.029404159545898437, 0.029129728317260743, 0.029077503204345705, 0.12596121978759767, 0.028948480606079102, 0.028915712356567383, 0.02889625549316406, 0.029248512268066407, 0.02892083168029785, 0.02918502426147461, 0.029222944259643554, 0.02925667190551758, 0.029230079650878905, 0.029304864883422852, 0.02944099235534668, 0.028855295181274415, 0.028848127365112306, 0.02894540786743164, 0.02918707275390625, 0.029305856704711915, 0.029314048767089845, 0.02911129570007324, 0.02893619155883789, 0.028952608108520506, 0.02901398468017578, 0.028831743240356447, 0.02894233512878418, 0.029148160934448244, 0.029076480865478517, 0.029130815505981445, 0.02973689651489258, 0.02954444885253906, 0.02938572883605957, 0.029336576461791993, 0.029337600708007814, 0.02932633590698242, 0.029377536773681642, 0.029461503982543946, 0.029343744277954102, 0.029378559112548826, 0.029437952041625977, 0.029445119857788086, 0.02962124824523926, 0.029524991989135742, 0.029664255142211913, 0.029371391296386717, 0.02939904022216797, 0.029429759979248047, 0.02943180847167969, 0.029429759979248047, 0.029386751174926756, 0.029454336166381836, 0.029267967224121092, 0.029076480865478517, 0.029563936233520507, 0.0296396484375, 0.029298688888549803, 0.029009920120239258, 0.02898841667175293, 0.029050880432128907, 0.029265920639038087, 0.02943180847167969, 0.029485055923461914, 0.029466623306274413, 0.02929254341125488, 0.02903654479980469, 0.12572876739501954, 0.028867584228515625, 0.028835840225219726, 0.028841983795166014, 0.028870655059814454, 0.0287825927734375, 0.0287457275390625, 0.028821504592895508, 0.029042688369750977, 0.029222911834716796, 0.02911027145385742, 0.029076480865478517, 0.029228031158447267, 0.029274112701416017, 0.029228031158447267, 0.029288448333740235, 0.029108224868774416, 0.02932633590698242, 0.02918400001525879, 0.029064191818237304, 0.029079551696777343, 0.028880895614624022, 0.028793855667114256, 0.028925952911376954, 0.028831743240356447, 0.02889625549316406, 0.029101055145263673, 0.02900377655029297, 0.029082624435424805, 0.02896895980834961, 0.029148160934448244, 0.0291276798248291, 0.029254655838012695, 0.029303808212280274, 0.02934067153930664, 0.02935398483276367, 0.029245439529418944, 0.029436927795410156, 0.02954854393005371, 0.029874176025390626, 0.02980768013000488, 0.02954745674133301, 0.029404159545898437, 0.029607936859130858, 0.029435903549194335, 0.029248512268066407, 0.02935091209411621, 0.029460479736328125, 0.029360128402709962, 0.02949017524719238, 0.029411327362060546, 0.029437952041625977, 0.02939289665222168, 0.029463552474975587, 0.02957619285583496, 0.02954035186767578, 0.029460479736328125, 0.02954035186767578, 0.029461503982543946, 0.029570047378540038, 0.029455360412597657, 0.029375520706176758, 0.029318111419677734, 0.12576358032226562, 0.029336576461791993, 0.029240320205688477, 0.02923520088195801, 0.029327360153198243, 0.029254655838012695, 0.028858367919921874, 0.02935091209411621, 0.029281280517578126, 0.02933145523071289, 0.029989887237548828, 0.02960383987426758, 0.029296640396118165, 0.029290496826171877, 0.029313024520874024, 0.02925056076049805, 0.029369344711303712, 0.029240320205688477, 0.029257728576660157, 0.02930790328979492, 0.029285375595092773, 0.029281280517578126, 0.0293703670501709, 0.029294591903686523, 0.02934783935546875, 0.029306880950927733, 0.029363199234008788, 0.02935910415649414, 0.029434879302978514, 0.02934169578552246, 0.029369344711303712, 0.02932633590698242, 0.02933247947692871, 0.029154304504394532, 0.02942464065551758, 0.029464576721191408, 0.02938368034362793, 0.0293621768951416, 0.029369344711303712, 0.02937343978881836, 0.029478912353515626, 0.029418560028076173, 0.029359039306640626, 0.02944819259643555, 0.029427711486816405, 0.029701120376586915, 0.029639680862426757, 0.029378559112548826, 0.029422592163085938, 0.029436927795410156, 0.029425664901733397, 0.029458431243896483, 0.02954854393005371, 0.029437952041625977, 0.029434944152832033, 0.029421503067016602, 0.029500415802001953, 0.029607936859130858, 0.029435903549194335, 0.02944819259643555, 0.029419519424438476, 0.029455360412597657, 0.029452287673950195, 0.1262366714477539, 0.02929151916503906, 0.029268991470336913, 0.02924236869812012, 0.029199359893798828, 0.029279232025146484, 0.029272064208984375, 0.029263872146606446, 0.029285375595092773, 0.029255680084228516, 0.029306880950927733, 0.029336576461791993, 0.029363199234008788, 0.029371391296386717, 0.02924236869812012, 0.029273088455200196, 0.029238271713256835, 0.029481983184814452, 0.029847551345825195, 0.029570047378540038, 0.029293567657470702, 0.029465599060058592, 0.02936422348022461, 0.029314048767089845, 0.029337600708007814, 0.029237247467041014, 0.029139968872070314, 0.029241344451904298, 0.02929254341125488, 0.029387775421142577, 0.029212671279907225, 0.029465599060058592, 0.02932326316833496, 0.029355007171630858, 0.029245471954345702, 0.02923619270324707, 0.02936524772644043, 0.02940108871459961, 0.029450239181518553, 0.029343744277954102, 0.0294021110534668, 0.029360128402709962, 0.02934886360168457, 0.029462528228759766, 0.0293621768951416, 0.029429759979248047, 0.029428735733032226, 0.02940108871459961, 0.029404159545898437, 0.029438976287841798, 0.029275136947631834, 0.029459455490112304, 0.029380607604980468, 0.029395967483520507, 0.02951372718811035, 0.029507583618164062, 0.029466623306274413, 0.029464576721191408, 0.029518848419189454, 0.029590591430664063, 0.02947782325744629, 0.029479936599731447, 0.029449216842651366, 0.1260953598022461, 0.029255680084228516, 0.029248512268066407, 0.029215744018554687, 0.029301759719848632, 0.029268991470336913, 0.029257728576660157, 0.029257728576660157, 0.029296640396118165, 0.029315071105957033, 0.029283327102661134, 0.029255680084228516, 0.029272064208984375, 0.029463552474975587, 0.029244415283203123, 0.029273088455200196, 0.02916454315185547, 0.028873727798461913, 0.029005823135375978, 0.02894643211364746, 0.028859392166137695, 0.028887039184570314, 0.028932096481323243, 0.028890111923217773, 0.028908544540405274, 0.02900275230407715, 0.029026336669921875, 0.028940256118774415, 0.028931072235107422, 0.029551616668701174, 0.029474815368652343, 0.02929151916503906, 0.029443071365356444, 0.029492223739624023, 0.028997631072998048, 0.029246463775634765, 0.029266944885253908, 0.02916454315185547, 0.029344768524169923, 0.02957107162475586, 0.029434879302978514, 0.028959743499755858, 0.028916736602783204, 0.028895231246948243, 0.029159423828125, 0.029484031677246093, 0.029386751174926756, 0.029351936340332032, 0.029464576721191408, 0.02943283271789551, 0.028957696914672853, 0.0291276798248291, 0.02956800079345703, 0.02944000053405762, 0.02939289665222168, 0.029436927795410156, 0.029062143325805666, 0.029420543670654296, 0.029449216842651366, 0.02936524772644043, 0.029820928573608397, 0.029611007690429687, 0.02944000053405762]",tokens/s,32.63345079402628,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,1,1,cuda,0,42,,,True,True,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/1/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669325ef-1d7e9cc93bf178490db05b44;4dd12aa1-1294-47f1-a187-77f09d21cd30) Repository Not Found for url: https://huggingface.co/1/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: 1 is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-7b,tiiuae/falcon-7b,cuda,0,42,,,True,True,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ self.load_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 250, in load_model_with_no_weights self.pretrained_model = self.automodel_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 559, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3710, in from_pretrained model = cls(config, *model_args, **model_kwargs) File ""/root/.cache/huggingface/modules/transformers_modules/tiiuae/falcon-7b/898df1396f35e447d5fe44e0a3ccaaaa69f30d36/modeling_falcon.py"", line 843, in __init__ self.transformer = FalconModel(config) File ""/root/.cache/huggingface/modules/transformers_modules/tiiuae/falcon-7b/898df1396f35e447d5fe44e0a3ccaaaa69f30d36/modeling_falcon.py"", line 650, in __init__ self.h = nn.ModuleList([FalconDecoderLayer(config) for _ in range(config.num_hidden_layers)]) File ""/root/.cache/huggingface/modules/transformers_modules/tiiuae/falcon-7b/898df1396f35e447d5fe44e0a3ccaaaa69f30d36/modeling_falcon.py"", line 650, in self.h = nn.ModuleList([FalconDecoderLayer(config) for _ in range(config.num_hidden_layers)]) File ""/root/.cache/huggingface/modules/transformers_modules/tiiuae/falcon-7b/898df1396f35e447d5fe44e0a3ccaaaa69f30d36/modeling_falcon.py"", line 420, in __init__ self.mlp = FalconMLP(config) File ""/root/.cache/huggingface/modules/transformers_modules/tiiuae/falcon-7b/898df1396f35e447d5fe44e0a3ccaaaa69f30d36/modeling_falcon.py"", line 405, in __init__ self.dense_4h_to_h = FalconLinear(4 * hidden_size, hidden_size, bias=config.bias) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/linear.py"", line 98, in __init__ self.weight = Parameter(torch.empty((out_features, in_features), **factory_kwargs)) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_device.py"", line 78, in __torch_function__ return func(*args, **kwargs) torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 316.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,meta-llama/Llama-2-7b-hf,meta-llama/Llama-2-7b-hf,cuda,0,42,,,True,True,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ self.load_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 250, in load_model_with_no_weights self.pretrained_model = self.automodel_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3710, in from_pretrained model = cls(config, *model_args, **model_kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 1101, in __init__ self.model = LlamaModel(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 884, in __init__ [LlamaDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 884, in [LlamaDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 679, in __init__ self.mlp = LlamaMLP(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 192, in __init__ self.gate_proj = nn.Linear(self.hidden_size, self.intermediate_size, bias=config.mlp_bias) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/linear.py"", line 98, in __init__ self.weight = Parameter(torch.empty((out_features, in_features), **factory_kwargs)) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_device.py"", line 78, in __torch_function__ return func(*args, **kwargs) torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 172.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-rw-1b,tiiuae/falcon-rw-1b,cuda,0,42,,,True,True,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1069.146112,6302.466048,0.0,5662.3104,5460.681728,s,11,0.5416408309936523,0.049240075544877485,0.0002039274825441834,0.049216064453125,0.04947452926635742,0.049598928451538085,0.04969844779968262,"[0.04972332763671875, 0.04910361480712891, 0.049191200256347656, 0.049352672576904295, 0.04921939086914062, 0.04907404708862305, 0.04901577758789062, 0.04947452926635742, 0.049216064453125, 0.04900406265258789, 0.04926614379882813]",tokens/s,5199.017206354227,kWh,6.003607692721196e-07,3.289697738260924e-07,4.197974311563383e-06,5.127304854661595e-06,tokens/kWh,49928765.16153557,MB,1069.473792,6304.5632,0.0,5664.407552,5560.067072,s,11,10.435399841308593,0.9486727128462357,0.009900882562527741,0.94589892578125,0.9587894287109375,0.9673311767578125,0.9741645751953125,"[0.9758729248046875, 0.9587894287109375, 0.94589892578125, 0.9464432373046875, 0.9484384155273438, 0.9483360595703125, 0.9409358520507812, 0.9458143310546875, 0.9432595825195312, 0.9397677612304688, 0.9418433227539063]",tokens/s,66.40857183610305,kWh,1.1131132538212399e-05,6.099426448319144e-06,4.237011849843808e-05,5.9600677484969626e-05,tokens/kWh,1057034.964340592,,s,692,10.911567865371714,0.015768161655161424,0.005921252649960189,0.014926848411560058,0.01554616289138794,0.015649655580520627,0.06447429748535156,"[0.015466496467590332, 0.015452159881591796, 0.015409152030944824, 0.01564367961883545, 0.01561302375793457, 0.015385472297668457, 0.015537152290344238, 0.015535103797912597, 0.01549721622467041, 0.015544320106506348, 0.015419391632080079, 0.015547391891479492, 0.015483903884887695, 0.01570918369293213, 0.015637503623962403, 0.015462400436401368, 0.016269311904907227, 0.016055295944213867, 0.015590399742126464, 0.015546367645263673, 0.015410176277160645, 0.01600921630859375, 0.016192512512207033, 0.016316415786743164, 0.015418368339538574, 0.015455231666564942, 0.015405055999755859, 0.015399935722351075, 0.015276032447814942, 0.01509887981414795, 0.015958016395568847, 0.016149503707885742, 0.016210943222045898, 0.01581772804260254, 0.015572992324829102, 0.015656959533691405, 0.015594495773315429, 0.015569919586181641, 0.015515647888183593, 0.015493120193481445, 0.015482912063598633, 0.015476703643798828, 0.015505439758300781, 0.015480799674987793, 0.015508480072021484, 0.015531007766723632, 0.015484928131103515, 0.015503359794616698, 0.015458304405212403, 0.015500288009643554, 0.015516672134399414, 0.014855168342590331, 0.014912511825561523, 0.014935040473937988, 0.014874624252319337, 0.014935040473937988, 0.014898176193237305, 0.014906368255615234, 0.014920703887939453, 0.01488486385345459, 0.014918656349182128, 0.014973952293395995, 0.06447615814208985, 0.01490944004058838, 0.014971903800964356, 0.014898176193237305, 0.014882816314697265, 0.014928895950317383, 0.01484390354156494, 0.01491763210296631, 0.014879743576049804, 0.01491152000427246, 0.014884832382202149, 0.014937088012695313, 0.014904319763183594, 0.01539583969116211, 0.015571968078613281, 0.014867456436157226, 0.014892031669616699, 0.01569587230682373, 0.01560371208190918, 0.01556275177001953, 0.015575039863586425, 0.01557094383239746, 0.015542271614074708, 0.015524864196777344, 0.015542271614074708, 0.015494144439697266, 0.015512576103210449, 0.015536128044128418, 0.015515647888183593, 0.015550463676452637, 0.015569919586181641, 0.015571968078613281, 0.015553536415100098, 0.015637503623962403, 0.015519743919372558, 0.015511551856994628, 0.01496780776977539, 0.01489408016204834, 0.01490227222442627, 0.014896160125732421, 0.01495036792755127, 0.014946304321289062, 0.01488486385345459, 0.01489305591583252, 0.01498624038696289, 0.014883839607238769, 0.014896127700805664, 0.014916607856750488, 0.01491763210296631, 0.01491763210296631, 0.014923775672912597, 0.014921728134155274, 0.014943231582641601, 0.014945280075073243, 0.015452159881591796, 0.01568569564819336, 0.015549375534057617, 0.01548902416229248, 0.015498240470886231, 0.015519743919372558, 0.015463423728942872, 0.015507488250732422, 0.015511520385742187, 0.06459801483154297, 0.014866432189941407, 0.014921728134155274, 0.01486847972869873, 0.014851072311401366, 0.0148787202835083, 0.01490944004058838, 0.014738431930541992, 0.015633440017700194, 0.015939552307128905, 0.01549516773223877, 0.015382528305053711, 0.015470656394958496, 0.015563712120056153, 0.015571968078613281, 0.015552512168884277, 0.01552998447418213, 0.015518719673156739, 0.01491763210296631, 0.014932991981506348, 0.014873600006103516, 0.014953472137451173, 0.014953472137451173, 0.014960639953613282, 0.014924799919128418, 0.014916607856750488, 0.014910495758056641, 0.014919648170471192, 0.01491763210296631, 0.014857215881347656, 0.014937088012695313, 0.01475276756286621, 0.01467193603515625, 0.01479468822479248, 0.014816255569458007, 0.014759936332702638, 0.014935040473937988, 0.01499238395690918, 0.014921728134155274, 0.01488691234588623, 0.014950495719909668, 0.014933919906616211, 0.014950400352478027, 0.014950400352478027, 0.014941184043884278, 0.01496780776977539, 0.014929951667785645, 0.014911487579345703, 0.01494217586517334, 0.014914560317993163, 0.014871552467346191, 0.014899200439453125, 0.014942208290100097, 0.014912511825561523, 0.014934016227722167, 0.014888992309570312, 0.014936032295227051, 0.014898176193237305, 0.014971903800964356, 0.015010815620422363, 0.014898176193237305, 0.014870528221130372, 0.014892031669616699, 0.06479666900634766, 0.014880767822265625, 0.01489305591583252, 0.014887935638427734, 0.014871552467346191, 0.014877696037292481, 0.014881792068481446, 0.014942208290100097, 0.014839808464050292, 0.014928895950317383, 0.01490227222442627, 0.014948351860046387, 0.014832639694213867, 0.014957568168640138, 0.014861311912536621, 0.014897151947021485, 0.014883839607238769, 0.014913536071777344, 0.014921728134155274, 0.01488486385345459, 0.014884896278381347, 0.014939104080200195, 0.014862336158752442, 0.014904319763183594, 0.014907391548156738, 0.014926848411560058, 0.014865407943725586, 0.014946335792541504, 0.014877663612365723, 0.014916607856750488, 0.014857215881347656, 0.014990336418151855, 0.014881792068481446, 0.014904319763183594, 0.014939135551452636, 0.014912511825561523, 0.014926848411560058, 0.014905344009399414, 0.015431679725646973, 0.015286272048950195, 0.015348735809326173, 0.015511551856994628, 0.015484928131103515, 0.014880767822265625, 0.014945280075073243, 0.014945280075073243, 0.014900223731994629, 0.01487564754486084, 0.014929920196533204, 0.015058943748474121, 0.016536575317382812, 0.01590169620513916, 0.015571968078613281, 0.015552512168884277, 0.014918656349182128, 0.015006719589233398, 0.014899200439453125, 0.014918656349182128, 0.014906368255615234, 0.014958592414855957, 0.014963711738586426, 0.015045632362365722, 0.01489305591583252, 0.06461644744873046, 0.01490944004058838, 0.014944255828857422, 0.01489305591583252, 0.01485209560394287, 0.01490944004058838, 0.01485209560394287, 0.014946304321289062, 0.014940159797668457, 0.015078399658203125, 0.014919679641723632, 0.01488588809967041, 0.014927871704101562, 0.015002623558044433, 0.014949376106262208, 0.014926848411560058, 0.015009792327880859, 0.014929920196533204, 0.014906368255615234, 0.014954527854919434, 0.014886879920959472, 0.014934016227722167, 0.014963711738586426, 0.01490944004058838, 0.014904319763183594, 0.014926912307739258, 0.0148919677734375, 0.014925824165344239, 0.015003647804260254, 0.014911487579345703, 0.014914560317993163, 0.015885343551635744, 0.015710176467895506, 0.015551487922668457, 0.015532032012939453, 0.015730751991271972, 0.015538175582885743, 0.01551968002319336, 0.015539199829101562, 0.015569919586181641, 0.015536128044128418, 0.01553929615020752, 0.01546127986907959, 0.015491071701049805, 0.015462431907653808, 0.01535587215423584, 0.015316991806030274, 0.014738431930541992, 0.014819328308105468, 0.014774271965026856, 0.01472000026702881, 0.01488588809967041, 0.014773311614990235, 0.014855104446411133, 0.014779392242431641, 0.014765055656433105, 0.014751744270324708, 0.014681183815002441, 0.014832544326782226, 0.014737407684326171, 0.014706687927246094, 0.01490124797821045, 0.01472000026702881, 0.064395263671875, 0.014733311653137206, 0.014753791809082031, 0.01484390354156494, 0.014733311653137206, 0.014731264114379883, 0.014766079902648926, 0.014715904235839844, 0.014674943923950195, 0.015159296035766602, 0.015340543746948243, 0.015343647956848144, 0.01535689640045166, 0.014739456176757813, 0.01470361614227295, 0.014748671531677245, 0.014736384391784667, 0.014743552207946778, 0.014749695777893066, 0.014721023559570312, 0.014706687927246094, 0.014750720024108887, 0.014914560317993163, 0.01475487995147705, 0.014751680374145508, 0.014948351860046387, 0.014908415794372559, 0.014971903800964356, 0.014979071617126465, 0.01491763210296631, 0.015027199745178223, 0.014943231582641601, 0.014940159797668457, 0.014909472465515137, 0.014960607528686524, 0.01491763210296631, 0.014880767822265625, 0.014926848411560058, 0.014947327613830566, 0.014925824165344239, 0.014922752380371093, 0.01489408016204834, 0.014962688446044922, 0.014944255828857422, 0.01493712043762207, 0.014937055587768555, 0.014938112258911132, 0.014920703887939453, 0.015631360054016115, 0.01577881622314453, 0.015610879898071289, 0.015682559967041015, 0.01562009620666504, 0.015463423728942872, 0.015535103797912597, 0.015501312255859375, 0.015506431579589844, 0.015511551856994628, 0.015514623641967774, 0.015455231666564942, 0.01547264003753662, 0.015518719673156739, 0.015492095947265624, 0.06447411346435547, 0.01490944004058838, 0.014888959884643555, 0.014881792068481446, 0.01489510440826416, 0.014937088012695313, 0.014931967735290527, 0.014869503974914551, 0.014914560317993163, 0.014933024406433106, 0.014920672416687012, 0.014958592414855957, 0.014944255828857422, 0.014915583610534668, 0.014934016227722167, 0.014908415794372559, 0.014919679641723632, 0.01578495979309082, 0.014949407577514649, 0.014887904167175293, 0.014928895950317383, 0.014895135879516602, 0.014882783889770507, 0.014935040473937988, 0.014922752380371093, 0.0148787202835083, 0.014945280075073243, 0.014916607856750488, 0.014943231582641601, 0.014979071617126465, 0.014854144096374512, 0.015034367561340332, 0.014910464286804198, 0.014867456436157226, 0.014892031669616699, 0.014922752380371093, 0.014870528221130372, 0.014866496086120605, 0.014942144393920899, 0.0149749755859375, 0.014896127700805664, 0.014906368255615234, 0.014871552467346191, 0.01488588809967041, 0.014901311874389648, 0.014875583648681641, 0.014907391548156738, 0.014914560317993163, 0.014921728134155274, 0.0149616641998291, 0.014934016227722167, 0.014889984130859376, 0.014844927787780762, 0.014989312171936036, 0.01489510440826416, 0.014850048065185547, 0.014922752380371093, 0.014891008377075195, 0.014979071617126465, 0.014915583610534668, 0.01488588809967041, 0.014823488235473633, 0.015007679939270019, 0.06488166046142578, 0.014866432189941407, 0.014928895950317383, 0.014915583610534668, 0.014842880249023438, 0.014896127700805664, 0.01491763210296631, 0.014919679641723632, 0.014911487579345703, 0.01486847972869873, 0.014931967735290527, 0.014904319763183594, 0.014881855964660644, 0.014996416091918945, 0.01489408016204834, 0.014889984130859376, 0.014892031669616699, 0.015239168167114257, 0.015006719589233398, 0.01488486385345459, 0.014914560317993163, 0.01487564754486084, 0.014862336158752442, 0.014881792068481446, 0.01490329647064209, 0.014953472137451173, 0.0149616641998291, 0.01499443244934082, 0.014874624252319337, 0.014939135551452636, 0.014826496124267579, 0.014888959884643555, 0.014913536071777344, 0.01489510440826416, 0.014921759605407715, 0.014885855674743651, 0.014938112258911132, 0.014913599967956543, 0.015020000457763671, 0.015045599937438965, 0.015779840469360353, 0.015703040122985838, 0.015554559707641602, 0.015608832359313965, 0.015474687576293946, 0.015490048408508301, 0.015453184127807617, 0.014861311912536621, 0.014911487579345703, 0.014962688446044922, 0.01490227222442627, 0.014925824165344239, 0.014941184043884278, 0.014951423645019531, 0.014910464286804198, 0.01497599983215332, 0.015137791633605957, 0.015069184303283691, 0.014981120109558106, 0.014958592414855957, 0.015031295776367188, 0.014938112258911132, 0.014900223731994629, 0.06468300628662109, 0.01485209560394287, 0.01490227222442627, 0.01488588809967041, 0.014870528221130372, 0.014918656349182128, 0.01489510440826416, 0.014941247940063476, 0.014903231620788574, 0.014904319763183594, 0.014950400352478027, 0.014951423645019531, 0.014906368255615234, 0.015022080421447754, 0.014942272186279297, 0.01490937614440918, 0.01496678352355957, 0.015029248237609863, 0.014957568168640138, 0.014919679641723632, 0.014940159797668457, 0.014946304321289062, 0.014925824165344239, 0.014952447891235352, 0.014958592414855957, 0.014940159797668457, 0.014943231582641601, 0.014871552467346191, 0.014995455741882324, 0.015989760398864745, 0.015640576362609862, 0.0155863676071167, 0.015521727561950683, 0.014954496383666992, 0.014952447891235352, 0.014937088012695313, 0.014951423645019531, 0.014905344009399414, 0.0149934720993042, 0.014972928047180176, 0.01499232006072998, 0.014906368255615234, 0.01506611156463623, 0.014930944442749023, 0.014904319763183594, 0.014918656349182128, 0.014872575759887695, 0.014934016227722167, 0.014810111999511719, 0.014921728134155274, 0.01498624038696289, 0.0147640323638916, 0.01476095962524414, 0.014840928077697754, 0.014763936042785645, 0.01469542407989502, 0.014821375846862793, 0.014828543663024902, 0.01490329647064209, 0.014899231910705567, 0.014949343681335448, 0.014919743537902833, 0.014947263717651368, 0.06439936065673828, 0.014844927787780762, 0.014942208290100097, 0.01490944004058838, 0.014800895690917968, 0.015002623558044433, 0.01486847972869873, 0.01477017593383789, 0.014844063758850098, 0.014779232025146484, 0.014717951774597168, 0.014718976020812988, 0.01470361614227295, 0.014924799919128418, 0.015164416313171386, 0.014927871704101562, 0.014970879554748535, 0.01498419189453125, 0.01490944004058838, 0.014908415794372559, 0.014924799919128418, 0.014919679641723632, 0.015027199745178223, 0.014910464286804198, 0.014964735984802247, 0.01495961570739746, 0.014914655685424804, 0.014933919906616211, 0.01497599983215332, 0.014932991981506348, 0.014882816314697265, 0.01489510440826416, 0.01489408016204834, 0.01489408016204834, 0.014948351860046387, 0.014960639953613282, 0.014888992309570312, 0.014881759643554688, 0.01488691234588623, 0.015017984390258789, 0.014880767822265625, 0.014923775672912597, 0.014989312171936036, 0.014939135551452636, 0.014932991981506348, 0.0148602876663208, 0.01491155242919922, 0.0148919677734375, 0.01500057601928711, 0.014928895950317383, 0.014914560317993163, 0.014861311912536621, 0.014876704216003418, 0.014899168014526367, 0.01489510440826416, 0.014907391548156738, 0.014872575759887695, 0.014927871704101562, 0.01496678352355957, 0.014921728134155274, 0.014941184043884278, 0.014993408203125, 0.014877696037292481, 0.06474547576904296, 0.01487667179107666, 0.014892031669616699, 0.014938112258911132, 0.01490124797821045, 0.014965760231018066, 0.014887935638427734, 0.0149616641998291, 0.014910464286804198, 0.014864383697509765, 0.014977024078369141, 0.014883839607238769, 0.014953472137451173, 0.0148787202835083, 0.014889984130859376, 0.014887935638427734, 0.014926912307739258, 0.014844863891601563, 0.015262720108032226, 0.015459327697753907, 0.01539583969116211, 0.015394816398620606, 0.015446016311645508, 0.014866432189941407, 0.014922752380371093, 0.01488588809967041, 0.014871552467346191, 0.014867456436157226, 0.014971903800964356, 0.014883839607238769, 0.014888959884643555, 0.014944255828857422, 0.014881792068481446, 0.0148787202835083, 0.014930944442749023, 0.014898176193237305, 0.014898176193237305, 0.014884896278381347, 0.014864352226257324, 0.014872575759887695, 0.01490944004058838, 0.01486847972869873, 0.014908415794372559, 0.014893088340759277, 0.014870495796203613, 0.01489510440826416, 0.014923775672912597, 0.01487564754486084, 0.014930944442749023, 0.01488588809967041, 0.014926848411560058, 0.014889984130859376, 0.014872575759887695, 0.01490124797821045, 0.014919679641723632, 0.014920703887939453, 0.014881792068481446, 0.014985216140747071, 0.014936063766479492, 0.014874624252319337, 0.01490124797821045, 0.014910495758056641, 0.014888928413391114]",tokens/s,63.4189337900825,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,8,8,cuda,0,42,,,True,True,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/8/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6693241f-015323c64946276a3a26656e;4bbb0464-afd6-4ccb-87fa-65e3ee21e468) Repository Not Found for url: https://huggingface.co/8/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: 8 is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-4B,,cuda,0,42,,,,,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,217063f5c507ed7cc255df7e1f64c4333a0b4dfe,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,,MB,1185.05472,18656.78848,0.0,18016.632832,17752.787968,s,10,1.20109903717041,0.120109903717041,0.00038338934699843507,0.1199699821472168,0.12053195648193359,0.12070700302124024,0.12084704025268556,"[0.12088204956054688, 0.11958604431152343, 0.119681884765625, 0.11992095947265625, 0.12049305725097656, 0.12030477142333984, 0.12041932678222657, 0.11989868927001954, 0.11989324951171874, 0.12001900482177734]",tokens/s,2131.3812772932824,kWh,1.4541449666346316e-06,7.968068603454079e-07,1.014631570512177e-05,1.2397267532101809e-05,tokens/kWh,20649711.66727724,MB,1185.05472,18656.78848,0.0,18016.632832,17960.87552,s,10,22.42018017578125,2.242018017578125,0.0027306777607785993,2.2409812011718753,2.2436963867187503,2.246627978515625,2.248973251953125,"[2.2495595703125, 2.2429951171875, 2.24173974609375, 2.240198486328125, 2.239986328125, 2.243044921875, 2.24075146484375, 2.2407548828125, 2.239942138671875, 2.24120751953125]",tokens/s,28.09968497401012,kWh,2.638014453778745e-05,1.4457266805983213e-05,0.00010865247377827962,0.0001494898851220503,tokens/kWh,421433.1956209877,,s,629,23.463265319824202,0.03730248858477618,0.014397221551963903,0.035542015075683595,0.03573063583374023,0.03584163818359375,0.15654248596191406,"[0.03633561706542969, 0.03611033630371094, 0.03634175872802734, 0.03616563034057617, 0.036103168487548826, 0.03606528091430664, 0.0361451530456543, 0.03607961654663086, 0.036155391693115234, 0.03592601776123047, 0.03586867141723633, 0.03606220626831055, 0.03559731292724609, 0.03544575881958008, 0.03543552017211914, 0.03560243225097656, 0.03548262405395508, 0.03543142318725586, 0.03540787124633789, 0.035448833465576174, 0.035465217590332034, 0.03547238540649414, 0.035460094451904296, 0.0354703369140625, 0.03546828842163086, 0.035490814208984374, 0.035465217590332034, 0.03548166275024414, 0.035493824005126955, 0.03554611206054688, 0.03548262405395508, 0.035538944244384765, 0.0354969596862793, 0.035547134399414065, 0.035517440795898435, 0.03554508972167969, 0.03556147384643555, 0.03555942535400391, 0.035566593170166014, 0.035590145111083986, 0.03559628677368164, 0.03557891082763672, 0.03555836868286133, 0.03559936141967773, 0.03562188720703125, 0.03563520050048828, 0.03563622283935547, 0.03563212966918945, 0.035664894104003905, 0.035729408264160156, 0.0356495361328125, 0.03567308807373047, 0.035678207397460936, 0.0356577262878418, 0.03564748764038086, 0.03568537521362305, 0.03570892715454101, 0.035730430603027344, 0.03571916961669922, 0.03576422500610352, 0.035772480010986325, 0.0357795524597168, 0.1562572479248047, 0.03554611206054688, 0.035434497833251956, 0.035514366149902346, 0.035381248474121094, 0.03537510299682617, 0.03539148712158203, 0.035399681091308595, 0.03541299057006836, 0.035410945892333984, 0.0354703369140625, 0.0358175048828125, 0.03563824081420899, 0.03542835235595703, 0.0354334716796875, 0.035454975128173825, 0.035631103515625, 0.035454975128173825, 0.03546316909790039, 0.03539865493774414, 0.03546623992919922, 0.035438591003417966, 0.03548672103881836, 0.03544063949584961, 0.03550310516357422, 0.03558607864379883, 0.03575088119506836, 0.03551129531860352, 0.035514366149902346, 0.03546112060546875, 0.03551232147216797, 0.035517440795898435, 0.035544063568115236, 0.03550310516357422, 0.03567923355102539, 0.035522560119628906, 0.03551641464233399, 0.03552870559692383, 0.03554816055297851, 0.03553484725952148, 0.035606529235839846, 0.03555123138427734, 0.03560451126098633, 0.036001758575439455, 0.0364400634765625, 0.035767295837402346, 0.035775489807128906, 0.03561471939086914, 0.035746814727783204, 0.03564851379394531, 0.03564851379394531, 0.03560550308227539, 0.03567103958129883, 0.03565465545654297, 0.03566387176513672, 0.03567103958129883, 0.035681278228759765, 0.03569664001464844, 0.03572531127929687, 0.035694591522216795, 0.03571814346313477, 0.035697662353515625, 0.03586764907836914, 0.15643852233886718, 0.035730430603027344, 0.03549798583984375, 0.035369983673095705, 0.03540172958374024, 0.0356495361328125, 0.035404800415039066, 0.03539046478271484, 0.03541196823120117, 0.035533824920654294, 0.036664321899414064, 0.03556966400146484, 0.03541196823120117, 0.03539251327514648, 0.03544780731201172, 0.03539865493774414, 0.03545087814331055, 0.03542323303222656, 0.03545609664916992, 0.035408798217773436, 0.03545907211303711, 0.03542118453979492, 0.03548160171508789, 0.03548364639282227, 0.035510337829589844, 0.035496894836425784, 0.035489791870117186, 0.03545804977416992, 0.03548364639282227, 0.03547238540649414, 0.03548876953125, 0.035476478576660156, 0.035519489288330076, 0.03549593734741211, 0.03551232147216797, 0.035487743377685545, 0.03552467346191406, 0.035573760986328126, 0.03558598327636719, 0.035535873413085936, 0.03557068634033203, 0.03557785415649414, 0.03565260696411133, 0.03575091171264649, 0.03560857772827149, 0.03559219360351563, 0.03561369705200195, 0.035622913360595705, 0.03563520050048828, 0.03560140609741211, 0.03566694259643555, 0.03564441680908203, 0.0356608657836914, 0.03566483306884766, 0.035659774780273434, 0.03561164855957031, 0.03566390228271484, 0.035668960571289064, 0.03568947219848633, 0.03567103958129883, 0.03574784088134766, 0.03569561767578125, 0.03575807952880859, 0.1567508544921875, 0.0358389778137207, 0.03558195114135742, 0.03541708755493164, 0.03542732620239258, 0.03541708755493164, 0.03552364730834961, 0.03536172866821289, 0.03538534545898438, 0.035381248474121094, 0.0354150390625, 0.03538844680786133, 0.035434463500976565, 0.03540582275390625, 0.03540889739990234, 0.035410945892333984, 0.035404800415039066, 0.03541708755493164, 0.03544780731201172, 0.03539558410644531, 0.03543558502197266, 0.03541907119750977, 0.035489791870117186, 0.0354252815246582, 0.03545907211303711, 0.03549184036254883, 0.0355153923034668, 0.035489791870117186, 0.03549193572998047, 0.03548662567138672, 0.03553484725952148, 0.03551846313476562, 0.03558707046508789, 0.03553484725952148, 0.03553177642822265, 0.0354969596862793, 0.03553484725952148, 0.03554304122924805, 0.03555433654785156, 0.035549152374267576, 0.03558198547363281, 0.03560752105712891, 0.03561471939086914, 0.03560140609741211, 0.03562496185302735, 0.035574783325195314, 0.035606529235839846, 0.035574783325195314, 0.035606529235839846, 0.03564748764038086, 0.035659774780273434, 0.03563827133178711, 0.03565260696411133, 0.035664894104003905, 0.0356577262878418, 0.03563008117675781, 0.03565875244140625, 0.03566592025756836, 0.035672065734863284, 0.03567411041259766, 0.03575807952880859, 0.03574169540405273, 0.0357314567565918, 0.15726591491699218, 0.035573760986328126, 0.0354703369140625, 0.03541299057006836, 0.03549491119384766, 0.03552870559692383, 0.03549388885498047, 0.03544678497314453, 0.035386367797851564, 0.03540070343017578, 0.03539865493774414, 0.035381248474121094, 0.03544575881958008, 0.035378177642822264, 0.03542323303222656, 0.0354150390625, 0.03545395278930664, 0.035434497833251956, 0.03544678497314453, 0.03543961715698242, 0.035519489288330076, 0.0354252815246582, 0.03545600128173828, 0.035434497833251956, 0.03554099273681641, 0.0354785270690918, 0.03549388885498047, 0.03546214294433594, 0.03546931076049804, 0.03548876953125, 0.03550207901000976, 0.035492862701416016, 0.03552460861206055, 0.03553279876708984, 0.03553283309936523, 0.03549795150756836, 0.035520511627197264, 0.035536895751953124, 0.03556147384643555, 0.035533824920654294, 0.035571712493896485, 0.03556454467773437, 0.03561983871459961, 0.03556966400146484, 0.03561369705200195, 0.0355860481262207, 0.035606529235839846, 0.03558297729492187, 0.03561881637573242, 0.03560345458984375, 0.03567411041259766, 0.03561881637573242, 0.03567308807373047, 0.03564963150024414, 0.035669921875, 0.0356495361328125, 0.03567308807373047, 0.035727359771728515, 0.03570995330810547, 0.03568025588989258, 0.035722240447998044, 0.03570892715454101, 0.03575193786621094, 0.15702117919921876, 0.035576831817626955, 0.035432449340820314, 0.03544985580444336, 0.03542118453979492, 0.03541196823120117, 0.0354334716796875, 0.035404800415039066, 0.035386463165283204, 0.03540470504760742, 0.035381248474121094, 0.03539763259887695, 0.03543040084838867, 0.03542732620239258, 0.03543142318725586, 0.03538022232055664, 0.03543961715698242, 0.0354334716796875, 0.03548876953125, 0.03543552017211914, 0.0354703369140625, 0.03543654251098633, 0.035460094451904296, 0.03544268798828125, 0.03545907211303711, 0.03548262405395508, 0.03549798583984375, 0.035665985107421874, 0.036125633239746095, 0.035643390655517575, 0.03550515365600586, 0.03546931076049804, 0.03551027297973633, 0.03553279876708984, 0.03551129531860352, 0.035522560119628906, 0.03554099273681641, 0.03554611206054688, 0.03558297729492187, 0.035542015075683595, 0.03566284942626953, 0.03619430541992188, 0.03571200180053711, 0.03555129623413086, 0.035611583709716794, 0.03558707046508789, 0.03584102249145508, 0.035757057189941405, 0.03562905502319336, 0.03563827133178711, 0.03568233489990234, 0.0356187858581543, 0.035639297485351565, 0.03565158462524414, 0.03565465545654297, 0.035645439147949216, 0.035694591522216795, 0.03565465545654297, 0.03580828857421875, 0.03634684753417969, 0.03604582214355469, 0.035732479095458985, 0.03580108642578125, 0.15733351135253906, 0.03557273483276367, 0.0354703369140625, 0.03540991973876953, 0.03541708755493164, 0.03542630386352539, 0.03536383819580078, 0.03537408065795898, 0.035418174743652345, 0.03539247894287109, 0.03542115020751953, 0.035388416290283206, 0.03543654251098633, 0.035404800415039066, 0.03541916656494141, 0.03542832183837891, 0.03541401672363281, 0.03544575881958008, 0.03544780731201172, 0.035448833465576174, 0.03546316909790039, 0.03547340774536133, 0.03555123138427734, 0.03554816055297851, 0.035471359252929685, 0.03557686233520508, 0.03573244857788086, 0.0355153923034668, 0.03549491119384766, 0.03549388885498047, 0.035522560119628906, 0.035571712493896485, 0.03552972793579102, 0.03550310516357422, 0.03552972793579102, 0.03553996658325195, 0.03563417434692383, 0.03554304122924805, 0.03555123138427734, 0.03554611206054688, 0.03556556701660156, 0.03558707046508789, 0.03559833526611328, 0.03559423828125, 0.03561471939086914, 0.03562393569946289, 0.035664894104003905, 0.03557273483276367, 0.03558911895751953, 0.035588096618652344, 0.03564748764038086, 0.0356864013671875, 0.03567308807373047, 0.035768318176269534, 0.03566284942626953, 0.03571916961669922, 0.035692577362060544, 0.03569558334350586, 0.035688449859619144, 0.03569049453735352, 0.03570687866210937, 0.035716094970703126, 0.03571507263183594, 0.15663206481933595, 0.03553792190551758, 0.03544166564941406, 0.035399681091308595, 0.03554099273681641, 0.03541708755493164, 0.03543654251098633, 0.03538025665283203, 0.035407840728759764, 0.03540991973876953, 0.03548876953125, 0.03537919998168945, 0.03544063949584961, 0.0354252815246582, 0.03542732620239258, 0.035388416290283206, 0.035745792388916016, 0.03567308807373047, 0.035566593170166014, 0.03543251037597656, 0.035513278961181644, 0.03547340774536133, 0.03545702362060547, 0.03542937469482422, 0.03547347259521484, 0.03549798583984375, 0.03553376007080078, 0.03549491119384766, 0.035523616790771484, 0.03548771286010742, 0.03551334381103516, 0.03546316909790039, 0.035506175994873046, 0.03549491119384766, 0.035509246826171875, 0.03553792190551758, 0.03556454467773437, 0.035573760986328126, 0.03555737686157227, 0.035533824920654294, 0.035558399200439454, 0.035552257537841796, 0.03561164855957031, 0.03562188720703125, 0.03562496185302735, 0.03567923355102539, 0.03564748764038086, 0.03562905502319336, 0.03565875244140625, 0.03561369705200195, 0.03563520050048828, 0.03559833526611328, 0.03565260696411133, 0.035661823272705076, 0.03567923355102539, 0.03565875244140625, 0.035699710845947266, 0.03568025588989258, 0.03571916961669922, 0.03567411041259766, 0.035692577362060544, 0.03568124771118164, 0.035746814727783204, 0.15658291625976561, 0.03551232147216797, 0.035416065216064455, 0.03538227081298828, 0.03537510299682617, 0.03545702362060547, 0.03540889739990234, 0.03537715148925781, 0.035369983673095705, 0.03538739013671875, 0.03541196823120117, 0.03539353561401367, 0.0354334716796875, 0.03541609573364258, 0.03544060897827148, 0.03538022232055664, 0.03541408157348633, 0.03550508880615234, 0.03543654251098633, 0.03544063949584961, 0.035451904296875, 0.03545395278930664, 0.035484672546386715, 0.035419136047363284, 0.0354785270690918, 0.03557273483276367, 0.035501056671142575, 0.0354856948852539, 0.035522560119628906, 0.035514366149902346, 0.03550310516357422, 0.035448833465576174, 0.035538944244384765, 0.03555430221557617, 0.03554816055297851, 0.03549900817871094, 0.03556249618530274, 0.035568641662597655, 0.03559219360351563, 0.035558399200439454, 0.035573760986328126, 0.03555737686157227, 0.03560345458984375, 0.035574783325195314, 0.03562496185302735, 0.03561983871459961, 0.03563520050048828, 0.03562803268432617, 0.035724288940429685, 0.03566902542114258, 0.0356607666015625, 0.03561369705200195, 0.035646495819091795, 0.03567612838745117, 0.035661823272705076, 0.03566080093383789, 0.03568435287475586, 0.0356864013671875, 0.035745792388916016, 0.035688449859619144, 0.03570687866210937, 0.03569868850708008, 0.03572633743286133, 0.15683584594726563, 0.03550207901000976, 0.035404800415039066, 0.03534745788574219, 0.035422206878662106, 0.0354252815246582, 0.035422206878662106, 0.035416065216064455, 0.03544678497314453, 0.03558092880249023, 0.03550822448730469, 0.035388416290283206, 0.03540070343017578, 0.03543654251098633, 0.035487743377685545, 0.03539353561401367, 0.03542425537109375, 0.03539148712158203, 0.03544575881958008, 0.03538022232055664, 0.03545395278930664, 0.03547750473022461, 0.03546623992919922, 0.03544889450073242, 0.035457984924316406, 0.0354785270690918, 0.03558297729492187, 0.035476478576660156, 0.03552665710449219, 0.03549593734741211, 0.03549593734741211, 0.03548364639282227, 0.03548364639282227, 0.035514366149902346, 0.03551232147216797, 0.035509246826171875, 0.03552358245849609, 0.035542015075683595, 0.035536895751953124, 0.035547134399414065, 0.0355491828918457, 0.03559833526611328, 0.03558092880249023, 0.03556249618530274, 0.03567103958129883, 0.03574169540405273, 0.03565158462524414, 0.03561164855957031, 0.03561369705200195, 0.03560857772827149, 0.03563417434692383, 0.03562393569946289, 0.03563212966918945, 0.035643390655517575, 0.03563315200805664, 0.03567308807373047, 0.03583283233642578, 0.036055038452148434, 0.03584204864501953, 0.035783679962158206, 0.03577552032470703, 0.035794910430908204, 0.03602431869506836]",tokens/s,26.807862905107037,,,main,False,False,,,,,,,,,,,,,,,,,,,,,,, float32-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,internlm/internlm2-20b,internlm/internlm2-20b,cuda,0,42,,,True,,,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",internlm2,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,i,i,cuda,0,42,,,True,True,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/i/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669474a6-738578c128f9da5b7eb7db0b;32cc0dc6-06ee-4ed9-aba2-9c4a58dac5cb) Repository Not Found for url: https://huggingface.co/i/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: i is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,stabilityai/stablelm-3b-4e1t,stabilityai/stablelm-3b-4e1t,cuda,0,42,,,True,True,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1129.033728,12505.841664,0.0,11865.686016,11620.236288,s,10,0.918577033996582,0.0918577033996582,0.00043350215481452887,0.09191893005371093,0.09225840759277344,0.09231416320800782,0.09235876770019531,"[0.09074518585205078, 0.0918354263305664, 0.0918609619140625, 0.09179357147216796, 0.09155689239501953, 0.09197689819335937, 0.0920439682006836, 0.092148193359375, 0.0922460174560547, 0.09236991882324219]",tokens/s,2786.9192296936153,kWh,1.1192137980736028e-06,6.132767045597094e-07,7.85152357681138e-06,9.584014079444692e-06,tokens/kWh,26711146.06864527,MB,1129.033728,12505.841664,0.0,11865.686016,11816.868352,s,10,16.879235107421877,1.6879235107421873,0.0016672162685396855,1.6872443237304688,1.689576135253906,1.6902889831542969,1.6908592614746094,"[1.68933984375, 1.6871929931640626, 1.6872318115234375, 1.685765625, 1.6910018310546875, 1.685685302734375, 1.6872568359375, 1.6893623046875, 1.689417724609375, 1.6869808349609374]",tokens/s,37.323966162600954,kWh,1.9839015902650688e-05,1.0871522825630067e-05,8.498205068998851e-05,0.00011569258941826925,tokens/kWh,544546.5462980772,,s,629,17.679265802383423,0.02810694086229479,0.011050221523148787,0.026746879577636717,0.02693222427368164,0.02698240089416504,0.1195253778076172,"[0.026831872940063478, 0.026682367324829103, 0.026606624603271484, 0.026670047760009766, 0.02670182418823242, 0.02671513557434082, 0.026742847442626953, 0.026713024139404296, 0.026712095260620117, 0.026702816009521485, 0.02673356819152832, 0.026730495452880858, 0.02671308708190918, 0.02674176025390625, 0.026672128677368165, 0.026663936614990235, 0.026657791137695314, 0.02662406349182129, 0.026696640014648436, 0.026987520217895508, 0.027026432037353516, 0.026737760543823243, 0.026735519409179686, 0.02676121520996094, 0.0267007999420166, 0.02671308708190918, 0.026702848434448243, 0.026693632125854492, 0.026694656372070313, 0.026736640930175783, 0.026711040496826172, 0.02670182418823242, 0.026719232559204102, 0.026720256805419923, 0.026702848434448243, 0.0267509765625, 0.02674278450012207, 0.02676736068725586, 0.02674995231628418, 0.026779647827148437, 0.026771455764770507, 0.026830848693847657, 0.026832895278930666, 0.026818559646606444, 0.026818559646606444, 0.026857471466064452, 0.026840063095092775, 0.026817535400390623, 0.026845184326171875, 0.026878976821899415, 0.026831872940063478, 0.026860544204711914, 0.026927104949951174, 0.02695884895324707, 0.026887168884277345, 0.027001855850219726, 0.02727731132507324, 0.027044864654541017, 0.02698240089416504, 0.026990591049194337, 0.027015167236328123, 0.02700595283508301, 0.11953254699707032, 0.026845184326171875, 0.0269117431640625, 0.026878976821899415, 0.02667622375488281, 0.026663936614990235, 0.026690559387207033, 0.02660966491699219, 0.026615808486938477, 0.026656768798828126, 0.026678272247314453, 0.02668441581726074, 0.026705984115600587, 0.02664134407043457, 0.026627071380615236, 0.026665983200073243, 0.02668339157104492, 0.026638336181640625, 0.026652671813964843, 0.026639360427856446, 0.026630144119262695, 0.026695680618286134, 0.026652671813964843, 0.026657791137695314, 0.02670182418823242, 0.026702848434448243, 0.026681343078613282, 0.026730495452880858, 0.02671308708190918, 0.02671718406677246, 0.026738687515258788, 0.02668339157104492, 0.02671001625061035, 0.026747903823852538, 0.02674278450012207, 0.02672537612915039, 0.026755071640014647, 0.026756128311157226, 0.02677756881713867, 0.026763263702392577, 0.026798080444335938, 0.02678374481201172, 0.02679910469055176, 0.02682368087768555, 0.026808319091796876, 0.026821632385253907, 0.026847232818603517, 0.026822656631469727, 0.026825727462768553, 0.026820608139038086, 0.026855424880981447, 0.026842111587524413, 0.026851327896118164, 0.026878976821899415, 0.02690457534790039, 0.02691276741027832, 0.02691481590270996, 0.02697932815551758, 0.026900480270385742, 0.02689945602416992, 0.02693222427368164, 0.02695475196838379, 0.027010047912597656, 0.11970662689208984, 0.026770431518554686, 0.026677248001098632, 0.026614784240722656, 0.026616832733154298, 0.026658815383911134, 0.02660147285461426, 0.026602495193481446, 0.026639360427856446, 0.026622976303100586, 0.02660966491699219, 0.026673152923583986, 0.026648576736450196, 0.026658815383911134, 0.026746879577636717, 0.02667622375488281, 0.026614784240722656, 0.02668339157104492, 0.026634239196777345, 0.026648576736450196, 0.02667519950866699, 0.026697727203369142, 0.02674278450012207, 0.026933248519897462, 0.026762239456176756, 0.026893375396728515, 0.026782655715942384, 0.02672435188293457, 0.02671001625061035, 0.02674278450012207, 0.026730495452880858, 0.02672332763671875, 0.02671615982055664, 0.02673971176147461, 0.026746879577636717, 0.02673356819152832, 0.026786815643310546, 0.026824703216552736, 0.026778623580932616, 0.026849279403686522, 0.026804224014282226, 0.026805248260498047, 0.026844160079956055, 0.026813440322875977, 0.026817535400390623, 0.026824703216552736, 0.026859519958496093, 0.026835968017578125, 0.026828800201416016, 0.026851327896118164, 0.026833919525146483, 0.02689740753173828, 0.026833919525146483, 0.02687283134460449, 0.026871807098388673, 0.026848255157470705, 0.02688204765319824, 0.02689740753173828, 0.0269117431640625, 0.026892288208007813, 0.02692915153503418, 0.02693222427368164, 0.02693939208984375, 0.11950694274902343, 0.026770431518554686, 0.026638336181640625, 0.026654720306396484, 0.02672537612915039, 0.026604543685913085, 0.026608640670776368, 0.02660966491699219, 0.02660147285461426, 0.026622976303100586, 0.026631168365478516, 0.026617855072021485, 0.026649599075317384, 0.026697727203369142, 0.026626047134399415, 0.026611711502075194, 0.026642431259155275, 0.026655744552612305, 0.026615808486938477, 0.026669055938720702, 0.026637311935424804, 0.026654720306396484, 0.02668339157104492, 0.026661888122558593, 0.026650623321533205, 0.02670694351196289, 0.026645503997802734, 0.026674175262451173, 0.026693632125854492, 0.02668339157104492, 0.026688512802124024, 0.02673766326904297, 0.02671820831298828, 0.02672332763671875, 0.026788864135742187, 0.02674278450012207, 0.026736640930175783, 0.026755071640014647, 0.026756095886230468, 0.02675814437866211, 0.026779680252075194, 0.02679292869567871, 0.026806272506713868, 0.026811391830444335, 0.026824735641479493, 0.02682159996032715, 0.026803199768066405, 0.026785791397094725, 0.026811391830444335, 0.026822656631469727, 0.026850303649902343, 0.026874879837036132, 0.026830848693847657, 0.026856447219848634, 0.026876928329467774, 0.026862592697143556, 0.026892288208007813, 0.02690255928039551, 0.026932191848754884, 0.026908672332763672, 0.02690764808654785, 0.02695577621459961, 0.02700595283508301, 0.1193338851928711, 0.02678169631958008, 0.026656768798828126, 0.026600448608398438, 0.026604543685913085, 0.026646528244018555, 0.026677248001098632, 0.026719232559204102, 0.026693632125854492, 0.026634239196777345, 0.026616832733154298, 0.02666294479370117, 0.02664854431152344, 0.026655744552612305, 0.026646528244018555, 0.026643455505371092, 0.026674175262451173, 0.026770431518554686, 0.02671718406677246, 0.0267007999420166, 0.02671615982055664, 0.02669875144958496, 0.026679296493530274, 0.02670182418823242, 0.026690559387207033, 0.026688512802124024, 0.0267007999420166, 0.0267007999420166, 0.02672537612915039, 0.02671615982055664, 0.026728448867797853, 0.026736640930175783, 0.02669875144958496, 0.02675302314758301, 0.026728448867797853, 0.02672947120666504, 0.02677555274963379, 0.02674995231628418, 0.026813440322875977, 0.026792959213256837, 0.026858495712280273, 0.026851327896118164, 0.026876928329467774, 0.02689023971557617, 0.026803199768066405, 0.026834943771362304, 0.030241792678833007, 0.027033599853515625, 0.026854400634765626, 0.02687283134460449, 0.02687286376953125, 0.026890207290649414, 0.026863616943359377, 0.0269486083984375, 0.026892288208007813, 0.02689023971557617, 0.026926080703735353, 0.02695782470703125, 0.02692300796508789, 0.02693222427368164, 0.02694041633605957, 0.026925056457519532, 0.02692198371887207, 0.11964723205566406, 0.026762239456176756, 0.026614784240722656, 0.02654617691040039, 0.02655948829650879, 0.02656358337402344, 0.02657689666748047, 0.026582015991210937, 0.02656972885131836, 0.026580991744995116, 0.026611711502075194, 0.026592256546020508, 0.026625024795532228, 0.026664960861206056, 0.026620927810668944, 0.026638336181640625, 0.026661888122558593, 0.026658815383911134, 0.026653696060180664, 0.02670694351196289, 0.026651647567749022, 0.026679296493530274, 0.026694656372070313, 0.026671104431152344, 0.02667622375488281, 0.026719232559204102, 0.02668441581726074, 0.02668339157104492, 0.02672127914428711, 0.02672640037536621, 0.02675712013244629, 0.026714111328125, 0.026719232559204102, 0.02673766326904297, 0.02675712013244629, 0.02675712013244629, 0.02674278450012207, 0.026778623580932616, 0.026806272506713868, 0.02679091262817383, 0.026772480010986328, 0.026831872940063478, 0.026825727462768553, 0.026801151275634767, 0.026840063095092775, 0.026836992263793946, 0.026838016510009766, 0.026833919525146483, 0.026840063095092775, 0.026852352142333984, 0.026825759887695314, 0.026859487533569336, 0.026838016510009766, 0.026855424880981447, 0.026895360946655275, 0.026875904083251953, 0.026875904083251953, 0.026883071899414062, 0.02690662384033203, 0.026900543212890624, 0.026977216720581055, 0.02690764808654785, 0.02692300796508789, 0.11979264068603515, 0.027056127548217773, 0.026646528244018555, 0.026635263442993166, 0.02663430404663086, 0.02661062431335449, 0.026659839630126952, 0.026610687255859376, 0.026599424362182617, 0.026671104431152344, 0.026615808486938477, 0.026640384674072266, 0.026669055938720702, 0.026642431259155275, 0.026637311935424804, 0.026663936614990235, 0.026631168365478516, 0.026647552490234375, 0.026682367324829103, 0.026640384674072266, 0.026625024795532228, 0.026677248001098632, 0.026659839630126952, 0.026689535140991212, 0.026698816299438478, 0.026673088073730467, 0.02666803169250488, 0.026711040496826172, 0.026664960861206056, 0.026682367324829103, 0.02670899200439453, 0.026672128677368165, 0.026694656372070313, 0.026744863510131837, 0.026701791763305664, 0.02673356819152832, 0.026738687515258788, 0.02673151969909668, 0.026754047393798826, 0.026762239456176756, 0.02674892807006836, 0.02674483108520508, 0.026770431518554686, 0.026969087600708007, 0.027518976211547853, 0.027000831604003905, 0.026909696578979493, 0.026883071899414062, 0.026853376388549805, 0.026847232818603517, 0.026839040756225587, 0.026853376388549805, 0.026888191223144533, 0.026874879837036132, 0.02690457534790039, 0.026883071899414062, 0.026874879837036132, 0.02693017578125, 0.02692198371887207, 0.02690559959411621, 0.026927104949951174, 0.02696499252319336, 0.026933248519897462, 0.11983155059814453, 0.026760223388671876, 0.026639328002929688, 0.026636287689208983, 0.026686464309692383, 0.026681343078613282, 0.026677248001098632, 0.026657791137695314, 0.026688512802124024, 0.02671820831298828, 0.026671104431152344, 0.026653696060180664, 0.026626047134399415, 0.026636287689208983, 0.026618879318237306, 0.026633216857910157, 0.026665983200073243, 0.026650623321533205, 0.026703872680664063, 0.02672537612915039, 0.02671001625061035, 0.02668137550354004, 0.026658784866333007, 0.02669977569580078, 0.026664960861206056, 0.026686464309692383, 0.02670796775817871, 0.02673459243774414, 0.02672537612915039, 0.02675916862487793, 0.02687593650817871, 0.026799072265625, 0.02674892807006836, 0.026806272506713868, 0.026798080444335938, 0.026779647827148437, 0.026875904083251953, 0.027041791915893554, 0.026986495971679687, 0.026853376388549805, 0.026862592697143556, 0.026874879837036132, 0.026867712020874023, 0.026875904083251953, 0.02690662384033203, 0.026891263961791992, 0.02691993522644043, 0.02692403221130371, 0.02689023971557617, 0.02690662384033203, 0.02687283134460449, 0.026933248519897462, 0.026918912887573244, 0.026944511413574217, 0.026934272766113283, 0.026977279663085937, 0.02692915153503418, 0.02695577621459961, 0.026959871292114256, 0.02698240089416504, 0.026999807357788085, 0.027040767669677734, 0.027001888275146484, 0.12090160369873047, 0.02694041633605957, 0.026679296493530274, 0.026606592178344726, 0.026634239196777345, 0.026631168365478516, 0.026631168365478516, 0.02669875144958496, 0.026682367324829103, 0.026685440063476562, 0.0267007999420166, 0.02666803169250488, 0.026643455505371092, 0.02666700744628906, 0.026651647567749022, 0.026634271621704102, 0.026737632751464843, 0.026720256805419923, 0.02668441581726074, 0.026711040496826172, 0.026705919265747072, 0.02673356819152832, 0.026682367324829103, 0.02672537612915039, 0.026696704864501954, 0.02672127914428711, 0.026788896560668945, 0.02679088020324707, 0.02679910469055176, 0.026827775955200195, 0.026833919525146483, 0.026772480010986328, 0.02673766326904297, 0.026738687515258788, 0.026755071640014647, 0.02673971176147461, 0.026746879577636717, 0.02676940727233887, 0.026843135833740234, 0.026884096145629883, 0.026869760513305665, 0.026862592697143556, 0.02688204765319824, 0.026841087341308592, 0.026884096145629883, 0.02689638328552246, 0.026912832260131837, 0.02682156753540039, 0.026854400634765626, 0.02693222427368164, 0.026862592697143556, 0.02691993522644043, 0.026853376388549805, 0.02688921546936035, 0.02688921546936035, 0.02689740753173828, 0.026879999160766603, 0.02696703910827637, 0.02689023971557617, 0.026908672332763672, 0.02692300796508789, 0.026942464828491212, 0.026944511413574217, 0.12003327941894532, 0.02673151969909668, 0.026654720306396484, 0.02659328079223633, 0.02657792091369629, 0.026621952056884765, 0.026611711502075194, 0.026591232299804687, 0.026627071380615236, 0.026647552490234375, 0.026603519439697267, 0.026650623321533205, 0.026638399124145506, 0.02664543914794922, 0.026646528244018555, 0.026642431259155275, 0.026623008728027343, 0.026658784866333007, 0.026640447616577148, 0.02664031982421875, 0.026642431259155275, 0.026663936614990235, 0.026694656372070313, 0.026673152923583986, 0.026662912368774414, 0.026709024429321288, 0.026737632751464843, 0.026720256805419923, 0.026704896926879884, 0.02677350425720215, 0.026770431518554686, 0.026697727203369142, 0.0267325439453125, 0.02671001625061035, 0.026702848434448243, 0.02675814437866211, 0.02673971176147461, 0.026762239456176756, 0.0267653751373291, 0.026771392822265625, 0.02676633644104004, 0.026810367584228514, 0.026796031951904296, 0.026831872940063478, 0.026803199768066405, 0.026843135833740234, 0.02691276741027832, 0.02733363151550293, 0.027019264221191407, 0.026952735900878905, 0.026917856216430665, 0.026876928329467774, 0.026859519958496093, 0.026895360946655275, 0.026864639282226564, 0.02687385559082031, 0.02689023971557617, 0.02695475196838379, 0.026969087600708007, 0.026894336700439454, 0.026944511413574217, 0.026934272766113283, 0.02694041633605957]",tokens/s,35.57840054167869,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,huggyllama/llama-13b,huggyllama/llama-13b,cuda,0,42,,,True,,,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",llama,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-2.7b,,cuda,0,42,,,,,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,d35829e539df8480b726c647eeabf91e41eae047,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,,MB,1053.85984,11903.95904,0.0,11263.803392,10971.005952,s,10,0.8630337219238281,0.08630337219238281,0.0005855146901819462,0.08610564804077149,0.08686894912719727,0.08726029930114745,0.08757337944030762,"[0.08765164947509765, 0.08608377838134766, 0.08595763397216796, 0.08605913543701171, 0.08641343688964843, 0.0866407699584961, 0.0853704605102539, 0.08612751770019532, 0.08594735717773437, 0.086781982421875]",tokens/s,2966.280383915227,kWh,1.0394877029789819e-06,5.695905701989097e-07,7.260797718964633e-06,8.869875992142523e-06,tokens/kWh,28861733.831090808,MB,1053.85984,11903.95904,0.0,11263.803392,11168.211456,s,10,17.5690556640625,1.75690556640625,0.006780376558086965,1.7566282348632813,1.7650084350585937,1.7657110656738282,1.7662731701660155,"[1.761355224609375, 1.756353515625, 1.747863525390625, 1.752947998046875, 1.7560301513671874, 1.7569029541015626, 1.7442252197265624, 1.762111083984375, 1.7664136962890624, 1.764852294921875]",tokens/s,35.858500994374154,kWh,2.0598839637306e-05,1.1288458643625504e-05,7.888091008323623e-05,0.00011076820836416775,tokens/kWh,568755.2496369507,,s,629,18.31616716384887,0.02911950264522874,0.010287310475246484,0.027885568618774413,0.028152269363403322,0.028269132614135745,0.11421908935546875,"[0.028080127716064454, 0.027995136260986327, 0.028368896484375, 0.02839347267150879, 0.02799001693725586, 0.028070911407470703, 0.027966463088989257, 0.028026880264282225, 0.02802182388305664, 0.028065727233886718, 0.028064767837524415, 0.028641279220581056, 0.02867715263366699, 0.0284169921875, 0.02838835144042969, 0.02857164764404297, 0.028508159637451173, 0.02832896041870117, 0.028240896224975585, 0.027785215377807617, 0.028045312881469726, 0.02755072021484375, 0.027585535049438475, 0.02756096076965332, 0.027708415985107423, 0.02756915283203125, 0.027617279052734374, 0.027578367233276366, 0.027618303298950195, 0.027610111236572265, 0.02773196792602539, 0.027800575256347656, 0.027603967666625977, 0.027655168533325194, 0.027626495361328125, 0.027642879486083984, 0.027821056365966795, 0.028091392517089843, 0.028318719863891603, 0.028218368530273437, 0.028124160766601562, 0.028131391525268553, 0.02826848030090332, 0.02800230407714844, 0.02775654411315918, 0.027699199676513672, 0.02776473617553711, 0.02796544075012207, 0.027848703384399414, 0.027853824615478515, 0.027711519241333006, 0.027854816436767578, 0.027869184494018553, 0.027896831512451172, 0.02792755126953125, 0.027784191131591796, 0.02775654411315918, 0.0277708797454834, 0.027860992431640624, 0.02772585678100586, 0.027850719451904298, 0.027810815811157227, 0.11423027038574218, 0.02753331184387207, 0.027620351791381836, 0.027468864440917968, 0.027832256317138673, 0.027971584320068358, 0.027495424270629884, 0.027634687423706054, 0.027620351791381836, 0.02800230407714844, 0.027837440490722655, 0.027676671981811524, 0.028087392807006836, 0.028052383422851563, 0.028108800888061523, 0.028116992950439453, 0.027871231079101562, 0.02751283264160156, 0.027627519607543945, 0.027826175689697266, 0.02779648017883301, 0.02800230407714844, 0.027983871459960938, 0.027635711669921875, 0.02758143997192383, 0.027603967666625977, 0.02811903953552246, 0.028081151962280275, 0.027715583801269532, 0.027572223663330078, 0.027619327545166016, 0.027833343505859375, 0.028087295532226563, 0.027603967666625977, 0.027674623489379883, 0.027653120040893556, 0.027707391738891602, 0.028151872634887696, 0.027927488327026368, 0.02772787284851074, 0.027650047302246093, 0.027679744720458983, 0.02814361572265625, 0.028027904510498046, 0.02778726387023926, 0.027650047302246093, 0.02878976058959961, 0.028808191299438478, 0.028521472930908204, 0.028207103729248048, 0.028184640884399415, 0.02820908737182617, 0.02775040054321289, 0.027708415985107423, 0.02819993591308594, 0.02775449562072754, 0.027806720733642577, 0.02775040054321289, 0.02776166343688965, 0.027802623748779298, 0.027888639450073242, 0.027884544372558592, 0.028231679916381838, 0.11419033813476563, 0.02750876808166504, 0.027622367858886717, 0.027703296661376952, 0.027897855758666993, 0.027864063262939453, 0.028017663955688478, 0.028047359466552735, 0.027889663696289063, 0.02795212745666504, 0.027907072067260744, 0.02795929527282715, 0.027817983627319336, 0.027843584060668947, 0.027896831512451172, 0.027691007614135742, 0.027636768341064454, 0.027590623855590822, 0.02752105522155762, 0.027622432708740235, 0.027602880477905274, 0.027635711669921875, 0.02752921676635742, 0.027657215118408202, 0.02753945541381836, 0.027639808654785155, 0.02757427215576172, 0.027637760162353517, 0.02755276870727539, 0.027637760162353517, 0.027554815292358398, 0.027634687423706054, 0.027580415725708008, 0.027640832901000976, 0.027588607788085938, 0.02772889518737793, 0.027675647735595704, 0.02768998336791992, 0.02774527931213379, 0.027831296920776367, 0.027857919692993165, 0.02796441650390625, 0.02795929527282715, 0.02798182487487793, 0.027687936782836913, 0.0277391357421875, 0.027684864044189454, 0.027775999069213866, 0.027695104598999022, 0.02771968078613281, 0.02773606491088867, 0.027675647735595704, 0.02771455955505371, 0.027806720733642577, 0.02772275161743164, 0.02777292823791504, 0.027828224182128908, 0.027852800369262694, 0.02773094367980957, 0.027786239624023438, 0.027786239624023438, 0.027842559814453126, 0.027743295669555666, 0.11437049865722657, 0.02759884834289551, 0.027520000457763674, 0.02754457664489746, 0.027513856887817382, 0.027692031860351563, 0.027481088638305663, 0.027661312103271486, 0.027546623229980468, 0.027687936782836913, 0.02755788803100586, 0.027672576904296874, 0.02750464057922363, 0.028065792083740236, 0.027636735916137696, 0.02757427215576172, 0.027621376037597657, 0.0275281925201416, 0.027672607421875, 0.02757219123840332, 0.027630592346191408, 0.02755686378479004, 0.027653120040893556, 0.027576320648193358, 0.0276112003326416, 0.027572160720825196, 0.027641855239868163, 0.027586559295654296, 0.027615232467651366, 0.027613183975219727, 0.027760639190673828, 0.027885568618774413, 0.027585535049438475, 0.027711488723754882, 0.028269567489624024, 0.028498943328857423, 0.028224512100219725, 0.02795212745666504, 0.027715583801269532, 0.028291072845458985, 0.028035072326660155, 0.028121088027954103, 0.028033023834228517, 0.028049407958984376, 0.027807743072509765, 0.027720703125, 0.02786204719543457, 0.028103744506835938, 0.028073888778686523, 0.028128255844116212, 0.028084224700927734, 0.028052480697631835, 0.02791116714477539, 0.02778726387023926, 0.028025856018066408, 0.02813132858276367, 0.02816819190979004, 0.028096511840820314, 0.028024896621704102, 0.027795391082763674, 0.027777023315429687, 0.02798182487487793, 0.02826144027709961, 0.11479341125488281, 0.027822080612182616, 0.027892736434936522, 0.027877376556396483, 0.028070911407470703, 0.027888639450073242, 0.027676671981811524, 0.02755891227722168, 0.02791731262207031, 0.0277258243560791, 0.0275230712890625, 0.02769715118408203, 0.027509759902954102, 0.02768998336791992, 0.02753945541381836, 0.027665407180786132, 0.02759065628051758, 0.02795008087158203, 0.027845632553100585, 0.02795622444152832, 0.027983871459960938, 0.02800127983093262, 0.02799308776855469, 0.02799001693725586, 0.028019712448120116, 0.028019712448120116, 0.028061695098876953, 0.02813542366027832, 0.027975679397583008, 0.02802895927429199, 0.027966432571411133, 0.02770227241516113, 0.027586591720581054, 0.027748319625854494, 0.027966463088989257, 0.02815488052368164, 0.028047359466552735, 0.028047359466552735, 0.028028928756713867, 0.02791219139099121, 0.02771968078613281, 0.027638784408569338, 0.027666431427001953, 0.02778112030029297, 0.027665407180786132, 0.027741184234619142, 0.02778009605407715, 0.028085248947143555, 0.028247039794921876, 0.028030975341796875, 0.027797504425048827, 0.02775551986694336, 0.02773401641845703, 0.02776678466796875, 0.027868160247802733, 0.027741184234619142, 0.027757568359375, 0.028112895965576173, 0.027979776382446288, 0.027806720733642577, 0.027995136260986327, 0.02819174385070801, 0.028076032638549804, 0.11479551696777343, 0.02749951934814453, 0.027642879486083984, 0.027485183715820313, 0.027657215118408202, 0.027830272674560546, 0.027696128845214843, 0.027802623748779298, 0.028034048080444338, 0.028077056884765625, 0.027992063522338868, 0.027915264129638673, 0.02795417594909668, 0.027961376190185547, 0.027890655517578126, 0.02775859260559082, 0.027798528671264647, 0.02770534324645996, 0.02778112030029297, 0.02775961685180664, 0.027708415985107423, 0.02753740882873535, 0.027652095794677735, 0.02775961685180664, 0.027890687942504884, 0.02791935920715332, 0.028314624786376953, 0.027650047302246093, 0.027888639450073242, 0.028021760940551758, 0.02798182487487793, 0.02800230407714844, 0.028000255584716797, 0.028031999588012696, 0.028006399154663086, 0.027999231338500977, 0.028012544631958007, 0.028005376815795898, 0.028091392517089843, 0.02794291114807129, 0.02810572814941406, 0.028227584838867188, 0.02772275161743164, 0.02774630355834961, 0.02815385627746582, 0.028048416137695313, 0.028056543350219728, 0.028008447647094727, 0.028041215896606447, 0.028006399154663086, 0.028056575775146485, 0.028035072326660155, 0.028014623641967773, 0.027736032485961914, 0.02779136085510254, 0.02775449562072754, 0.027752447128295898, 0.02778009605407715, 0.027870208740234374, 0.02771353530883789, 0.027786239624023438, 0.027807743072509765, 0.027907072067260744, 0.11358617401123047, 0.027505664825439452, 0.027651071548461914, 0.02753638458251953, 0.027662336349487306, 0.02751487922668457, 0.027646976470947264, 0.02772585678100586, 0.027819007873535157, 0.027647968292236327, 0.0275230712890625, 0.02772889518737793, 0.02755891227722168, 0.027657215118408202, 0.027631616592407225, 0.027588607788085938, 0.02750668716430664, 0.027673599243164062, 0.027520000457763674, 0.027676671981811524, 0.02755891227722168, 0.027657215118408202, 0.02754560089111328, 0.027630592346191408, 0.027554815292358398, 0.027662336349487306, 0.02754252815246582, 0.027655168533325194, 0.027675647735595704, 0.027632640838623046, 0.027583488464355467, 0.027668479919433595, 0.027633663177490234, 0.027667455673217774, 0.027792383193969726, 0.027634687423706054, 0.027644927978515626, 0.027576320648193358, 0.027676671981811524, 0.027851776123046876, 0.027604991912841798, 0.027715583801269532, 0.027693056106567384, 0.027683839797973633, 0.027720703125, 0.027669536590576173, 0.027687904357910156, 0.027885568618774413, 0.027825151443481445, 0.02769715118408203, 0.027709440231323244, 0.027793407440185547, 0.027676671981811524, 0.02775551986694336, 0.027836416244506838, 0.02771251106262207, 0.02775551986694336, 0.02790399932861328, 0.027841535568237305, 0.02776780891418457, 0.02777190399169922, 0.027844608306884764, 0.027865087509155274, 0.11437670135498047, 0.027510784149169923, 0.027650047302246093, 0.027494400024414063, 0.027682815551757813, 0.02752204895019531, 0.027663360595703124, 0.02753228759765625, 0.027656192779541015, 0.02816716766357422, 0.029268991470336913, 0.028785663604736327, 0.02800230407714844, 0.027871231079101562, 0.027876352310180662, 0.02792038345336914, 0.028018688201904295, 0.028121088027954103, 0.027979776382446288, 0.027982847213745117, 0.028009471893310548, 0.028060672760009765, 0.028021760940551758, 0.02791731262207031, 0.028020736694335937, 0.027992063522338868, 0.027987968444824218, 0.02794803237915039, 0.027696128845214843, 0.027620351791381836, 0.027665407180786132, 0.02812620735168457, 0.028035072326660155, 0.028004352569580077, 0.02775859260559082, 0.027633663177490234, 0.027885568618774413, 0.027736095428466796, 0.028077024459838867, 0.02810982322692871, 0.028082176208496092, 0.02833612823486328, 0.028010496139526365, 0.028064767837524415, 0.028115999221801757, 0.028073951721191405, 0.02814259147644043, 0.028062719345092774, 0.02774835205078125, 0.02775142478942871, 0.027711488723754882, 0.027894784927368164, 0.028168224334716798, 0.028183519363403322, 0.02815488052368164, 0.028136447906494142, 0.02817024040222168, 0.028242944717407226, 0.028076032638549804, 0.027825151443481445, 0.027842559814453126, 0.027852800369262694, 0.028216320037841795, 0.11450265502929688, 0.028048383712768556, 0.027884544372558592, 0.027856895446777344, 0.02790809631347656, 0.02796031951904297, 0.02794803237915039, 0.02815590476989746, 0.02799411201477051, 0.0279552001953125, 0.02795827293395996, 0.02798899269104004, 0.02791628837585449, 0.027922431945800782, 0.027850751876831056, 0.027883520126342775, 0.027906047821044923, 0.02798182487487793, 0.02795008087158203, 0.027967487335205078, 0.02794905662536621, 0.02802079963684082, 0.02800223922729492, 0.02816716766357422, 0.027996160507202147, 0.02797772789001465, 0.028047359466552735, 0.028013568878173828, 0.02795212745666504, 0.02796441650390625, 0.02800127983093262, 0.028025856018066408, 0.028037120819091797, 0.028014591217041016, 0.02795110321044922, 0.028019712448120116, 0.027967487335205078, 0.028011520385742186, 0.027987968444824218, 0.028019712448120116, 0.028009471893310548, 0.02797875213623047, 0.02792959976196289, 0.02777292823791504, 0.027667455673217774, 0.027839487075805663, 0.02896691131591797, 0.02836787223815918, 0.028201984405517577, 0.028120063781738282, 0.028007423400878906, 0.028030975341796875, 0.028034048080444338, 0.028049407958984376, 0.028076032638549804, 0.02807913589477539, 0.028106719970703124, 0.028218368530273437, 0.02815078353881836, 0.02819174385070801, 0.028201984405517577, 0.028220415115356445, 0.0283637752532959, 0.11547647857666016, 0.02796544075012207, 0.02817535972595215, 0.027922431945800782, 0.027983871459960938, 0.0279418888092041, 0.027894784927368164, 0.027890687942504884, 0.027836416244506838, 0.027876352310180662, 0.027872255325317383, 0.02794905662536621, 0.02793164825439453, 0.028009471893310548, 0.028054527282714844, 0.02793574333190918, 0.027974655151367187, 0.028014591217041016, 0.028051456451416015, 0.027974655151367187, 0.028073984146118162, 0.027971584320068358, 0.028007423400878906, 0.027975679397583008, 0.028025856018066408, 0.028090368270874022, 0.028021760940551758, 0.028028928756713867, 0.02809343910217285, 0.027991039276123047, 0.027966463088989257, 0.0279552001953125, 0.0281395206451416, 0.028059648513793944, 0.02794803237915039, 0.027969568252563477, 0.028091360092163085, 0.02792755126953125, 0.028051456451416015, 0.02794905662536621, 0.028045312881469726, 0.0279685115814209, 0.02797670364379883, 0.028100608825683594, 0.028039167404174805, 0.027966463088989257, 0.028047359466552735, 0.028009471893310548, 0.028026880264282225, 0.028008447647094727, 0.028077056884765625, 0.028064767837524415, 0.028096511840820314, 0.028144639968872072, 0.02813542366027832, 0.028060672760009765, 0.028059648513793944, 0.028063743591308594, 0.027874336242675782, 0.02803913688659668, 0.028051456451416015, 0.028053504943847656, 0.027853824615478515]",tokens/s,34.34124587165128,,,main,False,False,,,,,,,,,,,,,,,,,,,,,,, float32-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,microsoft/phi-1_5,microsoft/phi-1_5,cuda,0,42,,,True,True,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.1,,,,MB,1141.497856,6635.913216,0.0,5995.757568,5850.446848,s,10,0.550808479309082,0.05508084793090821,0.0001617596482639911,0.0550825424194336,0.05521057929992675,0.05531405849456787,0.055396841850280765,"[0.055417537689208984, 0.05473497772216797, 0.05506742477416992, 0.05500076675415039, 0.05510351943969727, 0.05518758392333984, 0.05513391876220703, 0.05499766540527344, 0.05506780624389648, 0.055097278594970704]",tokens/s,4647.713490560618,kWh,6.683542230034953e-07,3.662269313065473e-07,4.5717895959095885e-06,5.606370750219631e-06,tokens/kWh,45662338.68674689,MB,1141.82144,6635.913216,0.0,5995.757568,5949.83168,s,10,11.876027343749998,1.187602734375,0.01422515607898662,1.1834514770507814,1.1962282348632813,1.211889630126953,1.2244187463378906,"[1.227551025390625, 1.1927479248046875, 1.19101708984375, 1.1793822021484375, 1.1833125, 1.184025634765625, 1.1790487060546875, 1.1761890869140625, 1.1835904541015625, 1.1791627197265624]",tokens/s,53.048042225294324,kWh,1.4257345077232677e-05,7.812779941217495e-06,4.8761808626491046e-05,7.083193364494121e-05,tokens/kWh,889429.3401024417,,s,629,12.351732715606687,0.01963709493737153,0.006588905323580891,0.018717695236206054,0.019358899307250977,0.020042137145996095,0.07411335083007813,"[0.020109312057495117, 0.020265983581542968, 0.020355072021484375, 0.02022400093078613, 0.020257791519165038, 0.020154367446899413, 0.020115455627441405, 0.019994623184204103, 0.020978687286376953, 0.020195327758789062, 0.020339712142944336, 0.02008883285522461, 0.0196168327331543, 0.020621248245239258, 0.020124671936035156, 0.02003046417236328, 0.020140031814575195, 0.02048204803466797, 0.02050048065185547, 0.020132863998413086, 0.02000486373901367, 0.02004991912841797, 0.019919872283935547, 0.020135936737060548, 0.020150272369384766, 0.01977859115600586, 0.019357664108276367, 0.01945497512817383, 0.019698688507080078, 0.019506175994873046, 0.019386367797851564, 0.01944371223449707, 0.01945395278930664, 0.01960038375854492, 0.019546112060546874, 0.018711551666259766, 0.01866035270690918, 0.01868083190917969, 0.018696191787719727, 0.01862860870361328, 0.018694143295288086, 0.018670591354370117, 0.01863884735107422, 0.01863987159729004, 0.01859584045410156, 0.0186746883392334, 0.01879859161376953, 0.018710527420043945, 0.01863065528869629, 0.01944371223449707, 0.019556352615356445, 0.020141056060791016, 0.018969600677490234, 0.018732032775878905, 0.01863987159729004, 0.018676736831665038, 0.01867366409301758, 0.018692096710205077, 0.018480127334594726, 0.01848320007324219, 0.018701311111450195, 0.018737152099609376, 0.07384166717529297, 0.018704416275024414, 0.018743263244628907, 0.01859174346923828, 0.018670591354370117, 0.019156991958618166, 0.019118080139160155, 0.01905971145629883, 0.01920102310180664, 0.01944988822937012, 0.019177440643310548, 0.01885798454284668, 0.0188272647857666, 0.01869004821777344, 0.0186746883392334, 0.01866444778442383, 0.01862348747253418, 0.019345407485961915, 0.019290111541748048, 0.019186687469482423, 0.01998028755187988, 0.019188735961914064, 0.019140607833862306, 0.019134464263916014, 0.018651136398315428, 0.01939558410644531, 0.01925734329223633, 0.018996288299560547, 0.019123136520385744, 0.019165184020996092, 0.019165184020996092, 0.018691072463989256, 0.018728960037231446, 0.019313663482666017, 0.018758655548095703, 0.018667520523071288, 0.01862860870361328, 0.018675712585449217, 0.018661376953125, 0.019268608093261717, 0.019313663482666017, 0.01866854476928711, 0.018749439239501953, 0.018687999725341797, 0.018728960037231446, 0.018767871856689454, 0.01866035270690918, 0.01901260757446289, 0.018749439239501953, 0.01881292724609375, 0.018686975479125977, 0.018645055770874025, 0.018688959121704103, 0.018704383850097657, 0.0186562557220459, 0.018741247177124023, 0.01882931137084961, 0.018917375564575196, 0.019317760467529296, 0.019184640884399414, 0.019094528198242186, 0.0187064323425293, 0.018727935791015626, 0.07419801330566406, 0.018714624404907225, 0.018651136398315428, 0.018677791595458983, 0.018643999099731447, 0.01869919967651367, 0.018894847869873048, 0.01928704071044922, 0.01922662353515625, 0.019365888595581054, 0.01906790351867676, 0.018718719482421875, 0.018753599166870118, 0.018667455673217773, 0.019100671768188478, 0.019396608352661132, 0.01922355270385742, 0.019301376342773437, 0.0186060791015625, 0.018962432861328125, 0.01922355270385742, 0.01937718391418457, 0.01951740837097168, 0.019363840103149413, 0.018662431716918945, 0.01871254348754883, 0.018738176345825194, 0.01865216064453125, 0.01867263984680176, 0.018775039672851563, 0.019135488510131835, 0.019364864349365234, 0.019140607833862306, 0.0186746883392334, 0.01883443260192871, 0.018735103607177735, 0.01852524757385254, 0.018735040664672853, 0.019165184020996092, 0.019390464782714844, 0.018662399291992187, 0.018770944595336913, 0.018988031387329102, 0.018739200592041014, 0.018700288772583007, 0.01923481559753418, 0.01879347229003906, 0.01923379135131836, 0.018735103607177735, 0.018741247177124023, 0.01864908790588379, 0.018741247177124023, 0.018745344161987306, 0.01865318489074707, 0.01869824028015137, 0.01879756736755371, 0.0186562557220459, 0.01884160041809082, 0.019105791091918945, 0.018700288772583007, 0.01865216064453125, 0.0187064323425293, 0.019013631820678712, 0.0741365737915039, 0.018637823104858398, 0.01904640007019043, 0.018662399291992187, 0.018761728286743166, 0.01860403251647949, 0.018701311111450195, 0.019137535095214844, 0.019084287643432618, 0.01865216064453125, 0.018696191787719727, 0.01863270378112793, 0.018751487731933594, 0.018990079879760743, 0.01919692802429199, 0.018588672637939452, 0.0186746883392334, 0.01860812759399414, 0.01863577651977539, 0.01840127944946289, 0.01863884735107422, 0.018726911544799805, 0.018681888580322267, 0.01872585678100586, 0.018791423797607423, 0.01869824028015137, 0.018707456588745116, 0.01863577651977539, 0.018775039672851563, 0.018510847091674804, 0.018540544509887694, 0.01865318489074707, 0.018935808181762694, 0.01886207962036133, 0.01863065528869629, 0.018679807662963867, 0.018702335357666015, 0.018473983764648438, 0.01845043182373047, 0.018682912826538087, 0.01933513641357422, 0.018699264526367186, 0.01860812759399414, 0.018578432083129884, 0.018637823104858398, 0.018554880142211915, 0.018478080749511717, 0.018516992568969725, 0.0187064323425293, 0.018932735443115235, 0.018719743728637696, 0.018505727767944336, 0.018619392395019533, 0.018669567108154296, 0.018683904647827147, 0.01885593605041504, 0.01921232032775879, 0.01863983917236328, 0.018679807662963867, 0.018603008270263673, 0.018727935791015626, 0.018753536224365236, 0.018586624145507814, 0.07437619018554688, 0.01868083190917969, 0.01869004821777344, 0.0186562557220459, 0.018717695236206054, 0.01862451171875, 0.019183616638183593, 0.0185743350982666, 0.01861427116394043, 0.018741247177124023, 0.01861222457885742, 0.01903206443786621, 0.01884364891052246, 0.018663423538208008, 0.018687999725341797, 0.018687999725341797, 0.01864499282836914, 0.018752511978149415, 0.018692159652709962, 0.018665407180786135, 0.019187711715698243, 0.018765823364257812, 0.01942118453979492, 0.018970624923706055, 0.018677759170532226, 0.018700288772583007, 0.018739200592041014, 0.018961408615112304, 0.01879248046875, 0.018759647369384767, 0.018685951232910156, 0.018703359603881836, 0.019283967971801756, 0.019384319305419923, 0.0188723201751709, 0.018661376953125, 0.018738176345825194, 0.018868223190307617, 0.01908940887451172, 0.018722816467285155, 0.018735103607177735, 0.018767871856689454, 0.018768896102905275, 0.01863065528869629, 0.018594816207885743, 0.01879654312133789, 0.018720767974853517, 0.018689023971557618, 0.01865216064453125, 0.018724863052368163, 0.018697216033935548, 0.01863065528869629, 0.01863680076599121, 0.01866035270690918, 0.018778112411499022, 0.018728960037231446, 0.01863884735107422, 0.018723840713500976, 0.018723840713500976, 0.018900991439819336, 0.018686975479125977, 0.018655231475830078, 0.018758655548095703, 0.0745379867553711, 0.018719776153564453, 0.01865622329711914, 0.01867263984680176, 0.01901158332824707, 0.018695167541503906, 0.01863270378112793, 0.018970624923706055, 0.018780160903930664, 0.01881395149230957, 0.01866444778442383, 0.018693119049072265, 0.0186429443359375, 0.01867263984680176, 0.01863167953491211, 0.018751487731933594, 0.018974720001220705, 0.01878118324279785, 0.018693119049072265, 0.018693119049072265, 0.018977792739868164, 0.018725887298583984, 0.01867366409301758, 0.018971647262573242, 0.01886617660522461, 0.018729984283447267, 0.018684928894042968, 0.01881804847717285, 0.01859686470031738, 0.01882931137084961, 0.018736127853393555, 0.01905971145629883, 0.0187064323425293, 0.018481151580810547, 0.018551807403564453, 0.018751487731933594, 0.018852863311767578, 0.018782207489013672, 0.018719743728637696, 0.018692096710205077, 0.018734079360961914, 0.018944000244140623, 0.01903206443786621, 0.019696640014648437, 0.01904435157775879, 0.01876479911804199, 0.019050495147705078, 0.019079168319702147, 0.018727935791015626, 0.01869004821777344, 0.018659328460693358, 0.018769920349121092, 0.01883750343322754, 0.018753536224365236, 0.01864806365966797, 0.018766847610473633, 0.018669567108154296, 0.018945024490356444, 0.01877299118041992, 0.018733055114746093, 0.01863680076599121, 0.018760736465454102, 0.0186592960357666, 0.07426252746582031, 0.01864806365966797, 0.01884876823425293, 0.018686975479125977, 0.018678783416748047, 0.018586624145507814, 0.01865011215209961, 0.01865216064453125, 0.018754560470581053, 0.018686975479125977, 0.01862246322631836, 0.01841049575805664, 0.01862553596496582, 0.018729984283447267, 0.018750463485717773, 0.01863884735107422, 0.018699264526367186, 0.018563072204589845, 0.018761728286743166, 0.0186562557220459, 0.018738176345825194, 0.018675712585449217, 0.018755584716796874, 0.018670591354370117, 0.018489343643188477, 0.018679807662963867, 0.018724863052368163, 0.01864396858215332, 0.018712575912475587, 0.018488319396972656, 0.018761728286743166, 0.018686975479125977, 0.018713600158691408, 0.0186746883392334, 0.01883647918701172, 0.018747392654418944, 0.018562047958374024, 0.018769920349121092, 0.018723840713500976, 0.018768896102905275, 0.018780160903930664, 0.01869824028015137, 0.018761728286743166, 0.018722816467285155, 0.018725887298583984, 0.018795520782470702, 0.018689023971557618, 0.018682880401611326, 0.018889728546142577, 0.018723840713500976, 0.01903104019165039, 0.01868083190917969, 0.018717695236206054, 0.018637823104858398, 0.018732032775878905, 0.018845727920532226, 0.01891632080078125, 0.01865011215209961, 0.018662399291992187, 0.01862860870361328, 0.018770944595336913, 0.01878118324279785, 0.018884607315063476, 0.074429443359375, 0.018692127227783205, 0.0186777286529541, 0.01863167953491211, 0.01877299118041992, 0.01860812759399414, 0.018784255981445314, 0.018676736831665038, 0.0186562557220459, 0.01864499282836914, 0.01841971206665039, 0.01885081672668457, 0.018530336380004883, 0.018443231582641603, 0.01842278480529785, 0.018428991317749024, 0.018423776626586914, 0.018571231842041017, 0.018667520523071288, 0.01842790412902832, 0.018415615081787108, 0.01845043182373047, 0.01868083190917969, 0.018719743728637696, 0.01869004821777344, 0.01869824028015137, 0.018719743728637696, 0.01883750343322754, 0.018691072463989256, 0.018722816467285155, 0.018684928894042968, 0.018746368408203123, 0.01880985641479492, 0.018713600158691408, 0.018713600158691408, 0.018734079360961914, 0.018679807662963867, 0.018734079360961914, 0.01884876823425293, 0.01872489547729492, 0.0186777286529541, 0.018705408096313478, 0.018700288772583007, 0.018810943603515626, 0.018678720474243165, 0.01871673583984375, 0.01873196792602539, 0.018795520782470702, 0.018787328720092773, 0.0188221435546875, 0.018717695236206054, 0.018746400833129884, 0.01866339111328125, 0.018696191787719727, 0.018732032775878905, 0.018479103088378905, 0.018472959518432617, 0.018472959518432617, 0.018529279708862305, 0.01878835105895996, 0.018723840713500976, 0.018523136138916017, 0.018535423278808593, 0.07405363464355469, 0.018515968322753908, 0.018579456329345705, 0.018732032775878905, 0.01987993621826172, 0.01942527961730957, 0.018893823623657227, 0.018715648651123046, 0.018763776779174804, 0.018828319549560546, 0.01868284797668457, 0.018695167541503906, 0.01858252716064453, 0.018689023971557618, 0.018736127853393555, 0.018686975479125977, 0.01862246322631836, 0.018722816467285155, 0.020146175384521483, 0.01907302474975586, 0.018661376953125, 0.018654207229614257, 0.01862041664123535, 0.018923519134521484, 0.018676736831665038, 0.01864396858215332, 0.018670591354370117, 0.018746368408203123, 0.01862860870361328, 0.018707456588745116, 0.0186296329498291, 0.01865216064453125, 0.01865318489074707, 0.018670591354370117, 0.018654207229614257, 0.018695167541503906, 0.01863167953491211, 0.0187064323425293, 0.01864499282836914, 0.018869247436523438, 0.018721792221069337, 0.018686975479125977, 0.01863372802734375, 0.018712575912475587, 0.01865011215209961, 0.018662399291992187, 0.018655231475830078, 0.018686975479125977, 0.018556928634643553, 0.01860710334777832, 0.0186562557220459, 0.018472959518432617, 0.01861529541015625, 0.018696191787719727, 0.01859993553161621, 0.01902387237548828, 0.018939903259277344, 0.02040012741088867, 0.019106815338134766, 0.018686975479125977, 0.018670591354370117, 0.01884569549560547, 0.018630687713623046, 0.07429014587402344, 0.018662399291992187, 0.01879859161376953, 0.01865216064453125, 0.01866649627685547, 0.018581504821777343, 0.018717695236206054, 0.018610176086425782, 0.018745344161987306, 0.01867263984680176, 0.018743295669555664, 0.018739200592041014, 0.01876479911804199, 0.0186562557220459, 0.018693119049072265, 0.018589696884155273, 0.019129344940185547, 0.018580480575561522, 0.018709503173828124, 0.01863577651977539, 0.0187064323425293, 0.01865011215209961, 0.01878118324279785, 0.018515968322753908, 0.018719743728637696, 0.018692096710205077, 0.018735103607177735, 0.018744319915771485, 0.018676736831665038, 0.018654207229614257, 0.01875660705566406, 0.018694143295288086, 0.018724863052368163, 0.018702335357666015, 0.018737152099609376, 0.01866035270690918, 0.018803712844848632, 0.018817024230957033, 0.018884607315063476, 0.018670591354370117, 0.018670591354370117, 0.018715648651123046, 0.018728960037231446, 0.018795520782470702, 0.018892799377441406, 0.0186562557220459, 0.018687999725341797, 0.01860915184020996, 0.01882828712463379, 0.01864192008972168, 0.018750463485717773, 0.018634752273559572, 0.018700288772583007, 0.018744319915771485, 0.018693119049072265, 0.018709535598754882, 0.01871151924133301, 0.018725887298583984, 0.018676736831665038, 0.0186562557220459, 0.018686975479125977, 0.018643007278442383, 0.01877190399169922]",tokens/s,50.924029404007776,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,google/recurrentgemma-2b,google/recurrentgemma-2b,cuda,0,42,,,True,,,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.223-212.873.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.0,,,,1.21.4,,,,0.12.0,,,recurrent_gemma,MB,1415.626752,14164.688896,0.0,13524.533248,11787.729408,s,10,2.3343586883544925,0.23343586883544926,0.0006978004293051389,0.23334550476074217,0.23390711364746095,0.23453468017578125,0.2350367333984375,"[0.23516224670410157, 0.23376765441894531, 0.23341693115234374, 0.23238803100585936, 0.2334952392578125, 0.23365498352050781, 0.23270803833007814, 0.2332647705078125, 0.23327407836914063, 0.2332267150878906]",tokens/s,1096.6609427981969,kWh,2.841029781729603e-06,1.5567287986074208e-06,1.2416941150476218e-05,1.681469973081324e-05,tokens/kWh,15224773.804962771,MB,1440.804864,14168.8832,0.0,13526.6304,11790.353408,s,10,20.393506713867186,2.039350671386719,0.0039042916615324363,2.039412170410156,2.0449751586914062,2.0459625549316405,2.046752471923828,"[2.046949951171875, 2.0393221435546875, 2.03423583984375, 2.036394775390625, 2.0401373291015625, 2.039502197265625, 2.0447557373046874, 2.034259521484375, 2.0376094970703127, 2.0403397216796875]",tokens/s,30.89218587265388,kWh,2.4152499489034696e-05,1.3236467364270264e-05,9.072069691512312e-05,0.00012810966376842808,tokens/kWh,491766.18021478254,,s,630,20.39208451843262,0.03236838812449622,0.0002772989871540883,0.03227545547485351,0.03262801933288574,0.03283619899749756,0.0336297486114502,"[0.03308441543579101, 0.03257241439819336, 0.03241062545776367, 0.03234406280517578, 0.03231129455566406, 0.03232153701782227, 0.03231846237182617, 0.032295936584472655, 0.033035263061523434, 0.03236044692993164, 0.032336929321289065, 0.03258569717407227, 0.03285094451904297, 0.03229600143432617, 0.03225491333007813, 0.032247806549072264, 0.032228351593017575, 0.03225907135009766, 0.03229695892333984, 0.033020927429199216, 0.032718849182128903, 0.03230822372436523, 0.03228364944458008, 0.032276481628417966, 0.03233484649658203, 0.03225600051879883, 0.03225804901123047, 0.03234406280517578, 0.03251200103759765, 0.03229695892333984, 0.03276595306396484, 0.032320510864257815, 0.035320831298828126, 0.03325337600708008, 0.03231027221679687, 0.03226419067382812, 0.03263897705078125, 0.03227340698242188, 0.03227340698242188, 0.03228364944458008, 0.032307201385498044, 0.0323061752319336, 0.03229183959960937, 0.032277503967285154, 0.032299007415771484, 0.032301055908203126, 0.03225907135009766, 0.03226726531982422, 0.03281510543823242, 0.03224166488647461, 0.03267686462402344, 0.03253964614868164, 0.032263168334960936, 0.03225702285766602, 0.03224371337890625, 0.032322559356689456, 0.03337420654296875, 0.03257753753662109, 0.0323164176940918, 0.032535552978515625, 0.03297689437866211, 0.03234304046630859, 0.032290817260742184, 0.03330047988891602, 0.03272294235229492, 0.03240345764160156, 0.03243724822998047, 0.032330753326416016, 0.032301055908203126, 0.032317440032958986, 0.03232972717285156, 0.03232767868041992, 0.032336894989013674, 0.032345088958740234, 0.03224576187133789, 0.032290817260742184, 0.032282623291015625, 0.03227033615112305, 0.03236454391479492, 0.03233894348144531, 0.03253657531738281, 0.03225088119506836, 0.032691200256347655, 0.0322949104309082, 0.03231846237182617, 0.03233590316772461, 0.032567264556884766, 0.032285694122314454, 0.03266457748413086, 0.03239116668701172, 0.032246784210205076, 0.03262464141845703, 0.03227545547485351, 0.03229183959960937, 0.032295936584472655, 0.0322949104309082, 0.03228160095214844, 0.032277503967285154, 0.03226009750366211, 0.03227443313598633, 0.032246784210205076, 0.03226931381225586, 0.033073150634765625, 0.03228160095214844, 0.032282623291015625, 0.03269836807250977, 0.03261849594116211, 0.03226419067382812, 0.032266239166259765, 0.032263168334960936, 0.032284671783447266, 0.03227545547485351, 0.0322344970703125, 0.03226009750366211, 0.03224063873291016, 0.0324771842956543, 0.03229183959960937, 0.03227033615112305, 0.032385025024414066, 0.03225088119506836, 0.03224883270263672, 0.03226419067382812, 0.03226009750366211, 0.03228672027587891, 0.03224063873291016, 0.03224371337890625, 0.03234099197387695, 0.032290817260742184, 0.03225600051879883, 0.03224371337890625, 0.032290817260742184, 0.03227852630615234, 0.03229183959960937, 0.03222937774658203, 0.032290817260742184, 0.03225190353393555, 0.0324771842956543, 0.03234201431274414, 0.03230515289306641, 0.03275059127807617, 0.03268710327148437, 0.03225395202636719, 0.03238604736328125, 0.032249855041503905, 0.03224371337890625, 0.03221196746826172, 0.03225600051879883, 0.03225907135009766, 0.032246784210205076, 0.032247806549072264, 0.03227856063842773, 0.032289760589599606, 0.032299007415771484, 0.03226214218139648, 0.032247806549072264, 0.03226828765869141, 0.032266239166259765, 0.03224883270263672, 0.03222528076171875, 0.0322344970703125, 0.03225395202636719, 0.032230400085449216, 0.03226931381225586, 0.03229695892333984, 0.032233470916748046, 0.03228057479858398, 0.03225395202636719, 0.03229388809204101, 0.032295936584472655, 0.0322242546081543, 0.032309249877929686, 0.032247806549072264, 0.032254974365234376, 0.032323585510253904, 0.03226931381225586, 0.0322426872253418, 0.0322426872253418, 0.03225907135009766, 0.032279552459716795, 0.032266239166259765, 0.03225600051879883, 0.032299007415771484, 0.03226931381225586, 0.032266239166259765, 0.03225600051879883, 0.03224883270263672, 0.032292865753173826, 0.03227238464355469, 0.032304126739501955, 0.033446910858154294, 0.03272192001342773, 0.03237887954711914, 0.032328704833984374, 0.032307201385498044, 0.03239116668701172, 0.03230822372436523, 0.03228672027587891, 0.03244441604614258, 0.032271358489990236, 0.03236556625366211, 0.032276481628417966, 0.03226009750366211, 0.0322529296875, 0.032263168334960936, 0.032249855041503905, 0.03227443313598633, 0.03226521682739258, 0.032263168334960936, 0.03225088119506836, 0.0322426872253418, 0.032546817779541014, 0.032589824676513675, 0.032639999389648434, 0.03259494400024414, 0.032304126739501955, 0.03226726531982422, 0.032247806549072264, 0.03226009750366211, 0.032322559356689456, 0.03223756790161133, 0.032231422424316404, 0.03224371337890625, 0.03223756790161133, 0.032249855041503905, 0.03226214218139648, 0.03227545547485351, 0.03225395202636719, 0.032261119842529294, 0.03227238464355469, 0.03225088119506836, 0.03226009750366211, 0.03223551940917969, 0.03223654556274414, 0.032246784210205076, 0.03224576187133789, 0.03222937774658203, 0.032198654174804685, 0.03223244857788086, 0.03228057479858398, 0.032290817260742184, 0.03225190353393555, 0.03255295944213867, 0.032271358489990236, 0.0322949104309082, 0.03228876876831055, 0.03228672027587891, 0.03224063873291016, 0.032247806549072264, 0.03222937774658203, 0.03224063873291016, 0.032244735717773435, 0.03224883270263672, 0.03257241439819336, 0.03230822372436523, 0.03234611129760742, 0.03231129455566406, 0.032314369201660156, 0.0323133430480957, 0.032312320709228515, 0.03234201431274414, 0.03234304046630859, 0.03234304046630859, 0.03229183959960937, 0.03225907135009766, 0.03225702285766602, 0.032263168334960936, 0.03270348739624023, 0.032309249877929686, 0.03226521682739258, 0.03219968032836914, 0.032247806549072264, 0.0322426872253418, 0.03229388809204101, 0.03228160095214844, 0.03227443313598633, 0.03223961639404297, 0.03226214218139648, 0.03404185485839844, 0.03277004623413086, 0.032290817260742184, 0.03230003356933594, 0.032292865753173826, 0.032276481628417966, 0.03226009750366211, 0.03222732925415039, 0.032210945129394535, 0.03223859024047852, 0.03225395202636719, 0.03239321517944336, 0.03225088119506836, 0.0322355842590332, 0.03225900650024414, 0.03228979110717774, 0.03227443313598633, 0.03226931381225586, 0.03225190353393555, 0.03230310440063477, 0.03225600051879883, 0.032382976531982424, 0.03377664184570312, 0.032791553497314455, 0.032315391540527344, 0.03229695892333984, 0.032263168334960936, 0.032284671783447266, 0.03229388809204101, 0.0322242546081543, 0.03225600051879883, 0.03224166488647461, 0.03222016143798828, 0.032254974365234376, 0.03224063873291016, 0.03225907135009766, 0.03362611389160156, 0.032421886444091795, 0.033462272644042966, 0.03266355133056641, 0.03240038299560547, 0.03232153701782227, 0.03230822372436523, 0.032287742614746096, 0.03402444839477539, 0.032405502319335935, 0.0328540153503418, 0.03262771224975586, 0.03231846237182617, 0.0322426872253418, 0.032247806549072264, 0.03223551940917969, 0.032290817260742184, 0.032266239166259765, 0.032317440032958986, 0.03227238464355469, 0.032244735717773435, 0.03226419067382812, 0.032282623291015625, 0.03225088119506836, 0.032440319061279296, 0.03277721786499024, 0.03225600051879883, 0.03224063873291016, 0.03226214218139648, 0.03225088119506836, 0.03227852630615234, 0.0322529296875, 0.032279552459716795, 0.03226521682739258, 0.032276481628417966, 0.03228057479858398, 0.03225702285766602, 0.03220684814453125, 0.03226419067382812, 0.03223961639404297, 0.032247806549072264, 0.03230822372436523, 0.03227033615112305, 0.03236044692993164, 0.03262054443359375, 0.03266559982299805, 0.032720897674560545, 0.032295936584472655, 0.03222323226928711, 0.032215038299560544, 0.03226828765869141, 0.03224371337890625, 0.03229695892333984, 0.03225702285766602, 0.032304126739501955, 0.03226214218139648, 0.03227852630615234, 0.03225395202636719, 0.032290817260742184, 0.032266239166259765, 0.032282623291015625, 0.03224371337890625, 0.03224166488647461, 0.032263168334960936, 0.03226214218139648, 0.03299123382568359, 0.03257651138305664, 0.0323870735168457, 0.03235942459106445, 0.032331775665283204, 0.03253760147094727, 0.03258572769165039, 0.03231948852539063, 0.032500736236572264, 0.03270348739624023, 0.03363123321533203, 0.032317440032958986, 0.0322529296875, 0.03344384002685547, 0.032763904571533206, 0.032753662109375, 0.03226319885253906, 0.03225596618652344, 0.03226931381225586, 0.0322426872253418, 0.03225190353393555, 0.03225804901123047, 0.03225804901123047, 0.03226726531982422, 0.03225600051879883, 0.03223961639404297, 0.03240959930419922, 0.03226931381225586, 0.032277503967285154, 0.03234099197387695, 0.03225702285766602, 0.03277004623413086, 0.03260006332397461, 0.03223756790161133, 0.032292865753173826, 0.032470016479492186, 0.0323870735168457, 0.03278131103515625, 0.032328704833984374, 0.03284889602661133, 0.032276481628417966, 0.03241676712036133, 0.032285694122314454, 0.032263168334960936, 0.032290817260742184, 0.03230515289306641, 0.03229388809204101, 0.0322242546081543, 0.0322949104309082, 0.03227033615112305, 0.03225088119506836, 0.03224166488647461, 0.03225907135009766, 0.03299123382568359, 0.032589824676513675, 0.03263078308105469, 0.032611328125, 0.03246182250976563, 0.03226419067382812, 0.03223756790161133, 0.03226214218139648, 0.03224883270263672, 0.0333496322631836, 0.03284172821044922, 0.0325478401184082, 0.0324505615234375, 0.032292865753173826, 0.03233996963500976, 0.0323061752319336, 0.03236044692993164, 0.032309249877929686, 0.03234099197387695, 0.032314369201660156, 0.03228364944458008, 0.032261119842529294, 0.032279552459716795, 0.03226521682739258, 0.03226521682739258, 0.03225190353393555, 0.03226419067382812, 0.03223756790161133, 0.032249855041503905, 0.03225804901123047, 0.032266239166259765, 0.03225804901123047, 0.03228979110717774, 0.03225600051879883, 0.03228979110717774, 0.03225804901123047, 0.03229183959960937, 0.03224576187133789, 0.032277503967285154, 0.03224371337890625, 0.03227033615112305, 0.032217086791992186, 0.03225702285766602, 0.03231846237182617, 0.032277503967285154, 0.03226726531982422, 0.032317440032958986, 0.0322529296875, 0.032279552459716795, 0.03227545547485351, 0.032271358489990236, 0.032261119842529294, 0.03222630310058594, 0.03224063873291016, 0.03225804901123047, 0.032244735717773435, 0.0322426872253418, 0.032284671783447266, 0.032266239166259765, 0.03225088119506836, 0.032295936584472655, 0.03225395202636719, 0.032282623291015625, 0.03225600051879883, 0.032320510864257815, 0.03226726531982422, 0.032246784210205076, 0.03223756790161133, 0.03226726531982422, 0.032276481628417966, 0.0322529296875, 0.03224371337890625, 0.0322344970703125, 0.03347148895263672, 0.03262156677246094, 0.03236044692993164, 0.03235430526733398, 0.03233996963500976, 0.032320510864257815, 0.03233280181884766, 0.032307201385498044, 0.032328704833984374, 0.03231027221679687, 0.032304126739501955, 0.03223551940917969, 0.03232563018798828, 0.03222732925415039, 0.03225088119506836, 0.032233470916748046, 0.032228351593017575, 0.0322529296875, 0.03227852630615234, 0.032247806549072264, 0.03226931381225586, 0.03224576187133789, 0.032254974365234376, 0.03224166488647461, 0.03227545547485351, 0.03232153701782227, 0.032261119842529294, 0.03223551940917969, 0.03227443313598633, 0.0322344970703125, 0.032334880828857424, 0.0322191047668457, 0.03227238464355469, 0.032307201385498044, 0.032261119842529294, 0.03227238464355469, 0.032266239166259765, 0.03224576187133789, 0.032315391540527344, 0.03223859024047852, 0.03225804901123047, 0.0322344970703125, 0.03223859024047852, 0.03224166488647461, 0.03227238464355469, 0.03222016143798828, 0.032261119842529294, 0.032247806549072264, 0.03227545547485351, 0.03226521682739258, 0.03227340698242188, 0.03225702285766602, 0.032246784210205076, 0.0338974723815918, 0.0330250244140625, 0.032345088958740234, 0.032314369201660156, 0.03225702285766602, 0.032261119842529294, 0.032263168334960936, 0.032568321228027344, 0.03228979110717774, 0.03227340698242188, 0.03282944107055664, 0.03304345703125, 0.032456703186035156, 0.03231846237182617, 0.03278643035888672, 0.03364556884765625, 0.03284275054931641, 0.032780288696289066, 0.032323585510253904, 0.03230515289306641, 0.03230515289306641, 0.03231027221679687, 0.032290817260742184, 0.032271358489990236, 0.0325662727355957, 0.032658432006835936, 0.032292865753173826, 0.03226726531982422, 0.03225600051879883, 0.03225702285766602, 0.032246784210205076, 0.033312767028808594, 0.03235123062133789, 0.032263168334960936, 0.03227238464355469, 0.03226419067382812, 0.03224371337890625, 0.03225088119506836, 0.03225088119506836, 0.03223961639404297, 0.03226009750366211, 0.03223859024047852, 0.032263168334960936, 0.032317440032958986, 0.03229695892333984, 0.03224371337890625, 0.0328908805847168, 0.03264409637451172, 0.03225907135009766, 0.03224883270263672, 0.03227443313598633, 0.03227340698242188, 0.032333824157714845, 0.03225600051879883, 0.03226931381225586, 0.0322529296875, 0.03229388809204101, 0.032317440032958986, 0.032292865753173826, 0.032254974365234376, 0.032249855041503905, 0.03228057479858398, 0.032261119842529294, 0.03224883270263672, 0.03223654556274414, 0.03223859024047852, 0.032246784210205076, 0.03222937774658203, 0.03224166488647461, 0.032217086791992186, 0.03224371337890625, 0.032244735717773435, 0.03227545547485351]",tokens/s,30.89434037165433,,,,,,MB,1002.020864,14150.008832,0.0,13520.338944,13508.0832,s,1,7.04644287109375,7.04644287109375,0.0,7.04644287109375,7.04644287109375,7.04644287109375,7.04644287109375,[7.04644287109375],,kWh,2.7767686430624713e-06,1.5058699454657205e-06,7.335839202010641e-06,1.1618477790538832e-05, float32-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,stabilityai/stablelm-2-1_6b,stabilityai/stablelm-2-1_6b,cuda,0,42,,,True,True,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1146.896384,7590.117376,0.0,6949.961728,6804.719616,s,10,0.5250520591735839,0.0525052059173584,0.00029772145613171885,0.05241555213928223,0.0527534252166748,0.05302911167144775,0.05324966083526611,"[0.0533047981262207, 0.05244457626342774, 0.05252767944335938, 0.05232595062255859, 0.05269216156005859, 0.05252345657348633, 0.05238652801513672, 0.05235254287719727, 0.05219772720336914, 0.05229663848876953]",tokens/s,4875.706999472323,kWh,6.486730596913521e-07,3.5538707580584384e-07,4.55767820594539e-06,5.561738341442585e-06,tokens/kWh,46028774.50966159,MB,1146.896384,7590.117376,0.0,6949.961728,6904.4864,s,10,12.46740625,1.2467406249999997,0.012987626456919968,1.2483598022460938,1.2645401611328124,1.2647395629882812,1.2648990844726562,"[1.2483555908203126, 1.230513671875, 1.2362603759765625, 1.259568359375, 1.2252486572265624, 1.24107470703125, 1.26493896484375, 1.264495849609375, 1.248364013671875, 1.2485860595703124]",tokens/s,50.531761568289326,kWh,1.4604661830117939e-05,8.003160860817084e-06,5.1301283147852863e-05,7.390910583878787e-05,tokens/kWh,852398.3518000739,,s,629,12.919056360244769,0.020539040318354136,0.006232593580653284,0.019941375732421874,0.02016624641418457,0.020408320617675783,0.07202649169921875,"[0.02001203155517578, 0.019917823791503905, 0.020178943634033202, 0.019944448471069336, 0.020190208435058594, 0.020106239318847655, 0.01999667167663574, 0.019758079528808595, 0.019920896530151368, 0.020060159683227538, 0.020146175384521483, 0.019851264953613282, 0.019941375732421874, 0.019960832595825196, 0.02008678436279297, 0.01985740852355957, 0.019977216720581056, 0.02083839988708496, 0.02002841567993164, 0.019766271591186522, 0.019830848693847655, 0.019814336776733398, 0.01962700843811035, 0.02046771240234375, 0.020779008865356444, 0.02088140869140625, 0.020343807220458983, 0.020195327758789062, 0.02003660774230957, 0.01986457633972168, 0.020007936477661133, 0.02002227210998535, 0.019947519302368166, 0.019885055541992186, 0.019818496704101563, 0.019926015853881835, 0.019943424224853516, 0.01983590316772461, 0.019307519912719725, 0.01928294372558594, 0.01922969627380371, 0.01926553535461426, 0.01923891258239746, 0.019289087295532227, 0.01924198341369629, 0.019115007400512696, 0.01926348876953125, 0.01917647933959961, 0.019151840209960937, 0.01966592025756836, 0.019796991348266603, 0.020001792907714845, 0.020025344848632814, 0.019751935958862304, 0.019893247604370116, 0.01924812889099121, 0.019171327590942384, 0.01900851249694824, 0.01923686408996582, 0.01926553535461426, 0.019531776428222656, 0.019997695922851562, 0.07219097900390625, 0.01925324821472168, 0.019333120346069335, 0.019190784454345702, 0.01925632095336914, 0.01922355270385742, 0.019514368057250975, 0.019652639389038086, 0.01967100715637207, 0.019956735610961913, 0.02003455924987793, 0.01984716796875, 0.01964134407043457, 0.019927040100097656, 0.020113407135009767, 0.02002841567993164, 0.020016128540039063, 0.019959808349609375, 0.020007936477661133, 0.019955711364746095, 0.019994623184204103, 0.01983180809020996, 0.019955711364746095, 0.01966192054748535, 0.019778463363647462, 0.019951648712158203, 0.019997663497924804, 0.020172800064086914, 0.020002815246582033, 0.019902463912963866, 0.019948543548583983, 0.019762176513671875, 0.01901568031311035, 0.018940927505493164, 0.01925017547607422, 0.01926246452331543, 0.019194879531860352, 0.019817472457885742, 0.01988915252685547, 0.01988403129577637, 0.01990553665161133, 0.019770368576049805, 0.019280895233154297, 0.01922047996520996, 0.01920204734802246, 0.019140607833862306, 0.019219455718994142, 0.01922355270385742, 0.01923891258239746, 0.019162111282348633, 0.018958335876464845, 0.01900339126586914, 0.01901158332824707, 0.018970624923706055, 0.018885631561279297, 0.019139583587646485, 0.019144704818725586, 0.019174400329589843, 0.019397760391235353, 0.01949171257019043, 0.01920921516418457, 0.019118080139160155, 0.019160064697265625, 0.07202508544921875, 0.018894847869873048, 0.019054592132568358, 0.019107839584350587, 0.018957311630249024, 0.018782207489013672, 0.01904128074645996, 0.019203071594238282, 0.01928294372558594, 0.01923993682861328, 0.019151872634887695, 0.019175424575805664, 0.019324928283691405, 0.01924505615234375, 0.019117055892944337, 0.01942425537109375, 0.01925119972229004, 0.019250303268432616, 0.0190882568359375, 0.01927168083190918, 0.019301408767700194, 0.019954656600952147, 0.02008064079284668, 0.019869695663452147, 0.01969049644470215, 0.018979839324951172, 0.02008780860900879, 0.020140031814575195, 0.019156991958618166, 0.018994176864624023, 0.019647487640380858, 0.019810304641723633, 0.02004787254333496, 0.01964134407043457, 0.019693632125854493, 0.019654592514038085, 0.01924505615234375, 0.020039680480957032, 0.019739648818969727, 0.01920921516418457, 0.01969049644470215, 0.01986764717102051, 0.020494335174560546, 0.021016576766967773, 0.020304895401000975, 0.019336191177368164, 0.019478527069091797, 0.02000383949279785, 0.01992192077636719, 0.01986457633972168, 0.02100432014465332, 0.01991267204284668, 0.019941375732421874, 0.019893247604370116, 0.019910655975341796, 0.0198287353515625, 0.0198973445892334, 0.019909631729125975, 0.01985228729248047, 0.019825664520263672, 0.019957759857177734, 0.01999564743041992, 0.02001817512512207, 0.07238246154785156, 0.019957759857177734, 0.019902463912963866, 0.020093952178955078, 0.020057088851928712, 0.019907583236694337, 0.020074495315551756, 0.020059135437011717, 0.021238784790039062, 0.02005504035949707, 0.019870719909667968, 0.020012063980102537, 0.02010927963256836, 0.0198984317779541, 0.019869632720947265, 0.019898399353027344, 0.019865568161010743, 0.020091903686523437, 0.019911680221557617, 0.019937280654907227, 0.02002022361755371, 0.01991372871398926, 0.020063232421875, 0.019964927673339843, 0.019887104034423828, 0.019944448471069336, 0.019991552352905274, 0.020008960723876954, 0.019955711364746095, 0.019991552352905274, 0.019942399978637695, 0.019899391174316407, 0.02005401611328125, 0.019877887725830077, 0.019944448471069336, 0.019935232162475586, 0.02000592041015625, 0.020004831314086913, 0.019983360290527344, 0.019818496704101563, 0.019959808349609375, 0.019952640533447266, 0.020152320861816408, 0.019940351486206053, 0.019950592041015625, 0.02000486373901367, 0.02004991912841797, 0.020136959075927736, 0.0202608642578125, 0.019934207916259765, 0.020315135955810547, 0.0200898551940918, 0.01987379264831543, 0.020063232421875, 0.019895296096801757, 0.019212287902832033, 0.019652671813964843, 0.019998655319213868, 0.019939327239990236, 0.02009600067138672, 0.020068351745605468, 0.0200130558013916, 0.019969024658203126, 0.07241318511962891, 0.01904332733154297, 0.01923174476623535, 0.019211263656616212, 0.01924403190612793, 0.019094528198242186, 0.019289087295532227, 0.01921027183532715, 0.019331039428710936, 0.019260416030883788, 0.01920204734802246, 0.019292160034179686, 0.019215360641479492, 0.019150848388671874, 0.01923379135131836, 0.01922559928894043, 0.019212287902832033, 0.019054592132568358, 0.019098655700683594, 0.019611616134643555, 0.019408031463623045, 0.01974870491027832, 0.01988403129577637, 0.019969024658203126, 0.01985740852355957, 0.019738624572753907, 0.019742719650268553, 0.019945472717285157, 0.01987481689453125, 0.019870719909667968, 0.01982156753540039, 0.020006912231445313, 0.019929088592529298, 0.019954687118530275, 0.019720191955566405, 0.020059135437011717, 0.02003763198852539, 0.019894271850585937, 0.019833856582641602, 0.019713024139404296, 0.01978982353210449, 0.019909664154052733, 0.01973142433166504, 0.01904640007019043, 0.019156991958618166, 0.020281343460083007, 0.01999977684020996, 0.019775455474853515, 0.019183616638183593, 0.018946048736572265, 0.01914591979980469, 0.01901958465576172, 0.019117055892944337, 0.018998271942138673, 0.019002431869506835, 0.01896339225769043, 0.01901055908203125, 0.018921472549438476, 0.018982048034667968, 0.019191648483276365, 0.01922870445251465, 0.01903100776672363, 0.018943071365356445, 0.07202703857421874, 0.01900441551208496, 0.01923686408996582, 0.01922665596008301, 0.019155935287475587, 0.018956287384033203, 0.018982912063598634, 0.0192491512298584, 0.019178495407104493, 0.019080192565917968, 0.018968576431274413, 0.018927616119384767, 0.0192542724609375, 0.019091487884521485, 0.019215328216552734, 0.019087360382080077, 0.01923788833618164, 0.01926246452331543, 0.01920204734802246, 0.01924300765991211, 0.019283967971801756, 0.019296255111694336, 0.020113407135009767, 0.019927040100097656, 0.020048896789550782, 0.019900415420532228, 0.019983360290527344, 0.019952640533447266, 0.019955711364746095, 0.01999564743041992, 0.0198604793548584, 0.020100095748901366, 0.01997113609313965, 0.01997920036315918, 0.020008960723876954, 0.019916799545288084, 0.019941375732421874, 0.02003455924987793, 0.02003660774230957, 0.019875904083251954, 0.020047807693481447, 0.01999564743041992, 0.019990528106689453, 0.019969024658203126, 0.0198656005859375, 0.019968000411987305, 0.02000588798522949, 0.019917823791503905, 0.020028511047363282, 0.0198255672454834, 0.019981311798095702, 0.019994623184204103, 0.019960832595825196, 0.019958784103393554, 0.01986457633972168, 0.019921951293945313, 0.020371423721313477, 0.019973119735717772, 0.019904512405395508, 0.019994623184204103, 0.019764223098754884, 0.02044108772277832, 0.020068351745605468, 0.07218994903564453, 0.019598335266113282, 0.02006220817565918, 0.020033536911010744, 0.02027212715148926, 0.019972192764282228, 0.020255647659301757, 0.0200130558013916, 0.020008960723876954, 0.020091903686523437, 0.02043084716796875, 0.02087833595275879, 0.020446207046508787, 0.01923276710510254, 0.0192225284576416, 0.019194879531860352, 0.02007244873046875, 0.02005504035949707, 0.019969024658203126, 0.020001792907714845, 0.019997695922851562, 0.019978239059448243, 0.02004991912841797, 0.020009983062744142, 0.02002943992614746, 0.02000486373901367, 0.020122623443603514, 0.019983360290527344, 0.02002841567993164, 0.02009600067138672, 0.019938304901123048, 0.019977216720581056, 0.020074495315551756, 0.019977279663085937, 0.020011968612670898, 0.020025344848632814, 0.020017215728759766, 0.019993535995483397, 0.020084735870361328, 0.02003455924987793, 0.02006937599182129, 0.019999744415283204, 0.02001203155517578, 0.019973119735717772, 0.019904512405395508, 0.020011007308959963, 0.019974143981933593, 0.019959808349609375, 0.020008960723876954, 0.020017152786254884, 0.02002739143371582, 0.020106239318847655, 0.020187135696411132, 0.019989503860473632, 0.020198400497436524, 0.019927040100097656, 0.02063155174255371, 0.02286079978942871, 0.020477951049804686, 0.020207616806030275, 0.020115455627441405, 0.020180992126464844, 0.020106239318847655, 0.07206092834472656, 0.01927065658569336, 0.019795967102050782, 0.019978239059448243, 0.019975168228149414, 0.019997695922851562, 0.019992576599121094, 0.019945472717285157, 0.020053056716918944, 0.01997715187072754, 0.019981311798095702, 0.02001919937133789, 0.020067327499389647, 0.020059167861938478, 0.02005193519592285, 0.020033536911010744, 0.02000486373901367, 0.020031488418579102, 0.02000588798522949, 0.019801088333129883, 0.0200447998046875, 0.019973119735717772, 0.02004377555847168, 0.02001203155517578, 0.020011007308959963, 0.020727807998657227, 0.021276735305786134, 0.020643775939941406, 0.020557823181152343, 0.020023296356201172, 0.020039680480957032, 0.02001817512512207, 0.020174848556518556, 0.020178943634033202, 0.02007756805419922, 0.02008166313171387, 0.01992192077636719, 0.019985408782958985, 0.02028339195251465, 0.019945472717285157, 0.020362239837646484, 0.019999744415283204, 0.019985408782958985, 0.020082687377929686, 0.02003865623474121, 0.019924991607666014, 0.020124671936035156, 0.020141056060791016, 0.020178943634033202, 0.020237312316894532, 0.020389888763427736, 0.019959808349609375, 0.020335615158081053, 0.02001919937133789, 0.02008166313171387, 0.019998720169067383, 0.020025344848632814, 0.020146175384521483, 0.020121599197387697, 0.02003455924987793, 0.01986764717102051, 0.01964851188659668, 0.020051967620849608, 0.0719482879638672, 0.019336191177368164, 0.020343807220458983, 0.019769344329833984, 0.019886079788208007, 0.01992192077636719, 0.019968000411987305, 0.019962879180908204, 0.020226079940795897, 0.019179487228393556, 0.019164159774780275, 0.019190784454345702, 0.020024320602416993, 0.020131839752197265, 0.019924991607666014, 0.01988198471069336, 0.01987900733947754, 0.020139936447143555, 0.019946495056152345, 0.019981311798095702, 0.019935232162475586, 0.01984819221496582, 0.020164608001708984, 0.019993600845336915, 0.019963903427124022, 0.01998028755187988, 0.019993600845336915, 0.019937280654907227, 0.019951616287231445, 0.019984384536743165, 0.019970048904418947, 0.020001792907714845, 0.019940351486206053, 0.019949567794799804, 0.019918880462646484, 0.019850208282470704, 0.019313663482666017, 0.019999744415283204, 0.019938304901123048, 0.019940351486206053, 0.01970278358459473, 0.019954687118530275, 0.02001817512512207, 0.019991552352905274, 0.019204095840454103, 0.019280895233154297, 0.019191808700561523, 0.019561471939086913, 0.019730432510375977, 0.019378175735473634, 0.019208192825317383, 0.019191808700561523, 0.01923072052001953, 0.019883007049560548, 0.019954687118530275, 0.01997007942199707, 0.01986249542236328, 0.019948543548583983, 0.019914751052856446, 0.020074495315551756, 0.020033536911010744, 0.019947551727294923, 0.019976160049438477, 0.07208448028564453, 0.01923072052001953, 0.01940787124633789, 0.02044825553894043, 0.020187135696411132, 0.019972127914428713, 0.020087776184082032, 0.020221952438354493, 0.020182016372680665, 0.019933183670043944, 0.020011007308959963, 0.019963903427124022, 0.02004991912841797, 0.02009702491760254, 0.019928064346313477, 0.01998847961425781, 0.019945472717285157, 0.020016128540039063, 0.02001203155517578, 0.01991372871398926, 0.01984614372253418, 0.01943244743347168, 0.019556352615356445, 0.020009983062744142, 0.020107263565063475, 0.019978239059448243, 0.02003763198852539, 0.019968000411987305, 0.019811328887939454, 0.01922150421142578, 0.019307519912719725, 0.0192225284576416, 0.019108863830566408, 0.019908607482910155, 0.019689472198486328, 0.01993011283874512, 0.019190784454345702, 0.019216384887695313, 0.02003455924987793, 0.019999744415283204, 0.019957759857177734, 0.019915775299072267, 0.019923967361450197, 0.020130815505981444, 0.020239360809326173, 0.0200263671875, 0.019920896530151368, 0.019970048904418947, 0.020420608520507814, 0.020454399108886717, 0.020231168746948244, 0.02004377555847168, 0.020009983062744142, 0.02082099151611328, 0.019579904556274414, 0.01919491195678711, 0.01927471923828125, 0.019171327590942384, 0.019357696533203125, 0.019185663223266602, 0.019305503845214844, 0.01927676773071289, 0.01925939178466797]",tokens/s,48.6877665411844,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,togethercomputer/RedPajama-INCITE-Base-3B-v1,togethercomputer/RedPajama-INCITE-Base-3B-v1,cuda,0,42,,,True,True,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1150.078976,12604.407808,0.0,11964.25216,11746.409472,s,10,1.0169193267822265,0.10169193267822266,0.0009483642245462167,0.10140919876098634,0.10186781387329102,0.10318830528259278,0.10424469841003418,"[0.10450879669189453, 0.10122940826416016, 0.10129225921630859, 0.10140169525146485, 0.10157437133789063, 0.10141670227050781, 0.10117424011230469, 0.10125529479980469, 0.10149622344970703, 0.1015703353881836]",tokens/s,2517.4071655226044,kWh,1.2220457657096313e-06,6.696180996721185e-07,8.594731961690943e-06,1.048639582707269e-05,tokens/kWh,24412582.189496007,MB,1150.078976,12604.407808,0.0,11964.25216,11913.600512,s,10,19.049793212890627,1.9049793212890624,0.005443109354300446,1.9031524047851562,1.9121210449218748,1.9150102294921876,1.9173215771484375,"[1.9178994140625, 1.91147900390625, 1.90205126953125, 1.903072998046875, 1.9004093017578125, 1.9000986328125, 1.90039111328125, 1.907216796875, 1.90394287109375, 1.9032318115234375]",tokens/s,33.071225128768916,kWh,2.242578291589179e-05,1.2289913078731146e-05,8.600714927530603e-05,0.00012072284526992896,tokens/kWh,521856.48755325325,,s,629,19.93117596054077,0.031687084197998046,0.012111300472524103,0.03018035125732422,0.030503936767578125,0.03068375015258789,0.13200947814941405,"[0.0307906551361084, 0.03074764823913574, 0.030760959625244142, 0.03060940742492676, 0.030835712432861328, 0.030688255310058594, 0.03078963279724121, 0.030898176193237304, 0.030715904235839843, 0.031127552032470703, 0.03077324867248535, 0.030600191116333008, 0.030510080337524413, 0.03124019241333008, 0.031023103713989256, 0.03083776092529297, 0.030734336853027344, 0.030307327270507813, 0.030212095260620117, 0.030631935119628906, 0.030637056350708007, 0.030449663162231445, 0.030518272399902343, 0.030439424514770507, 0.030123008728027343, 0.030227455139160156, 0.030482431411743165, 0.030268415451049805, 0.03015884780883789, 0.030096384048461915, 0.030142463684082032, 0.03040460777282715, 0.030475263595581056, 0.03012505531311035, 0.03018649673461914, 0.030121023178100587, 0.030373823165893554, 0.030529535293579102, 0.030298175811767577, 0.030184383392333983, 0.030209024429321288, 0.030143487930297853, 0.030460927963256838, 0.030233600616455077, 0.030130176544189452, 0.030087167739868165, 0.03013222312927246, 0.030300159454345704, 0.03059916877746582, 0.030445568084716795, 0.03016908836364746, 0.030103551864624024, 0.03054707145690918, 0.03043008041381836, 0.03014143943786621, 0.030144512176513674, 0.030119935989379884, 0.030136320114135744, 0.03026860809326172, 0.030415679931640623, 0.03018239974975586, 0.030104576110839845, 0.1322670135498047, 0.030503936767578125, 0.030093311309814453, 0.030434303283691407, 0.030242816925048828, 0.030107648849487304, 0.030075904846191406, 0.030119935989379884, 0.030055423736572266, 0.030297088623046874, 0.030398464202880858, 0.030093311309814453, 0.030102527618408204, 0.030104671478271484, 0.03011471939086914, 0.030635007858276365, 0.030462976455688476, 0.030294015884399415, 0.03039232063293457, 0.03063091278076172, 0.030587903976440428, 0.030270463943481447, 0.030104576110839845, 0.030130176544189452, 0.030111743927001954, 0.03017830467224121, 0.030587903976440428, 0.030466047286987305, 0.03015475273132324, 0.030257152557373046, 0.030135295867919923, 0.030341119766235353, 0.030079999923706056, 0.03058380889892578, 0.0301527042388916, 0.030238719940185548, 0.030246912002563478, 0.03054899215698242, 0.030304256439208983, 0.030650367736816408, 0.030535680770874023, 0.030503936767578125, 0.030482431411743165, 0.03079270362854004, 0.030736383438110353, 0.030660608291625976, 0.03058687973022461, 0.03059097671508789, 0.03037696075439453, 0.030539775848388673, 0.030328832626342773, 0.030248992919921874, 0.030278623580932616, 0.03023052787780762, 0.030248960494995116, 0.030320640563964843, 0.03020595169067383, 0.030291967391967774, 0.030291967391967774, 0.030262271881103517, 0.03017932891845703, 0.030236671447753907, 0.030225471496582033, 0.1319976348876953, 0.03014143943786621, 0.030082048416137694, 0.030127103805541993, 0.030065664291381834, 0.03014860725402832, 0.030081024169921877, 0.03014860725402832, 0.030091264724731445, 0.030142463684082032, 0.030070783615112305, 0.030130176544189452, 0.03013734436035156, 0.030123008728027343, 0.030144512176513674, 0.030087167739868165, 0.03012403106689453, 0.030083072662353515, 0.03014041519165039, 0.030127103805541993, 0.030135295867919923, 0.03013324737548828, 0.030143487930297853, 0.03017011260986328, 0.03014656066894531, 0.03012403106689453, 0.030139392852783203, 0.03017830467224121, 0.030094335556030274, 0.03014656066894531, 0.03013222312927246, 0.03017728042602539, 0.030130176544189452, 0.030200832366943358, 0.0301711368560791, 0.030234624862670898, 0.03015372848510742, 0.030251007080078125, 0.030159872055053712, 0.030208000183105467, 0.030229503631591798, 0.030209024429321288, 0.03017523193359375, 0.03020595169067383, 0.030212160110473632, 0.030186431884765625, 0.03017523193359375, 0.030202880859375, 0.030204927444458008, 0.030246912002563478, 0.0301844482421875, 0.030213119506835938, 0.03019059181213379, 0.03019366455078125, 0.030222368240356446, 0.03026131248474121, 0.030215072631835937, 0.030262271881103517, 0.030267391204833984, 0.030237695693969727, 0.030213151931762695, 0.03063702392578125, 0.03079270362854004, 0.13209088134765626, 0.030073856353759764, 0.030023679733276368, 0.03002470397949219, 0.030089216232299806, 0.030121023178100587, 0.03007174491882324, 0.030094335556030274, 0.030062591552734375, 0.03016499137878418, 0.030023712158203125, 0.03010966491699219, 0.030089216232299806, 0.030159872055053712, 0.030057632446289062, 0.03009014320373535, 0.030010303497314452, 0.03018547248840332, 0.030134271621704102, 0.0301844482421875, 0.030121984481811522, 0.030118911743164063, 0.030115840911865234, 0.030123008728027343, 0.03019366455078125, 0.030489599227905274, 0.030486528396606444, 0.030273536682128906, 0.03019059181213379, 0.03014553642272949, 0.030195711135864257, 0.030232576370239257, 0.03037593650817871, 0.030280704498291015, 0.03015884780883789, 0.030259199142456054, 0.03017318344116211, 0.030272512435913085, 0.03016089630126953, 0.03018547248840332, 0.03017215919494629, 0.03018956756591797, 0.03018239974975586, 0.03019468879699707, 0.030202880859375, 0.030229503631591798, 0.030252031326293945, 0.030345216751098632, 0.030166015625, 0.030267391204833984, 0.03019161605834961, 0.030300159454345704, 0.030422016143798827, 0.03031449508666992, 0.030248960494995116, 0.03029097557067871, 0.03044655990600586, 0.030258176803588867, 0.030267391204833984, 0.03027052879333496, 0.030250944137573243, 0.03027561569213867, 0.030279647827148436, 0.13243391418457032, 0.030187519073486328, 0.030086143493652344, 0.030118911743164063, 0.030036991119384765, 0.030094335556030274, 0.03011686325073242, 0.030070783615112305, 0.030069759368896484, 0.030138368606567382, 0.030060543060302734, 0.030097408294677733, 0.030113792419433592, 0.030078975677490235, 0.030078975677490235, 0.030057472229003908, 0.030098432540893554, 0.030096384048461915, 0.030075904846191406, 0.03009542465209961, 0.030006208419799806, 0.02998374366760254, 0.030077951431274414, 0.030042112350463866, 0.030115840911865234, 0.03017932891845703, 0.030107648849487304, 0.030034944534301757, 0.030075904846191406, 0.030088191986083986, 0.03038822364807129, 0.030403711318969726, 0.030036863327026368, 0.030055423736572266, 0.030019584655761718, 0.030091264724731445, 0.030078975677490235, 0.030136320114135744, 0.030054399490356445, 0.030099456787109374, 0.030068735122680663, 0.030098432540893554, 0.030038015365600586, 0.03015679931640625, 0.030077951431274414, 0.03015475273132324, 0.03011686325073242, 0.030311424255371092, 0.030084096908569335, 0.030499839782714845, 0.03081625556945801, 0.03059404754638672, 0.030448640823364258, 0.03019468879699707, 0.03022643280029297, 0.030232576370239257, 0.030213119506835938, 0.03031449508666992, 0.030234624862670898, 0.030245887756347657, 0.030213119506835938, 0.030278656005859376, 0.03022336006164551, 0.13214515686035155, 0.030110719680786133, 0.03018035125732422, 0.030054399490356445, 0.030123008728027343, 0.030093311309814453, 0.030103551864624024, 0.030034944534301757, 0.030014656066894532, 0.030233407974243166, 0.030306304931640625, 0.03039334487915039, 0.030034944534301757, 0.030108671188354492, 0.030048288345336915, 0.030127071380615236, 0.030054399490356445, 0.030089216232299806, 0.030066688537597655, 0.030143487930297853, 0.030004255294799803, 0.030021600723266602, 0.030010368347167967, 0.03013324737548828, 0.030057472229003908, 0.030147584915161132, 0.030121984481811522, 0.0301711368560791, 0.030097408294677733, 0.030139392852783203, 0.030112768173217775, 0.03012505531311035, 0.030111743927001954, 0.030121984481811522, 0.03018137550354004, 0.030128128051757814, 0.030110719680786133, 0.03015065574645996, 0.030192640304565428, 0.03015679931640625, 0.03016089630126953, 0.03017011260986328, 0.030090240478515624, 0.030128128051757814, 0.030187551498413085, 0.030172128677368164, 0.03018239974975586, 0.030216224670410158, 0.030147552490234375, 0.030235647201538086, 0.030073856353759764, 0.030561279296875, 0.030261247634887696, 0.030187519073486328, 0.030208000183105467, 0.030219263076782226, 0.03019980812072754, 0.030219263076782226, 0.030246912002563478, 0.030249984741210937, 0.03020595169067383, 0.030266368865966797, 0.03019161605834961, 0.13190042114257813, 0.03017523193359375, 0.030059520721435546, 0.030086143493652344, 0.030055423736572266, 0.030070783615112305, 0.030053375244140625, 0.030097408294677733, 0.030081024169921877, 0.03011382484436035, 0.030075872421264648, 0.03013222312927246, 0.030103551864624024, 0.030104576110839845, 0.030102527618408204, 0.030111743927001954, 0.030090240478515624, 0.030079999923706056, 0.030064640045166017, 0.03016703987121582, 0.030078975677490235, 0.03017318344116211, 0.03037798309326172, 0.030213119506835938, 0.030083072662353515, 0.030128128051757814, 0.030097408294677733, 0.030035968780517577, 0.030047231674194336, 0.03003289604187012, 0.030070783615112305, 0.030096384048461915, 0.03005955123901367, 0.030054367065429688, 0.030055423736572266, 0.030070783615112305, 0.030071807861328126, 0.030070783615112305, 0.030104671478271484, 0.030089120864868164, 0.030092287063598632, 0.03018035125732422, 0.030305280685424804, 0.030323711395263672, 0.030893056869506837, 0.030600223541259765, 0.030264287948608398, 0.030047231674194336, 0.03015884780883789, 0.030094335556030274, 0.030099456787109374, 0.03016089630126953, 0.03023155212402344, 0.030282751083374023, 0.03022233581542969, 0.030262271881103517, 0.030209024429321288, 0.030204927444458008, 0.030251007080078125, 0.03021414375305176, 0.030232576370239257, 0.030276607513427735, 0.030213119506835938, 0.13201408386230468, 0.03041279983520508, 0.03040563201904297, 0.03012505531311035, 0.030118911743164063, 0.03013222312927246, 0.03015679931640625, 0.030127103805541993, 0.030101503372192383, 0.030094335556030274, 0.030106624603271483, 0.030409727096557617, 0.030345216751098632, 0.030229503631591798, 0.030111743927001954, 0.03015577507019043, 0.030307327270507813, 0.030256128311157225, 0.03013222312927246, 0.03040563201904297, 0.03057766342163086, 0.030187519073486328, 0.030443519592285157, 0.030195711135864257, 0.030095392227172852, 0.030183391571044924, 0.03014553642272949, 0.030329856872558594, 0.030098432540893554, 0.030192640304565428, 0.030126079559326172, 0.03018239974975586, 0.030139392852783203, 0.030225408554077147, 0.030135295867919923, 0.030247936248779295, 0.030166015625, 0.030242816925048828, 0.03040768051147461, 0.03023052787780762, 0.03019878387451172, 0.030229503631591798, 0.030249984741210937, 0.030291967391967774, 0.030219263076782226, 0.030242816925048828, 0.030391359329223634, 0.03023766326904297, 0.030305248260498047, 0.030406656265258788, 0.03021107292175293, 0.030266368865966797, 0.03037900733947754, 0.030294015884399415, 0.030292991638183595, 0.030292991638183595, 0.030291967391967774, 0.030307392120361328, 0.03026937675476074, 0.030279680252075194, 0.030216192245483397, 0.030299135208129883, 0.03156582450866699, 0.1324830780029297, 0.03021004867553711, 0.030123008728027343, 0.030099456787109374, 0.030054399490356445, 0.030044160842895507, 0.030103551864624024, 0.03013324737548828, 0.030047231674194336, 0.03008620834350586, 0.030052288055419922, 0.030089216232299806, 0.030083072662353515, 0.030075904846191406, 0.030088191986083986, 0.03018956756591797, 0.030112768173217775, 0.030287872314453124, 0.03041279983520508, 0.030487615585327147, 0.030404544830322264, 0.03014963150024414, 0.030109695434570313, 0.030096384048461915, 0.030069759368896484, 0.030066688537597655, 0.03008723258972168, 0.03008198356628418, 0.030095359802246095, 0.03012819290161133, 0.030113727569580077, 0.030127103805541993, 0.03015372848510742, 0.030136320114135744, 0.03012403106689453, 0.03014656066894531, 0.03016294479370117, 0.030444543838500978, 0.030127103805541993, 0.030159872055053712, 0.0301527042388916, 0.03019161605834961, 0.030131200790405273, 0.030232576370239257, 0.030121984481811522, 0.03019980812072754, 0.03016908836364746, 0.030202880859375, 0.030491647720336915, 0.030505983352661133, 0.03042508888244629, 0.03037081527709961, 0.03058380889892578, 0.030238719940185548, 0.03022233581542969, 0.030232576370239257, 0.030195711135864257, 0.030207008361816407, 0.03023049545288086, 0.030264320373535155, 0.030229503631591798, 0.030749696731567383, 0.030898176193237304, 0.13244723510742187, 0.030264320373535155, 0.03012915229797363, 0.030054399490356445, 0.030030847549438477, 0.030083072662353515, 0.030069759368896484, 0.030095359802246095, 0.030028799057006835, 0.030118911743164063, 0.030039039611816407, 0.030120960235595705, 0.03003392028808594, 0.030127103805541993, 0.030044160842895507, 0.030131200790405273, 0.030073856353759764, 0.030115840911865234, 0.030058496475219725, 0.03012505531311035, 0.03018137550354004, 0.030121984481811522, 0.030085119247436523, 0.030088224411010743, 0.030090208053588866, 0.030087167739868165, 0.03012403106689453, 0.030118911743164063, 0.030098432540893554, 0.03011686325073242, 0.030118911743164063, 0.030068735122680663, 0.030168064117431642, 0.03017420768737793, 0.0301844482421875, 0.03017932891845703, 0.03015372848510742, 0.03020595169067383, 0.030304256439208983, 0.030676992416381835, 0.030451711654663087, 0.030517248153686522, 0.030283775329589844, 0.030292991638183595, 0.03018345642089844, 0.030187488555908203, 0.0302336311340332, 0.030220256805419923, 0.030418943405151368, 0.03100569534301758, 0.030485504150390624, 0.030253055572509766, 0.030435327529907227, 0.030345216751098632, 0.030220287322998047, 0.030229503631591798, 0.030239744186401366, 0.030241792678833007, 0.030273536682128906, 0.030219263076782226, 0.030192640304565428, 0.030241792678833007, 0.03020697593688965]",tokens/s,31.558599514914626,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,meta-llama/Meta-Llama-3-8B,meta-llama/Meta-Llama-3-8B,cuda,0,42,,,True,True,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ self.load_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 250, in load_model_with_no_weights self.pretrained_model = self.automodel_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3710, in from_pretrained model = cls(config, *model_args, **model_kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 1101, in __init__ self.model = LlamaModel(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 884, in __init__ [LlamaDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 884, in [LlamaDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 679, in __init__ self.mlp = LlamaMLP(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 193, in __init__ self.up_proj = nn.Linear(self.hidden_size, self.intermediate_size, bias=config.mlp_bias) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/linear.py"", line 98, in __init__ self.weight = Parameter(torch.empty((out_features, in_features), **factory_kwargs)) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_device.py"", line 78, in __torch_function__ return func(*args, **kwargs) torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 224.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen2-beta-14B,Qwen/Qwen2-beta-14B,cuda,0,42,,,True,,,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",qwen2,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-32B,Qwen/Qwen1.5-32B,cuda,0,42,,,True,,,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",qwen2,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,0,0,cuda,0,42,,,True,True,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/0/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66932588-0c4353c817b00c67294be211;748facb7-b8f1-4eb8-87b6-fd807a3220ec) Repository Not Found for url: https://huggingface.co/0/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: 0 is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/gpt-neo-125m,EleutherAI/gpt-neo-125m,cuda,0,42,,,True,True,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1092.927488,1449.65632,0.0,809.500672,638.67392,s,19,0.19452700710296628,0.010238263531735068,0.0002581533230812105,0.010151968002319336,0.010361830711364746,0.010485673427581785,0.011090542888641359,"[0.01124176025390625, 0.010122207641601563, 0.010151968002319336, 0.010176735877990723, 0.010149472236633301, 0.010126303672790528, 0.010160320281982422, 0.010351872444152832, 0.010225024223327637, 0.010343263626098633, 0.010401663780212402, 0.010349791526794433, 0.01008902359008789, 0.010111040115356445, 0.010201631546020508, 0.010143136024475098, 0.01006668758392334, 0.01008176040649414, 0.010033344268798828]",tokens/s,25004.240143505656,kWh,1.2298004725638506e-07,6.738695076012285e-08,6.209300839791911e-07,8.11297081995699e-07,tokens/kWh,315544090.6680805,MB,1092.927488,1449.65632,0.0,809.500672,657.857536,s,19,10.221581420898438,0.5379779695209704,0.010023447601295787,0.535698486328125,0.5520516357421875,0.5531655395507813,0.5531682641601563,"[0.535698486328125, 0.5300238647460938, 0.5284521484375, 0.5290758666992188, 0.5369417724609375, 0.5414525146484375, 0.5472347412109375, 0.5517476196289063, 0.5449077758789063, 0.5531651611328126, 0.5517732543945313, 0.548458740234375, 0.5278040771484375, 0.5298796997070313, 0.528615478515625, 0.5281749877929688, 0.5277741088867187, 0.527232177734375, 0.5531689453125]",tokens/s,117.10516706863821,kWh,6.3349377197050535e-06,3.4712833805649743e-06,1.2234050960231598e-05,2.2040272060501635e-05,tokens/kWh,2858403.9174771477,,s,1196,10.395724808692918,0.008692077599241583,0.0013070878803914776,0.00841420841217041,0.008794624328613282,0.008842496156692504,0.018980044174194334,"[0.00922316837310791, 0.009273344039916993, 0.009462783813476563, 0.010035200119018555, 0.008651776313781738, 0.008361984252929687, 0.008360960006713868, 0.008417280197143554, 0.008386560440063476, 0.008473600387573242, 0.00840499210357666, 0.008429568290710449, 0.008396800041198731, 0.008775679588317872, 0.008435711860656739, 0.008444928169250488, 0.008393728256225585, 0.00840396785736084, 0.00841215991973877, 0.008364031791687012, 0.008399871826171875, 0.008378368377685547, 0.008461312294006347, 0.008407039642333984, 0.008383487701416016, 0.008398847579956055, 0.00840499210357666, 0.008416255950927735, 0.008384511947631837, 0.008401920318603515, 0.008425472259521484, 0.008392704010009766, 0.008555520057678222, 0.008447999954223634, 0.00840396785736084, 0.008385536193847656, 0.008373248100280761, 0.008398847579956055, 0.008357888221740722, 0.008423487663269042, 0.008473535537719727, 0.008375295639038086, 0.008489983558654785, 0.008390656471252441, 0.008388607978820802, 0.008423423767089844, 0.008516608238220215, 0.0084203519821167, 0.008385536193847656, 0.00837939167022705, 0.008364031791687012, 0.008354816436767578, 0.008391679763793946, 0.008317952156066894, 0.008350720405578613, 0.008351743698120117, 0.008341504096984862, 0.008356863975524903, 0.008370176315307617, 0.008368127822875977, 0.008458239555358887, 0.008382464408874512, 0.019121152877807617, 0.008368127822875977, 0.008374272346496582, 0.008367103576660156, 0.008377344131469726, 0.008515583992004394, 0.008545280456542969, 0.008493056297302246, 0.008540160179138183, 0.00851865577697754, 0.008546303749084473, 0.008438783645629883, 0.008479743957519532, 0.008378368377685547, 0.008447999954223634, 0.008450048446655273, 0.00850227165222168, 0.00841318416595459, 0.008398847579956055, 0.008391679763793946, 0.008410112380981445, 0.008444928169250488, 0.008378368377685547, 0.008378368377685547, 0.008418304443359375, 0.008354816436767578, 0.008375295639038086, 0.008357888221740722, 0.00838758373260498, 0.008377344131469726, 0.008368127822875977, 0.008352767944335938, 0.008476672172546386, 0.008351743698120117, 0.008393728256225585, 0.008391679763793946, 0.00838758373260498, 0.008386560440063476, 0.008348671913146973, 0.008357888221740722, 0.008377344131469726, 0.008380415916442872, 0.008431615829467774, 0.00838963222503662, 0.008368127822875977, 0.008375295639038086, 0.008427519798278809, 0.008384511947631837, 0.008409088134765624, 0.008368127822875977, 0.008373248100280761, 0.008380415916442872, 0.008410112380981445, 0.00841215991973877, 0.008368127822875977, 0.00839475154876709, 0.008346624374389648, 0.008358912467956543, 0.008368127822875977, 0.008392736434936524, 0.00834659194946289, 0.008372223854064942, 0.008370176315307617, 0.019144704818725586, 0.008285183906555176, 0.008301568031311036, 0.008297472000122071, 0.008359935760498047, 0.008274944305419921, 0.008307711601257324, 0.008367103576660156, 0.008366080284118652, 0.008358912467956543, 0.00842240047454834, 0.008385536193847656, 0.00838758373260498, 0.00840601634979248, 0.00837939167022705, 0.008442879676818848, 0.008378368377685547, 0.008453120231628418, 0.008423423767089844, 0.008364031791687012, 0.008372223854064942, 0.008360960006713868, 0.008361984252929687, 0.008416255950927735, 0.008367103576660156, 0.008446975708007813, 0.008369152069091796, 0.008386560440063476, 0.008380415916442872, 0.00841215991973877, 0.008391679763793946, 0.008382495880126952, 0.008396767616271972, 0.008461312294006347, 0.008341504096984862, 0.00839577579498291, 0.008339455604553223, 0.008428544044494628, 0.00841420841217041, 0.008369152069091796, 0.008384511947631837, 0.008385536193847656, 0.008360960006713868, 0.008407039642333984, 0.008380415916442872, 0.008380415916442872, 0.008415231704711914, 0.008416255950927735, 0.008410112380981445, 0.008382464408874512, 0.00837939167022705, 0.008380415916442872, 0.008348671913146973, 0.008375295639038086, 0.008354816436767578, 0.008349696159362792, 0.008372223854064942, 0.008376319885253907, 0.008358912467956543, 0.008390720367431641, 0.0084202880859375, 0.008373248100280761, 0.008367103576660156, 0.01901568031311035, 0.008373248100280761, 0.008390656471252441, 0.00838144016265869, 0.008388607978820802, 0.008391679763793946, 0.008341504096984862, 0.008375295639038086, 0.008393728256225585, 0.008377344131469726, 0.008378368377685547, 0.008391679763793946, 0.008399871826171875, 0.008378368377685547, 0.008377344131469726, 0.008440832138061523, 0.008464415550231934, 0.008415200233459473, 0.00838963222503662, 0.008441856384277344, 0.00831283187866211, 0.008281087875366211, 0.008299519538879394, 0.008399871826171875, 0.008344575881958008, 0.008308735847473145, 0.008315903663635254, 0.008393759727478028, 0.00837936019897461, 0.008445952415466309, 0.00838144016265869, 0.008416255950927735, 0.008353792190551757, 0.008501248359680176, 0.00838044834136963, 0.008385503768920898, 0.008455167770385743, 0.008374272346496582, 0.008458239555358887, 0.008374336242675782, 0.008362943649291993, 0.008429568290710449, 0.008398847579956055, 0.008451071739196778, 0.00841113567352295, 0.008396800041198731, 0.008398847579956055, 0.00840499210357666, 0.008378368377685547, 0.008432640075683593, 0.00838758373260498, 0.008373248100280761, 0.008384511947631837, 0.008364031791687012, 0.008371199607849121, 0.008437760353088379, 0.00838144016265869, 0.008372223854064942, 0.008464384078979491, 0.008466431617736817, 0.008396800041198731, 0.008367103576660156, 0.008376352310180664, 0.019056608200073242, 0.008368127822875977, 0.00829036808013916, 0.008282048225402832, 0.008903679847717285, 0.008747008323669434, 0.008754176139831543, 0.008772607803344726, 0.00880128002166748, 0.008792063713073731, 0.008755200386047364, 0.008752127647399903, 0.008765439987182617, 0.008760319709777833, 0.008819711685180665, 0.008839167594909669, 0.008796159744262694, 0.0088156156539917, 0.008781824111938476, 0.008764415740966798, 0.008736767768859864, 0.008758272171020508, 0.008878080368041993, 0.008836095809936523, 0.008469504356384277, 0.008434687614440918, 0.008448032379150391, 0.008458208084106446, 0.008463359832763672, 0.0084203519821167, 0.008430591583251953, 0.008380415916442872, 0.00840601634979248, 0.008373248100280761, 0.00839782428741455, 0.008400896072387695, 0.008384511947631837, 0.008421376228332519, 0.008297472000122071, 0.008357888221740722, 0.008393728256225585, 0.008436736106872558, 0.008453120231628418, 0.008436736106872558, 0.008309760093688966, 0.008324095726013184, 0.00831283187866211, 0.008338432312011718, 0.008416255950927735, 0.008377344131469726, 0.00841318416595459, 0.008339455604553223, 0.008368127822875977, 0.008371199607849121, 0.00838963222503662, 0.008340479850769043, 0.008428544044494628, 0.008416255950927735, 0.008391712188720703, 0.008377311706542968, 0.00841932773590088, 0.008351743698120117, 0.008339455604553223, 0.018983936309814452, 0.008537088394165039, 0.00838144016265869, 0.008385536193847656, 0.008365056037902833, 0.008361984252929687, 0.008372223854064942, 0.008356863975524903, 0.008428544044494628, 0.008359935760498047, 0.008473600387573242, 0.00840601634979248, 0.008409088134765624, 0.00837939167022705, 0.008366080284118652, 0.00841932773590088, 0.00838758373260498, 0.008421376228332519, 0.008385536193847656, 0.008396800041198731, 0.00839577579498291, 0.008382464408874512, 0.008368127822875977, 0.008486911773681641, 0.00909004783630371, 0.008928256034851074, 0.008820735931396484, 0.009273344039916993, 0.00955084800720215, 0.008878080368041993, 0.008839167594909669, 0.008904704093933105, 0.008819711685180665, 0.008804351806640624, 0.00866105556488037, 0.008689599990844727, 0.008729599952697753, 0.008804351806640624, 0.008845312118530273, 0.008779775619506837, 0.00878489589691162, 0.008797247886657715, 0.008779711723327636, 0.008755200386047364, 0.008762368202209473, 0.008748031616210938, 0.008745984077453613, 0.008770560264587402, 0.008570879936218261, 0.008571935653686523, 0.008586208343505859, 0.008592415809631347, 0.008590304374694824, 0.008279040336608886, 0.008342528343200683, 0.008373248100280761, 0.008376319885253907, 0.008367103576660156, 0.008452095985412598, 0.008386560440063476, 0.008391679763793946, 0.008555520057678222, 0.008407039642333984, 0.01903206443786621, 0.008431615829467774, 0.008377344131469726, 0.008364031791687012, 0.008393728256225585, 0.00838144016265869, 0.008368127822875977, 0.00840396785736084, 0.008372223854064942, 0.008398847579956055, 0.008367103576660156, 0.008363007545471191, 0.00890675163269043, 0.008798208236694336, 0.00881049633026123, 0.008742912292480469, 0.008690688133239746, 0.008749055862426757, 0.009241600036621094, 0.008890368461608887, 0.008731648445129395, 0.00872652816772461, 0.00878489589691162, 0.008765439987182617, 0.008757247924804687, 0.008767487525939942, 0.008785920143127441, 0.008747008323669434, 0.00871628761291504, 0.008707072257995606, 0.008713215827941894, 0.008707072257995606, 0.008715264320373535, 0.008722432136535644, 0.0087193603515625, 0.00872447967529297, 0.008812543869018554, 0.008739839553833008, 0.008738816261291504, 0.008742912292480469, 0.008621088027954102, 0.008573920249938964, 0.00858726406097412, 0.008612863540649414, 0.008652799606323243, 0.008747008323669434, 0.008820735931396484, 0.008852479934692382, 0.008813568115234375, 0.008816639900207519, 0.008792063713073731, 0.008745984077453613, 0.008806400299072266, 0.008612863540649414, 0.008626175880432128, 0.008578047752380372, 0.00867024040222168, 0.008725472450256348, 0.008785920143127441, 0.00872652816772461, 0.008811519622802735, 0.008788991928100585, 0.008736767768859864, 0.019560447692871095, 0.008747008323669434, 0.008765439987182617, 0.008723456382751465, 0.008744959831237792, 0.008751104354858399, 0.008738816261291504, 0.008755200386047364, 0.008626175880432128, 0.008562687873840333, 0.008629280090332031, 0.008626144409179687, 0.00859340763092041, 0.008607744216918945, 0.008781824111938476, 0.008773632049560547, 0.008804351806640624, 0.008793120384216308, 0.008781791687011719, 0.008775679588317872, 0.008757247924804687, 0.008767487525939942, 0.008754176139831543, 0.008759296417236329, 0.00880128002166748, 0.008851455688476563, 0.008778752326965332, 0.00875113582611084, 0.00874390411376953, 0.008798208236694336, 0.00871833610534668, 0.008750080108642578, 0.008744959831237792, 0.008803327560424805, 0.009347071647644043, 0.008887295722961425, 0.008828927993774414, 0.008772607803344726, 0.008763392448425293, 0.008717311859130859, 0.00873574447631836, 0.008888319969177246, 0.008775679588317872, 0.008775679588317872, 0.008730624198913574, 0.008731648445129395, 0.008722432136535644, 0.008879103660583497, 0.008791040420532227, 0.008720383644104004, 0.008751104354858399, 0.00878489589691162, 0.008730624198913574, 0.008667136192321777, 0.008587295532226563, 0.00857084846496582, 0.008674304008483886, 0.008754176139831543, 0.008848383903503418, 0.008783871650695801, 0.008759296417236329, 0.008629247665405274, 0.008590335845947266, 0.019110912322998046, 0.00840601634979248, 0.008368127822875977, 0.008390656471252441, 0.008335424423217773, 0.008372159957885742, 0.008349696159362792, 0.008385536193847656, 0.008382464408874512, 0.008372223854064942, 0.008396800041198731, 0.00838758373260498, 0.008398880004882812, 0.008592351913452149, 0.008740863800048827, 0.00880025577545166, 0.00877670383453369, 0.008844287872314453, 0.008822784423828126, 0.008754176139831543, 0.00873574447631836, 0.008744959831237792, 0.008759296417236329, 0.008706048011779785, 0.00880128002166748, 0.008806400299072266, 0.00872652816772461, 0.008720447540283203, 0.008734656333923339, 0.008758272171020508, 0.008772607803344726, 0.008750080108642578, 0.008722432136535644, 0.008743935585021973, 0.008720383644104004, 0.008736831665039062, 0.00875820827484131, 0.008804351806640624, 0.008729599952697753, 0.008798208236694336, 0.008756223678588868, 0.008773632049560547, 0.008759296417236329, 0.008766464233398438, 0.008783871650695801, 0.008762368202209473, 0.008769536018371582, 0.008785920143127441, 0.008713215827941894, 0.008749055862426757, 0.008734720230102539, 0.00877670383453369, 0.008744959831237792, 0.008688639640808106, 0.008706048011779785, 0.008751104354858399, 0.008756223678588868, 0.008396800041198731, 0.008299519538879394, 0.008340479850769043, 0.008279040336608886, 0.008295424461364746, 0.008780799865722656, 0.019544063568115236, 0.00871833610534668, 0.008745984077453613, 0.008777728080749512, 0.008742912292480469, 0.008755200386047364, 0.00879308795928955, 0.008753151893615722, 0.008751104354858399, 0.008750080108642578, 0.008763392448425293, 0.008722432136535644, 0.00880128002166748, 0.009272319793701172, 0.009025535583496093, 0.008854528427124024, 0.008788991928100585, 0.008742912292480469, 0.00894976043701172, 0.00873574447631836, 0.008766464233398438, 0.008751104354858399, 0.008723456382751465, 0.008791040420532227, 0.008773632049560547, 0.00873574447631836, 0.008814592361450196, 0.008763392448425293, 0.00880844783782959, 0.008750080108642578, 0.008762368202209473, 0.008757247924804687, 0.008811519622802735, 0.008767487525939942, 0.008767487525939942, 0.008758272171020508, 0.00873574447631836, 0.008734720230102539, 0.00873574447631836, 0.00872652816772461, 0.008778752326965332, 0.008721407890319824, 0.008756223678588868, 0.008769536018371582, 0.008783871650695801, 0.008839167594909669, 0.008782848358154297, 0.008728575706481934, 0.008763392448425293, 0.00873574447631836, 0.008785920143127441, 0.00882585620880127, 0.009454591751098633, 0.008700927734375, 0.00885043239593506, 0.00872447967529297, 0.008728575706481934, 0.008677375793457032, 0.008604672431945801, 0.008560640335083008, 0.008635392189025879, 0.008561663627624512, 0.008544256210327148, 0.01963929557800293, 0.008729599952697753, 0.008764415740966798, 0.008766464233398438, 0.008748031616210938, 0.008744959831237792, 0.008737792015075683, 0.008742912292480469, 0.008851455688476563, 0.008740927696228027, 0.008765376091003418, 0.008762368202209473, 0.008715264320373535, 0.008759360313415527, 0.008715200424194335, 0.008774656295776367, 0.008749055862426757, 0.008766464233398438, 0.008736767768859864, 0.008770560264587402, 0.008722432136535644, 0.008713215827941894, 0.008715264320373535, 0.008733695983886718, 0.008767487525939942, 0.00882585620880127, 0.008752127647399903, 0.0087193603515625, 0.0087326717376709, 0.008738816261291504, 0.008740863800048827, 0.008742912292480469, 0.008792063713073731, 0.008721407890319824, 0.008733695983886718, 0.008747008323669434, 0.008766528129577636, 0.008723391532897948, 0.00871833610534668, 0.008758272171020508, 0.008762368202209473, 0.008734720230102539, 0.008773632049560547, 0.008795136451721192, 0.008794112205505371, 0.008749088287353516, 0.008780768394470215, 0.008753151893615722, 0.008751104354858399, 0.008744959831237792, 0.008753151893615722, 0.008863743782043456, 0.008674304008483886, 0.008571904182434082, 0.008592384338378906, 0.008765439987182617, 0.008820735931396484, 0.00879308795928955, 0.008788991928100585, 0.008760319709777833, 0.008766464233398438, 0.00879923152923584, 0.008769536018371582, 0.019553279876708983, 0.008734720230102539, 0.008714240074157715, 0.008754176139831543, 0.00871833610534668, 0.008748031616210938, 0.008741888046264648, 0.008740863800048827, 0.008730624198913574, 0.008744959831237792, 0.00873574447631836, 0.008707072257995606, 0.008719391822814941, 0.008712160110473632, 0.008744959831237792, 0.00871628761291504, 0.008743935585021973, 0.008742912292480469, 0.00871833610534668, 0.008756223678588868, 0.008737792015075683, 0.00882688045501709, 0.008705023765563966, 0.008759296417236329, 0.008771583557128907, 0.008747008323669434, 0.008737792015075683, 0.008816639900207519, 0.008796159744262694, 0.008788991928100585, 0.008796159744262694, 0.008773632049560547, 0.008795136451721192, 0.008803327560424805, 0.008729599952697753, 0.008697855949401855, 0.00871014404296875, 0.008752127647399903, 0.00869478416442871, 0.008722432136535644, 0.008697855949401855, 0.008730624198913574, 0.008812543869018554, 0.008778752326965332, 0.008714240074157715, 0.008747008323669434, 0.008752127647399903, 0.00873574447631836, 0.00881049633026123, 0.008769536018371582, 0.008752127647399903, 0.008737792015075683, 0.008820735931396484, 0.008752127647399903, 0.008745984077453613, 0.0084203519821167, 0.008409088134765624, 0.00828825569152832, 0.008403008460998534, 0.00830457592010498, 0.008354816436767578, 0.008367103576660156, 0.008300543785095215, 0.018953216552734374, 0.008355839729309082, 0.008357888221740722, 0.008382464408874512, 0.008367103576660156, 0.00838963222503662, 0.00838963222503662, 0.008374272346496582, 0.00838144016265869, 0.008347647666931152, 0.008350720405578613, 0.008366080284118652, 0.008396800041198731, 0.008377344131469726, 0.008401920318603515, 0.00834563159942627, 0.008266719818115234, 0.008273920059204102, 0.008285183906555176, 0.00828006362915039, 0.008325119972229005, 0.008386560440063476, 0.008384511947631837, 0.00841427230834961, 0.008424384117126465, 0.00840294361114502, 0.008423423767089844, 0.008357888221740722, 0.00840294361114502, 0.008377344131469726, 0.00840396785736084, 0.008371199607849121, 0.008377344131469726, 0.008376319885253907, 0.008481792449951172, 0.008378368377685547, 0.00841113567352295, 0.008365056037902833, 0.008348735809326172, 0.008361920356750489, 0.008319999694824219, 0.008357888221740722, 0.008386560440063476, 0.008382464408874512, 0.00840294361114502, 0.008366080284118652, 0.008359935760498047, 0.008355839729309082, 0.008343551635742187, 0.008369152069091796, 0.00841420841217041, 0.008341504096984862, 0.008380415916442872, 0.008368127822875977, 0.008376319885253907, 0.008357888221740722, 0.008380415916442872, 0.008349696159362792, 0.008391679763793946, 0.008343551635742187, 0.008313856124877929, 0.008353856086730957, 0.00850937557220459, 0.018929664611816405, 0.008374272346496582, 0.00840396785736084, 0.008377344131469726, 0.008350720405578613, 0.008363007545471191, 0.008352767944335938, 0.008372223854064942, 0.008384511947631837, 0.008398847579956055, 0.00839475154876709, 0.008371199607849121, 0.008410112380981445, 0.008388607978820802, 0.008382464408874512, 0.00841113567352295, 0.0084203519821167, 0.008386560440063476, 0.008397855758666993, 0.008434656143188477, 0.008371199607849121, 0.008384511947631837, 0.00839475154876709, 0.008550399780273438, 0.008671232223510742, 0.008820735931396484, 0.008500224113464355, 0.008493056297302246, 0.008434687614440918, 0.00848588752746582, 0.008461312294006347, 0.008467455863952637, 0.008408063888549805, 0.00840601634979248, 0.008386560440063476, 0.00838758373260498, 0.008371199607849121, 0.008384511947631837, 0.008357888221740722, 0.008368127822875977, 0.008398847579956055, 0.008462335586547852, 0.00840601634979248, 0.008386560440063476, 0.008408127784729004, 0.008380352020263672, 0.008359935760498047, 0.00840396785736084, 0.00832102394104004, 0.008358912467956543, 0.008364031791687012, 0.008369152069091796, 0.008373248100280761, 0.008332287788391114, 0.008356863975524903, 0.008361984252929687, 0.008372223854064942, 0.008363007545471191, 0.008352767944335938, 0.008355839729309082, 0.00839577579498291, 0.008375295639038086, 0.008357888221740722, 0.019066879272460938, 0.00838758373260498, 0.008361984252929687, 0.008368127822875977, 0.008374272346496582, 0.00840499210357666, 0.008388640403747559, 0.008392671585083007, 0.008361984252929687, 0.008390687942504882, 0.00837936019897461, 0.008368127822875977, 0.008352767944335938, 0.008427519798278809, 0.008358912467956543, 0.008368127822875977, 0.008365056037902833, 0.008348671913146973, 0.00841420841217041, 0.008343551635742187, 0.008372223854064942, 0.008327168464660644, 0.00837939167022705, 0.008385536193847656, 0.00839475154876709, 0.008360960006713868, 0.00841113567352295, 0.008346624374389648, 0.008380415916442872, 0.008373248100280761, 0.008350720405578613, 0.008353792190551757, 0.008384511947631837, 0.00849407958984375, 0.008427519798278809, 0.008447999954223634, 0.008421376228332519, 0.008439807891845704, 0.008430591583251953, 0.008445952415466309, 0.00850534439086914, 0.008429568290710449, 0.00838144016265869, 0.008370176315307617, 0.00838963222503662, 0.008384511947631837, 0.008410112380981445, 0.008348671913146973, 0.008399871826171875, 0.008367103576660156, 0.008398847579956055, 0.008324095726013184, 0.008356863975524903, 0.008380415916442872, 0.008371199607849121, 0.008346624374389648, 0.008322048187255859, 0.008346624374389648, 0.008399871826171875, 0.008349696159362792, 0.008369152069091796, 0.008396800041198731, 0.008345600128173827, 0.018979839324951172, 0.008372223854064942, 0.008396800041198731, 0.008341504096984862, 0.008391679763793946, 0.008376319885253907, 0.008492032051086425, 0.008365056037902833, 0.008386560440063476, 0.008374272346496582, 0.008455167770385743, 0.008402976036071777, 0.008395744323730468, 0.008356863975524903, 0.008350720405578613, 0.008372223854064942, 0.008388607978820802, 0.008408063888549805, 0.008388607978820802, 0.008386560440063476, 0.008365056037902833, 0.008366144180297852, 0.008355775833129882, 0.00837939167022705, 0.008416255950927735, 0.008396800041198731, 0.008435711860656739, 0.008351743698120117, 0.008465408325195312, 0.008285183906555176, 0.008293375968933106, 0.008302592277526855, 0.00828825569152832, 0.008315903663635254, 0.00838963222503662, 0.008350720405578613, 0.00841113567352295, 0.008348671913146973, 0.008385536193847656, 0.008368127822875977, 0.008374272346496582, 0.008366080284118652, 0.008385536193847656, 0.00838144016265869, 0.00840396785736084, 0.008343551635742187, 0.008386560440063476, 0.008363007545471191, 0.008348671913146973, 0.008356863975524903, 0.008359935760498047, 0.008367103576660156, 0.008358912467956543, 0.008349696159362792, 0.008390656471252441, 0.008492032051086425, 0.008363007545471191, 0.008375295639038086, 0.008396800041198731, 0.008378368377685547, 0.008421376228332519, 0.008357888221740722, 0.008384511947631837, 0.018932735443115235, 0.008388607978820802, 0.008382464408874512, 0.008343551635742187, 0.00838144016265869, 0.008365056037902833, 0.008410112380981445, 0.008310784339904785, 0.008364031791687012, 0.008438783645629883, 0.00840601634979248, 0.008355839729309082, 0.008343551635742187, 0.008353792190551757, 0.008416255950927735, 0.008357888221740722, 0.008393728256225585, 0.008371199607849121, 0.008363007545471191, 0.008358912467956543, 0.008341504096984862, 0.008330240249633789, 0.008343551635742187, 0.008357888221740722, 0.008348671913146973, 0.008475647926330567, 0.008388607978820802, 0.008345600128173827, 0.008373248100280761, 0.008345600128173827, 0.008347647666931152, 0.008360960006713868, 0.008353792190551757, 0.008365056037902833, 0.008347647666931152, 0.008348671913146973, 0.008373248100280761, 0.008367103576660156, 0.008350751876831054, 0.008400863647460937, 0.008343551635742187, 0.008358912467956543, 0.008370176315307617, 0.00839475154876709, 0.008368127822875977, 0.008340479850769043, 0.008398847579956055, 0.008377344131469726, 0.008338432312011718, 0.008514559745788575, 0.00840396785736084, 0.008371199607849121, 0.008347647666931152, 0.008349696159362792, 0.008376319885253907, 0.008415231704711914, 0.00838758373260498, 0.008330240249633789, 0.008360960006713868, 0.008351743698120117, 0.008383487701416016, 0.008326144218444824, 0.008347647666931152, 0.01889587211608887, 0.0083374080657959, 0.008363039970397949, 0.0083721923828125, 0.008349760055541992, 0.008354751586914063, 0.008396800041198731, 0.008426527976989746, 0.008357855796813965, 0.008371199607849121, 0.008387647628784179, 0.008368063926696777, 0.008371199607849121, 0.008359935760498047, 0.008378368377685547, 0.00839782428741455, 0.008370176315307617, 0.008353792190551757, 0.008361984252929687, 0.008350720405578613, 0.008336383819580078, 0.00840294361114502, 0.008373248100280761, 0.008302592277526855, 0.00828825569152832, 0.008364031791687012, 0.008417311668395997, 0.008330240249633789, 0.008353759765625, 0.008360960006713868, 0.008392704010009766, 0.00831494426727295, 0.008345536231994629, 0.008332287788391114, 0.008311807632446289, 0.00841318416595459, 0.00832102394104004, 0.008390656471252441, 0.008340479850769043, 0.008334336280822753, 0.008346624374389648, 0.008325119972229005, 0.008377344131469726, 0.008342528343200683, 0.008342528343200683, 0.008335359573364258, 0.008390656471252441, 0.00832307243347168, 0.008328191757202149, 0.008364031791687012, 0.008393728256225585, 0.008388607978820802, 0.008359935760498047, 0.008359935760498047, 0.008370176315307617, 0.00838144016265869, 0.00841318416595459, 0.008393792152404786, 0.008323007583618165, 0.008384511947631837, 0.008374272346496582, 0.008449024200439453, 0.008339455604553223, 0.01887027168273926, 0.008372223854064942, 0.009064448356628419, 0.00913920021057129, 0.008843263626098634, 0.0088156156539917, 0.008779775619506837, 0.008791040420532227, 0.00880128002166748, 0.00882688045501709, 0.008762368202209473, 0.00877670383453369, 0.008759296417236329, 0.00880742359161377, 0.00878489589691162, 0.008824831962585449, 0.008889408111572265, 0.008783807754516601, 0.008737792015075683, 0.00880128002166748, 0.008885248184204102, 0.00889241600036621, 0.008729599952697753, 0.008768511772155761, 0.008752127647399903, 0.008737792015075683, 0.00871116828918457, 0.008757247924804687, 0.008747008323669434, 0.008923135757446288, 0.008760319709777833, 0.008851455688476563, 0.008721440315246583, 0.00875823974609375, 0.008743935585021973, 0.008766464233398438, 0.008750080108642578, 0.008737792015075683, 0.0087326717376709, 0.00884224033355713, 0.0087326717376709, 0.008749055862426757, 0.008765439987182617, 0.008769536018371582, 0.008761343955993652, 0.00877670383453369, 0.008781824111938476, 0.008754176139831543, 0.008730624198913574, 0.00880742359161377, 0.008819711685180665, 0.008750111579895019, 0.008725536346435548, 0.008697792053222657, 0.008707072257995606, 0.008736767768859864, 0.008765439987182617, 0.008869888305664063, 0.008722432136535644, 0.00872652816772461, 0.008677375793457032, 0.008853504180908203, 0.008720383644104004]",tokens/s,115.04729319113002,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,2,2,cuda,0,42,,,True,True,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/2/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669324bb-388cbd4e505a7569300918fb;bcbd8348-3562-4370-b945-9eb973923a16) Repository Not Found for url: https://huggingface.co/2/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: 2 is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/xglm-564M,,cuda,0,42,,,,,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,d35829e539df8480b726c647eeabf91e41eae047,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,,MB,1298.6368,4031.250432,0.0,3391.094784,2592.661504,s,11,0.28360147285461423,0.025781952077692207,0.00035487888937498554,0.02561180877685547,0.02600396728515625,0.02639412784576416,0.02670625629425049,"[0.02678428840637207, 0.02592259216308594, 0.02557980728149414, 0.025543327331542968, 0.025554912567138672, 0.0255830078125, 0.02551251220703125, 0.025654176712036132, 0.025851072311401366, 0.02600396728515625, 0.02561180877685547]",tokens/s,9929.426570515721,kWh,3.0907990536829473e-07,1.6935861220584613e-07,2.0540551680470518e-06,2.5324936856211925e-06,tokens/kWh,101086135.55622984,MB,1298.6368,4031.250432,0.0,3391.094784,2642.977792,s,11,10.472120300292968,0.9520109363902698,0.013099083517054742,0.9473740234375,0.963303955078125,0.9760693969726563,0.9862817504882813,"[0.9888348388671875, 0.9539652709960937, 0.9455442504882813, 0.9513214111328125, 0.9479583740234375, 0.963303955078125, 0.9473740234375, 0.9443287353515625, 0.9465650024414063, 0.9435420532226563, 0.9393823852539063]",tokens/s,66.17571037458505,kWh,1.1270201758285531e-05,6.173530932770213e-06,2.8999479775767524e-05,4.644321246682328e-05,tokens/kWh,1356495.3123129471,,s,692,10.712844279289246,0.015480988842903533,0.003124733432381309,0.014949376106262208,0.015536025619506835,0.015922022390365602,0.04091054950714111,"[0.0160184326171875, 0.01581056022644043, 0.0158023681640625, 0.015582207679748536, 0.015929344177246094, 0.015978495597839357, 0.01562828826904297, 0.015900671958923338, 0.015797247886657716, 0.0157521915435791, 0.01640652847290039, 0.015970303535461427, 0.016157695770263672, 0.016227327346801757, 0.01586176013946533, 0.01635327911376953, 0.016343040466308592, 0.016292896270751953, 0.016146400451660155, 0.015844351768493654, 0.015820799827575685, 0.015730688095092774, 0.015756287574768067, 0.01597439956665039, 0.01601740837097168, 0.015890432357788087, 0.016269311904907227, 0.01619046401977539, 0.016082944869995116, 0.01591603183746338, 0.01618943977355957, 0.016149503707885742, 0.016356351852416993, 0.015933440208435058, 0.015552512168884277, 0.015534079551696778, 0.015545344352722168, 0.015362048149108886, 0.015499263763427735, 0.01488588809967041, 0.014892031669616699, 0.014939135551452636, 0.014955519676208496, 0.015471615791320802, 0.015396863937377929, 0.015431679725646973, 0.01538150405883789, 0.015413248062133789, 0.015411199569702149, 0.015443967819213868, 0.015065119743347169, 0.01503228759765625, 0.015508480072021484, 0.015417344093322754, 0.015425536155700683, 0.015463423728942872, 0.015453184127807617, 0.015147007942199708, 0.01546444797515869, 0.01498624038696289, 0.014948351860046387, 0.015285280227661133, 0.041831390380859375, 0.014970879554748535, 0.01496678352355957, 0.014939135551452636, 0.014946304321289062, 0.015589376449584962, 0.01601433563232422, 0.015641599655151366, 0.015416319847106934, 0.014947327613830566, 0.01516748809814453, 0.015494175910949707, 0.01544598388671875, 0.015351840019226074, 0.015407072067260743, 0.015448063850402831, 0.01538150405883789, 0.015351807594299317, 0.015448063850402831, 0.015385600090026855, 0.015520768165588379, 0.015226880073547363, 0.01540608024597168, 0.015428607940673827, 0.014952447891235352, 0.014971903800964356, 0.014947423934936524, 0.014968735694885254, 0.014947392463684083, 0.01490937614440918, 0.014919679641723632, 0.014898176193237305, 0.014924799919128418, 0.014918656349182128, 0.01490227222442627, 0.014856191635131836, 0.014910464286804198, 0.014929920196533204, 0.014922752380371093, 0.015003647804260254, 0.014927871704101562, 0.014905344009399414, 0.014918656349182128, 0.014914560317993163, 0.014927871704101562, 0.0149616641998291, 0.015049728393554687, 0.014942208290100097, 0.01490329647064209, 0.014905344009399414, 0.014936063766479492, 0.014926848411560058, 0.014947327613830566, 0.014952447891235352, 0.014913536071777344, 0.014882816314697265, 0.01622425651550293, 0.01582703971862793, 0.01556879997253418, 0.014924799919128418, 0.0149749755859375, 0.01496678352355957, 0.014919679641723632, 0.040970241546630856, 0.014889984130859376, 0.014948351860046387, 0.014908415794372559, 0.01487667179107666, 0.01489408016204834, 0.014970879554748535, 0.015441920280456543, 0.015380479812622071, 0.015418368339538574, 0.014913536071777344, 0.015001600265502929, 0.01549721622467041, 0.015362048149108886, 0.015431679725646973, 0.014965760231018066, 0.014941184043884278, 0.0148602876663208, 0.014973952293395995, 0.014898176193237305, 0.01489408016204834, 0.0149268798828125, 0.014881759643554688, 0.014935040473937988, 0.014927871704101562, 0.01492073631286621, 0.014906335830688476, 0.014945280075073243, 0.014873600006103516, 0.014941184043884278, 0.015028223991394044, 0.014945280075073243, 0.014951423645019531, 0.014948351860046387, 0.014897151947021485, 0.014911487579345703, 0.014927871704101562, 0.014923775672912597, 0.014887935638427734, 0.014905344009399414, 0.014979071617126465, 0.01489305591583252, 0.014930944442749023, 0.014857215881347656, 0.014915583610534668, 0.014899200439453125, 0.014865407943725586, 0.014980095863342285, 0.01480191993713379, 0.015287296295166015, 0.01547878360748291, 0.015399935722351075, 0.015364095687866211, 0.015329279899597169, 0.014888959884643555, 0.014888959884643555, 0.014874688148498535, 0.015035327911376952, 0.014929920196533204, 0.014862336158752442, 0.014888959884643555, 0.014880767822265625, 0.014896160125732421, 0.040896480560302734, 0.014889984130859376, 0.0148787202835083, 0.014845952033996582, 0.014869503974914551, 0.014912511825561523, 0.014864383697509765, 0.014869503974914551, 0.014832639694213867, 0.014914560317993163, 0.014836735725402832, 0.0149749755859375, 0.014836735725402832, 0.01490329647064209, 0.015380479812622071, 0.015900671958923338, 0.01542246437072754, 0.015510527610778809, 0.015336447715759278, 0.015408127784729005, 0.01537945556640625, 0.015431679725646973, 0.015580160140991211, 0.015512576103210449, 0.015391743659973145, 0.014944255828857422, 0.014881792068481446, 0.014955519676208496, 0.014899264335632325, 0.015049663543701172, 0.01490124797821045, 0.014941184043884278, 0.014923775672912597, 0.015024127960205079, 0.014925824165344239, 0.01496780776977539, 0.014887935638427734, 0.014960639953613282, 0.01519206428527832, 0.014987263679504394, 0.014945280075073243, 0.015044608116149903, 0.014951423645019531, 0.01566412830352783, 0.015524864196777344, 0.015456319808959961, 0.015360960006713867, 0.01537331199645996, 0.015322112083435058, 0.01549830436706543, 0.015356863975524902, 0.015042559623718262, 0.014931967735290527, 0.014942208290100097, 0.014874624252319337, 0.014955519676208496, 0.014882816314697265, 0.01496678352355957, 0.014971903800964356, 0.014960639953613282, 0.014918656349182128, 0.015012864112854005, 0.014924799919128418, 0.040932350158691407, 0.014906368255615234, 0.014945280075073243, 0.014916607856750488, 0.014899200439453125, 0.014898176193237305, 0.014814208030700684, 0.014941184043884278, 0.014823488235473633, 0.014946240425109864, 0.015453184127807617, 0.015503359794616698, 0.015353856086730957, 0.015461376190185547, 0.015461376190185547, 0.01530470371246338, 0.015396863937377929, 0.015473664283752442, 0.01538150405883789, 0.015466496467590332, 0.01547878360748291, 0.015509504318237305, 0.014882816314697265, 0.014948351860046387, 0.014888959884643555, 0.014908415794372559, 0.014932991981506348, 0.014926848411560058, 0.014865407943725586, 0.014918656349182128, 0.014866432189941407, 0.015003647804260254, 0.01488486385345459, 0.01499135971069336, 0.01494431972503662, 0.014961600303649902, 0.01488486385345459, 0.01491763210296631, 0.014879743576049804, 0.014914560317993163, 0.014925824165344239, 0.01492073631286621, 0.014845919609069825, 0.014910464286804198, 0.014963711738586426, 0.014982144355773925, 0.014941184043884278, 0.01490124797821045, 0.01496780776977539, 0.014881792068481446, 0.014871552467346191, 0.015033344268798828, 0.014889984130859376, 0.014954496383666992, 0.0148787202835083, 0.014945280075073243, 0.0148787202835083, 0.014987296104431152, 0.015433695793151856, 0.015419391632080079, 0.015426560401916504, 0.014929920196533204, 0.014870528221130372, 0.0408996467590332, 0.014974911689758301, 0.015799296379089355, 0.016542720794677734, 0.015876095771789552, 0.015491071701049805, 0.01561190414428711, 0.01548697566986084, 0.015475711822509765, 0.015452192306518555, 0.015387616157531737, 0.015511551856994628, 0.015458304405212403, 0.015427583694458008, 0.015388671875, 0.015599616050720215, 0.015432703971862792, 0.0157327356338501, 0.015482879638671876, 0.015446016311645508, 0.015935487747192383, 0.015737855911254883, 0.015673343658447265, 0.015622143745422363, 0.015446016311645508, 0.015543295860290527, 0.015456255912780761, 0.015385600090026855, 0.01548902416229248, 0.015494144439697266, 0.015461376190185547, 0.015535103797912597, 0.015706111907958984, 0.014913536071777344, 0.015161343574523926, 0.015469568252563477, 0.015439871788024903, 0.015364095687866211, 0.015019007682800293, 0.014982144355773925, 0.014977024078369141, 0.015068160057067872, 0.01495961570739746, 0.014916607856750488, 0.014943231582641601, 0.0148787202835083, 0.014892031669616699, 0.014874624252319337, 0.01491763210296631, 0.014928895950317383, 0.014940159797668457, 0.0149749755859375, 0.01490227222442627, 0.014968832015991211, 0.014923775672912597, 0.014956543922424317, 0.014912511825561523, 0.014899200439453125, 0.014951423645019531, 0.014970879554748535, 0.014943231582641601, 0.01495961570739746, 0.014921728134155274, 0.040908798217773434, 0.014915583610534668, 0.014943231582641601, 0.01477836799621582, 0.014763008117675782, 0.01480191993713379, 0.014854144096374512, 0.014936063766479492, 0.014908415794372559, 0.01491763210296631, 0.014944255828857422, 0.014930944442749023, 0.014955519676208496, 0.014972928047180176, 0.014936063766479492, 0.014938112258911132, 0.014960639953613282, 0.01491763210296631, 0.014945280075073243, 0.014896127700805664, 0.01500876808166504, 0.014879743576049804, 0.014949376106262208, 0.014916607856750488, 0.014946304321289062, 0.014910464286804198, 0.01496780776977539, 0.014880767822265625, 0.014996479988098145, 0.014897151947021485, 0.014942208290100097, 0.014921728134155274, 0.014975040435791016, 0.01489401626586914, 0.014964768409729004, 0.014881759643554688, 0.01499238395690918, 0.014955519676208496, 0.014957568168640138, 0.014962688446044922, 0.01499135971069336, 0.014920703887939453, 0.014977024078369141, 0.01490944004058838, 0.015168512344360351, 0.01537331199645996, 0.01542246437072754, 0.015362048149108886, 0.015421440124511718, 0.015460415840148926, 0.015432640075683594, 0.015370240211486816, 0.015401984214782715, 0.015359999656677246, 0.015435775756835938, 0.01540403175354004, 0.015484928131103515, 0.01537331199645996, 0.014972928047180176, 0.014899200439453125, 0.015005696296691894, 0.014908415794372559, 0.014958592414855957, 0.04103680038452148, 0.014954527854919434, 0.014946271896362304, 0.014950400352478027, 0.014922752380371093, 0.014906368255615234, 0.014830592155456544, 0.01488486385345459, 0.014882816314697265, 0.014919679641723632, 0.014877696037292481, 0.014939135551452636, 0.014979071617126465, 0.014932991981506348, 0.014919679641723632, 0.014954496383666992, 0.014931967735290527, 0.014922752380371093, 0.014953472137451173, 0.014965760231018066, 0.014939135551452636, 0.014954496383666992, 0.014941184043884278, 0.014935040473937988, 0.014934016227722167, 0.014939135551452636, 0.014935040473937988, 0.014948351860046387, 0.01490944004058838, 0.01497599983215332, 0.014921728134155274, 0.014914560317993163, 0.014907391548156738, 0.015002623558044433, 0.014921728134155274, 0.014896127700805664, 0.01489408016204834, 0.014951423645019531, 0.014919679641723632, 0.014982144355773925, 0.01490329647064209, 0.014923775672912597, 0.01495961570739746, 0.014899200439453125, 0.014925824165344239, 0.015014911651611328, 0.014960639953613282, 0.01490944004058838, 0.01496678352355957, 0.014981120109558106, 0.014982144355773925, 0.014952447891235352, 0.014988287925720215, 0.014949376106262208, 0.014970879554748535, 0.0149616641998291, 0.014938112258911132, 0.014952447891235352, 0.014981120109558106, 0.015536128044128418, 0.015246335983276366, 0.016184320449829103, 0.015575039863586425, 0.04124671936035156, 0.014911487579345703, 0.014989312171936036, 0.014985216140747071, 0.014898176193237305, 0.014891008377075195, 0.01488486385345459, 0.014907391548156738, 0.014936063766479492, 0.014957568168640138, 0.015211520195007324, 0.01540608024597168, 0.01537331199645996, 0.015309823989868163, 0.015416319847106934, 0.015432703971862792, 0.01539583969116211, 0.015427583694458008, 0.014948351860046387, 0.014922752380371093, 0.014919679641723632, 0.014923775672912597, 0.015070207595825195, 0.015481856346130371, 0.015430656433105469, 0.015378432273864746, 0.015324159622192383, 0.01497599983215332, 0.014897151947021485, 0.01491763210296631, 0.014908415794372559, 0.014907391548156738, 0.014932991981506348, 0.01495043182373047, 0.014892000198364258, 0.014936063766479492, 0.014892095565795899, 0.01490835189819336, 0.014927871704101562, 0.014945280075073243, 0.014928895950317383, 0.014871552467346191, 0.015254528045654296, 0.015094783782958985, 0.014950400352478027, 0.014956543922424317, 0.014929920196533204, 0.01488486385345459, 0.014866432189941407, 0.014904319763183594, 0.014883839607238769, 0.014926848411560058, 0.014922752380371093, 0.014949376106262208, 0.014872575759887695, 0.014864447593688965, 0.014913472175598144, 0.014914560317993163, 0.014858240127563477, 0.01489305591583252, 0.014922752380371093, 0.01490124797821045, 0.015011839866638184, 0.04181708908081055, 0.015363072395324707, 0.015396863937377929, 0.01548799991607666, 0.015387680053710938, 0.01493609619140625, 0.014859199523925781, 0.014949376106262208, 0.014953472137451173, 0.01498624038696289, 0.01489510440826416, 0.014922752380371093, 0.014839808464050292, 0.014931967735290527, 0.014993408203125, 0.01496678352355957, 0.014859264373779296, 0.014942208290100097, 0.014858240127563477, 0.01496780776977539, 0.014883839607238769, 0.014981120109558106, 0.014879743576049804, 0.01507737636566162, 0.014944255828857422, 0.014999551773071289, 0.014888959884643555, 0.014977024078369141, 0.01487667179107666, 0.014977055549621582, 0.014886879920959472, 0.014965760231018066, 0.014913536071777344, 0.014943231582641601, 0.01489408016204834, 0.0149616641998291, 0.01489305591583252, 0.014965760231018066, 0.015179776191711425, 0.014989312171936036, 0.014900223731994629, 0.014949376106262208, 0.014897151947021485, 0.014916607856750488, 0.01495961570739746, 0.014950400352478027, 0.01487564754486084, 0.01492790412902832, 0.014885855674743651, 0.015009792327880859, 0.014934016227722167, 0.014924799919128418, 0.014881792068481446, 0.01496780776977539, 0.014891008377075195, 0.01496070384979248, 0.014905344009399414, 0.014932928085327148, 0.0148603515625, 0.014996416091918945, 0.014861311912536621, 0.014899200439453125, 0.014919679641723632, 0.04092825698852539, 0.014874688148498535, 0.014879679679870606, 0.014939135551452636, 0.014899200439453125, 0.015017984390258789, 0.01488486385345459, 0.014898176193237305, 0.014841856002807617, 0.014915583610534668, 0.01489510440826416, 0.01490124797821045, 0.014888959884643555, 0.014929920196533204, 0.014892031669616699, 0.014941184043884278, 0.014870528221130372, 0.014908415794372559, 0.014828543663024902, 0.014979071617126465, 0.014851072311401366, 0.014963711738586426, 0.014879743576049804, 0.014918656349182128, 0.01487667179107666, 0.014937088012695313, 0.014867456436157226, 0.015007743835449219, 0.014899200439453125, 0.014898176193237305, 0.014863360404968261, 0.014907391548156738, 0.014889984130859376, 0.014958592414855957, 0.014866432189941407, 0.01490944004058838, 0.014951423645019531, 0.014910464286804198, 0.01489305591583252, 0.014950400352478027, 0.014915583610534668, 0.014905344009399414, 0.014849023818969726, 0.014932991981506348, 0.014898176193237305, 0.014944255828857422, 0.014873600006103516, 0.014960639953613282, 0.014857215881347656, 0.01477017593383789, 0.01470464038848877, 0.01479372787475586, 0.014821375846862793, 0.014937088012695313, 0.014877696037292481, 0.014925824165344239, 0.0149616641998291, 0.0151910400390625, 0.01493404769897461, 0.014923744201660157, 0.0148787202835083, 0.01498316764831543, 0.014870528221130372]",tokens/s,64.5953569340888,,,main,False,False,,,,,,,,,,,,,,,,,,,,,,, float32-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-14B,Qwen/Qwen-14B,cuda,0,42,,,True,,,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",qwen,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,huggyllama/llama-7b,huggyllama/llama-7b,cuda,0,42,,,True,,,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",llama,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-180B,tiiuae/falcon-180B,cuda,0,42,,,True,True,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 403 Client Error: Forbidden for url: https://huggingface.co/tiiuae/falcon-180B/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-669326f3-50d1d5e74790930f2557b730;23705765-6a1d-41ba-94d2-9aaa7a3a4305) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/tiiuae/falcon-180B/resolve/main/config.json. If you are trying to create or update content,make sure you have a token with the `write` role. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1826, in _raise_on_head_call_error raise LocalEntryNotFoundError( huggingface_hub.utils._errors.LocalEntryNotFoundError: An error happened while trying to locate the file on the Hub and we cannot find the requested files in the local cache. Please check your connection and try again or make sure your Internet connection is on. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 445, in cached_file raise EnvironmentError( OSError: We couldn't connect to 'https://huggingface.co' to load this file, couldn't find it in the cached files and it looks like tiiuae/falcon-180B is not the path to a directory containing a file named config.json. Checkout your internet connection or see how to run the library in offline mode at 'https://huggingface.co/docs/transformers/installation#offline-mode'. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/gpt-neo-1.3B,EleutherAI/gpt-neo-1.3B,cuda,0,42,,,True,True,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1099.452416,6442.975232,0.0,5802.819584,5525.733376,s,10,0.5362704963684082,0.05362704963684082,0.0010259161074046576,0.05331676864624023,0.05393291015624999,0.055296790695190425,0.05638789512634277,"[0.056660671234130856, 0.05333660888671875, 0.05334508895874023, 0.0532325439453125, 0.053267326354980465, 0.05362982559204101, 0.05307635116577149, 0.05345491027832031, 0.05297024154663086, 0.053296928405761716]",tokens/s,4773.710314731404,kWh,6.428221690553974e-07,3.5223568752272786e-07,4.487571524836959e-06,5.482629381415084e-06,tokens/kWh,46692924.54233439,MB,1099.452416,6442.975232,0.0,5802.819584,5626.041856,s,10,10.849396728515622,1.0849396728515623,0.005517432657076276,1.0839310913085938,1.0908426147460937,1.0937900817871093,1.0961480554199219,"[1.096737548828125, 1.0848692626953125, 1.088586669921875, 1.0873590087890626, 1.0901876220703124, 1.082992919921875, 1.0795074462890626, 1.078352294921875, 1.0798287353515625, 1.0809752197265625]",tokens/s,58.0677447571036,kWh,1.2833512459076256e-05,7.03239351157307e-06,4.624558017276269e-05,6.611148614341202e-05,tokens/kWh,952935.7706970549,,s,629,11.310397439956658,0.01798155395859565,0.0064204726478463555,0.017150976181030272,0.017442610931396485,0.017541548538208006,0.07105478393554689,"[0.017547264099121093, 0.017250303268432618, 0.017135616302490234, 0.01720012855529785, 0.017447999954223633, 0.017585088729858398, 0.01740492820739746, 0.01779814338684082, 0.017123327255249024, 0.017117183685302736, 0.017142784118652343, 0.017183744430541992, 0.017316864013671874, 0.017438720703125, 0.01738956832885742, 0.01741107177734375, 0.017554431915283202, 0.01744179153442383, 0.01746227264404297, 0.017398784637451172, 0.01738751983642578, 0.01739673614501953, 0.017415168762207032, 0.017500160217285156, 0.017539072036743163, 0.01741209602355957, 0.01742336082458496, 0.017450048446655275, 0.017486783981323244, 0.017491968154907226, 0.018719743728637696, 0.01804287910461426, 0.017523712158203125, 0.017382400512695313, 0.017257471084594727, 0.01721548843383789, 0.017254400253295898, 0.017231903076171874, 0.017789920806884765, 0.01741619110107422, 0.017327104568481445, 0.01738444709777832, 0.01744691276550293, 0.017356800079345702, 0.017313791275024415, 0.017370111465454103, 0.017340415954589843, 0.01725542449951172, 0.01717452812194824, 0.01716633605957031, 0.017143808364868163, 0.017099775314331055, 0.01722368049621582, 0.017155071258544922, 0.017126399993896483, 0.0172410888671875, 0.01724825668334961, 0.01722777557373047, 0.017257471084594727, 0.017533952713012696, 0.01741004753112793, 0.017342464447021484, 0.07163187408447266, 0.01745510482788086, 0.01741007995605469, 0.01727280044555664, 0.017085439682006837, 0.017152000427246093, 0.01718988800048828, 0.0174202880859375, 0.017282047271728516, 0.017509376525878906, 0.01742646408081055, 0.017318880081176758, 0.017228799819946287, 0.017337343215942384, 0.017298431396484376, 0.017291263580322267, 0.017294336318969726, 0.017153024673461914, 0.017708032608032227, 0.01739366340637207, 0.01724415969848633, 0.017137664794921875, 0.01725644874572754, 0.017318912506103516, 0.017022975921630858, 0.017286144256591796, 0.01721343994140625, 0.017290239334106446, 0.017124351501464845, 0.017110015869140623, 0.017124351501464845, 0.017261568069458007, 0.017081344604492187, 0.017099775314331055, 0.01706188774108887, 0.017079296112060546, 0.017075199127197266, 0.017105920791625977, 0.017063936233520507, 0.017140735626220704, 0.017124351501464845, 0.01721958351135254, 0.017080320358276366, 0.017154048919677735, 0.017148927688598634, 0.01719808006286621, 0.017185792922973633, 0.017138687133789063, 0.017104896545410156, 0.017129472732543945, 0.017154048919677735, 0.017122304916381836, 0.017125375747680666, 0.017163263320922852, 0.017121280670166016, 0.017121280670166016, 0.017133567810058595, 0.01720012855529785, 0.017140735626220704, 0.017157119750976564, 0.017142784118652343, 0.017120256423950195, 0.017113151550292968, 0.0713061752319336, 0.01719500732421875, 0.017472511291503907, 0.017566719055175782, 0.0172728328704834, 0.017117183685302736, 0.01722060775756836, 0.017089536666870117, 0.017172479629516603, 0.017122304916381836, 0.017167360305786132, 0.01728102493286133, 0.017069055557250978, 0.017096704483032226, 0.017051647186279297, 0.017106943130493164, 0.017127424240112304, 0.017074176788330078, 0.01705881690979004, 0.01718169593811035, 0.017304576873779298, 0.017087488174438475, 0.01719808006286621, 0.017477632522583008, 0.017145856857299805, 0.017138687133789063, 0.017146879196166993, 0.017128448486328125, 0.017131519317626954, 0.017148927688598634, 0.01715814399719238, 0.01716223907470703, 0.017163263320922852, 0.01745305633544922, 0.01783705520629883, 0.01755340766906738, 0.017438720703125, 0.017511423110961915, 0.017458175659179686, 0.01742950439453125, 0.017334272384643554, 0.017183744430541992, 0.017175552368164062, 0.017107967376708985, 0.017177600860595704, 0.017113088607788086, 0.017464319229125978, 0.01746227264404297, 0.017314847946166993, 0.01740284729003906, 0.01739263916015625, 0.01726464080810547, 0.01721446418762207, 0.017502208709716797, 0.017435647964477538, 0.017386495590209963, 0.01759436798095703, 0.01721241569519043, 0.01718272018432617, 0.017329151153564454, 0.017290239334106446, 0.017330175399780275, 0.01723187255859375, 0.07124479675292969, 0.017326080322265625, 0.017139711380004884, 0.017278976440429687, 0.017701887130737306, 0.01763430404663086, 0.017538047790527343, 0.01742131233215332, 0.017133567810058595, 0.017288192749023438, 0.017093631744384767, 0.01717043113708496, 0.01740390396118164, 0.017159168243408202, 0.017286144256591796, 0.017086463928222655, 0.017097728729248047, 0.01707526397705078, 0.017082304000854493, 0.017130495071411133, 0.01724928092956543, 0.01747148895263672, 0.017130495071411133, 0.017370111465454103, 0.017233951568603516, 0.01709872055053711, 0.01720012855529785, 0.01724723243713379, 0.017596416473388672, 0.018000896453857423, 0.017481727600097655, 0.01745305633544922, 0.0173885440826416, 0.01721139144897461, 0.017119232177734374, 0.017099775314331055, 0.017090560913085938, 0.017114112854003907, 0.017111040115356444, 0.017106943130493164, 0.017160192489624023, 0.017144832611083984, 0.017275903701782228, 0.017148927688598634, 0.017187839508056642, 0.01719603157043457, 0.017157119750976564, 0.01723391914367676, 0.01720217514038086, 0.017122304916381836, 0.017112064361572265, 0.017266687393188478, 0.017228799819946287, 0.01721036720275879, 0.017201152801513672, 0.017266687393188478, 0.01725951957702637, 0.01718681526184082, 0.01721958351135254, 0.01719398307800293, 0.01719500732421875, 0.01720729637145996, 0.01721343994140625, 0.07125196838378907, 0.017332223892211913, 0.017126399993896483, 0.017125375747680666, 0.017315839767456053, 0.0172359676361084, 0.017709056854248048, 0.01720319938659668, 0.017071104049682616, 0.017110015869140623, 0.017253376007080077, 0.017337343215942384, 0.01746329689025879, 0.017383424758911133, 0.017354751586914064, 0.017331199645996095, 0.017344512939453126, 0.01738956832885742, 0.017325056076049804, 0.017353727340698243, 0.017297439575195313, 0.01724720001220703, 0.017154048919677735, 0.017120256423950195, 0.017184768676757813, 0.017107967376708985, 0.017137664794921875, 0.017072128295898437, 0.017093631744384767, 0.017092607498168946, 0.017086463928222655, 0.017054719924926756, 0.017102848052978514, 0.017160192489624023, 0.01720217514038086, 0.017080320358276366, 0.01716531181335449, 0.017116159439086915, 0.01717350387573242, 0.01719193649291992, 0.01714384078979492, 0.01710793685913086, 0.017086463928222655, 0.01742438316345215, 0.017296384811401368, 0.01721139144897461, 0.01721548843383789, 0.017167360305786132, 0.01716223907470703, 0.017209344863891602, 0.01723904037475586, 0.01716633605957031, 0.017133567810058595, 0.017167360305786132, 0.018226175308227538, 0.019099647521972657, 0.017625087738037108, 0.017481727600097655, 0.017358848571777344, 0.017344512939453126, 0.01744588851928711, 0.017830911636352538, 0.01747148895263672, 0.0717158432006836, 0.017460224151611328, 0.01739263916015625, 0.01754319953918457, 0.0174366397857666, 0.017103904724121093, 0.017095647811889648, 0.017340415954589843, 0.01703014373779297, 0.017069055557250978, 0.017064960479736328, 0.01705881690979004, 0.01703321647644043, 0.017124351501464845, 0.017134592056274413, 0.017138687133789063, 0.017055744171142577, 0.01719193649291992, 0.01716223907470703, 0.017125375747680666, 0.017102848052978514, 0.017184768676757813, 0.017093631744384767, 0.01716633605957031, 0.017161216735839844, 0.017144832611083984, 0.01715814399719238, 0.017128480911254882, 0.017109024047851563, 0.017097663879394532, 0.017100799560546876, 0.017130495071411133, 0.017080320358276366, 0.017085439682006837, 0.017070144653320313, 0.017196992874145507, 0.017177600860595704, 0.01720524787902832, 0.017150976181030272, 0.017768447875976562, 0.017509376525878906, 0.01721548843383789, 0.017099775314331055, 0.017144832611083984, 0.017147903442382813, 0.017137664794921875, 0.017087488174438475, 0.017209344863891602, 0.017122304916381836, 0.017140735626220704, 0.017133567810058595, 0.01724825668334961, 0.017161216735839844, 0.017152000427246093, 0.01707827186584473, 0.017157119750976564, 0.017136640548706054, 0.017134592056274413, 0.017155071258544922, 0.017262592315673828, 0.01719910430908203, 0.0172677116394043, 0.017187839508056642, 0.07090585327148438, 0.01716531181335449, 0.017074176788330078, 0.01708140754699707, 0.01732703971862793, 0.01725132751464844, 0.017138687133789063, 0.017073152542114257, 0.017052671432495118, 0.017059839248657227, 0.017060928344726563, 0.017066944122314454, 0.017153024673461914, 0.01707827186584473, 0.017057823181152343, 0.017076192855834962, 0.017063936233520507, 0.017069055557250978, 0.01706188774108887, 0.017121280670166016, 0.017075199127197266, 0.017043455123901367, 0.017083391189575196, 0.017119232177734374, 0.017060863494873048, 0.017060863494873048, 0.01707827186584473, 0.017092607498168946, 0.017132543563842775, 0.017119232177734374, 0.01721958351135254, 0.01722675132751465, 0.017063936233520507, 0.017122304916381836, 0.017142784118652343, 0.017097728729248047, 0.017069055557250978, 0.017133567810058595, 0.017083391189575196, 0.017081344604492187, 0.017133567810058595, 0.017126399993896483, 0.017107967376708985, 0.017101823806762697, 0.017140735626220704, 0.017142784118652343, 0.017098751068115235, 0.017150976181030272, 0.017100799560546876, 0.017113088607788086, 0.017104896545410156, 0.017171455383300782, 0.017217536926269532, 0.01738035202026367, 0.017291263580322267, 0.017141759872436522, 0.017185855865478515, 0.017175487518310547, 0.017139711380004884, 0.017129472732543945, 0.017133567810058595, 0.01717862319946289, 0.017116159439086915, 0.07127961730957032, 0.017172479629516603, 0.017075199127197266, 0.017148927688598634, 0.017123327255249024, 0.017062911987304686, 0.017079296112060546, 0.017094655990600584, 0.017077247619628907, 0.017135616302490234, 0.017093631744384767, 0.017048608779907225, 0.01703318405151367, 0.01705062484741211, 0.017062911987304686, 0.01703321647644043, 0.017112064361572265, 0.017077247619628907, 0.017082368850708008, 0.017088512420654296, 0.017055744171142577, 0.017031167984008787, 0.017035295486450195, 0.01705673599243164, 0.01705062484741211, 0.017090560913085938, 0.017094655990600584, 0.017069055557250978, 0.017062911987304686, 0.017096704483032226, 0.017100799560546876, 0.017079296112060546, 0.017082368850708008, 0.017072128295898437, 0.017188863754272463, 0.017092607498168946, 0.017123327255249024, 0.017124351501464845, 0.017110015869140623, 0.017106943130493164, 0.017121280670166016, 0.017104896545410156, 0.017175552368164062, 0.017132543563842775, 0.017154048919677735, 0.017127424240112304, 0.017089536666870117, 0.017187839508056642, 0.017133567810058595, 0.017096704483032226, 0.017108991622924806, 0.017124351501464845, 0.017157119750976564, 0.017092607498168946, 0.017131519317626954, 0.017119232177734374, 0.017099775314331055, 0.01717452812194824, 0.017154048919677735, 0.017130495071411133, 0.017145856857299805, 0.01725644874572754, 0.017142784118652343, 0.07079116821289062, 0.017130495071411133, 0.017062911987304686, 0.01723187255859375, 0.01724313545227051, 0.017228799819946287, 0.017073152542114257, 0.017073152542114257, 0.017067007064819336, 0.01704140853881836, 0.017037343978881837, 0.01707209587097168, 0.017051647186279297, 0.017145856857299805, 0.01704652786254883, 0.017079296112060546, 0.017047552108764647, 0.017075199127197266, 0.017090560913085938, 0.017079328536987303, 0.017043487548828125, 0.017075136184692384, 0.0172677116394043, 0.017304576873779298, 0.017153024673461914, 0.017176576614379883, 0.017104896545410156, 0.017072128295898437, 0.01706188774108887, 0.01716633605957031, 0.017276927947998046, 0.017080320358276366, 0.017152000427246093, 0.017099775314331055, 0.017084415435791016, 0.017106943130493164, 0.01716633605957031, 0.017148927688598634, 0.017111040115356444, 0.017140735626220704, 0.017096704483032226, 0.017118207931518553, 0.017094655990600584, 0.017209344863891602, 0.017125375747680666, 0.017114112854003907, 0.017111040115356444, 0.01724415969848633, 0.01722675132751465, 0.017155071258544922, 0.01719603157043457, 0.017123327255249024, 0.017132543563842775, 0.01718681526184082, 0.017142784118652343, 0.017121280670166016, 0.017150976181030272, 0.017169408798217774, 0.017152000427246093, 0.017123327255249024, 0.017160192489624023, 0.017107967376708985, 0.017111040115356444, 0.07111270141601563, 0.017130495071411133, 0.017077247619628907, 0.01706188774108887, 0.017072128295898437, 0.017044479370117188, 0.01703731155395508, 0.017114112854003907, 0.01702604866027832, 0.01702604866027832, 0.017062911987304686, 0.017126399993896483, 0.01702092742919922, 0.01706502342224121, 0.017078208923339843, 0.017060863494873048, 0.017072128295898437, 0.017039360046386717, 0.017055744171142577, 0.017063936233520507, 0.0170700798034668, 0.01705062484741211, 0.017081344604492187, 0.017125408172607423, 0.0170546875, 0.017262592315673828, 0.017217536926269532, 0.017083391189575196, 0.017083391189575196, 0.017133567810058595, 0.017060863494873048, 0.017092607498168946, 0.01717043113708496, 0.017136640548706054, 0.017117183685302736, 0.017500160217285156, 0.017333248138427734, 0.01741721534729004, 0.01738751983642578, 0.017152000427246093, 0.017097728729248047, 0.017142784118652343, 0.017265663146972657, 0.017351680755615235, 0.017139711380004884, 0.0172042236328125, 0.017122304916381836, 0.017315839767456053, 0.017192960739135742, 0.017188863754272463, 0.01721651268005371, 0.017129472732543945, 0.017146879196166993, 0.01725644874572754, 0.017091583251953125, 0.017175552368164062, 0.017116159439086915, 0.017142784118652343, 0.017104896545410156, 0.017118207931518553, 0.017361919403076173, 0.017252351760864256, 0.017275903701782228]",tokens/s,55.61254618497385,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,mistralai/Mistral-7B-v0.1,mistralai/Mistral-7B-v0.1,cuda,0,42,,,True,True,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ self.load_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 250, in load_model_with_no_weights self.pretrained_model = self.automodel_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3710, in from_pretrained model = cls(config, *model_args, **model_kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/mistral/modeling_mistral.py"", line 1126, in __init__ self.model = MistralModel(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/mistral/modeling_mistral.py"", line 879, in __init__ [MistralDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/mistral/modeling_mistral.py"", line 879, in [MistralDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/mistral/modeling_mistral.py"", line 679, in __init__ self.mlp = MistralMLP(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/mistral/modeling_mistral.py"", line 165, in __init__ self.gate_proj = nn.Linear(self.hidden_size, self.intermediate_size, bias=False) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/linear.py"", line 98, in __init__ self.weight = Parameter(torch.empty((out_features, in_features), **factory_kwargs)) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_device.py"", line 78, in __torch_function__ return func(*args, **kwargs) torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 224.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,google/recurrentgemma-7b,google/recurrentgemma-7b,cuda,0,42,,,True,True,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/google/recurrentgemma-7b/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66947027-61f3eed65ebd874e31134d16;89bd6c95-d9f8-4e68-ab55-5a0e7c96ded9) Repository Not Found for url: https://huggingface.co/google/recurrentgemma-7b/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: google/recurrentgemma-7b is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,mistralai/Mixtral-8x7B-v0.1,mistralai/Mixtral-8x7B-v0.1,cuda,0,42,,,True,True,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ self.load_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 250, in load_model_with_no_weights self.pretrained_model = self.automodel_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3710, in from_pretrained model = cls(config, *model_args, **model_kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/mixtral/modeling_mixtral.py"", line 1323, in __init__ self.model = MixtralModel(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/mixtral/modeling_mixtral.py"", line 1087, in __init__ [MixtralDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/mixtral/modeling_mixtral.py"", line 1087, in [MixtralDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/mixtral/modeling_mixtral.py"", line 877, in __init__ self.block_sparse_moe = MixtralSparseMoeBlock(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/mixtral/modeling_mixtral.py"", line 824, in __init__ self.experts = nn.ModuleList([MixtralBlockSparseTop2MLP(config) for _ in range(self.num_experts)]) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/mixtral/modeling_mixtral.py"", line 824, in self.experts = nn.ModuleList([MixtralBlockSparseTop2MLP(config) for _ in range(self.num_experts)]) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/mixtral/modeling_mixtral.py"", line 791, in __init__ self.w2 = nn.Linear(self.ffn_dim, self.hidden_dim, bias=False) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/linear.py"", line 98, in __init__ self.weight = Parameter(torch.empty((out_features, in_features), **factory_kwargs)) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_device.py"", line 78, in __torch_function__ return func(*args, **kwargs) torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 224.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,t,t,cuda,0,42,,,True,True,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/t/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66947383-60fe704b1dc0830f382c28c2;e9956a5c-60a2-4d48-838d-3e80c4033e7b) Repository Not Found for url: https://huggingface.co/t/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: t is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-72B,Qwen/Qwen-72B,cuda,0,42,,,True,,,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",qwen,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-30b,facebook/opt-30b,cuda,0,42,,,True,,,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",opt,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-MoE-A2.7B,Qwen/Qwen1.5-MoE-A2.7B,cuda,0,42,,,True,,,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",qwen2_moe,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-1.3b,EleutherAI/pythia-1.3b,cuda,0,42,,,True,True,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1135.607808,6809.976832,0.0,6169.821184,6038.342656,s,10,0.4774934082031251,0.0477493408203125,0.0009943449519893373,0.0473834228515625,0.04809864120483399,0.049400473022460936,0.0504419384765625,"[0.05070230484008789, 0.04743491363525391, 0.04748246383666992, 0.04735868835449219, 0.04733385467529297, 0.04736966323852539, 0.04780934524536133, 0.047322303771972656, 0.04728268814086914, 0.04739718246459961]",tokens/s,5361.330556653423,kWh,5.837328303819417e-07,3.198557834281386e-07,4.061099307990119e-06,4.9646879218001994e-06,tokens/kWh,51564167.58360397,MB,1135.607808,6809.976832,0.0,6169.821184,6138.651648,s,10,11.794597167968748,1.179459716796875,0.006327697856050478,1.1794486694335937,1.1890898925781248,1.1891889892578125,1.1892682666015624,"[1.18018701171875, 1.173260986328125, 1.1726756591796874, 1.1892880859375, 1.18906787109375, 1.18304736328125, 1.18335009765625, 1.171730224609375, 1.173279541015625, 1.1787103271484376]",tokens/s,53.41428715436983,kWh,1.3728726102017813e-05,7.521564732477926e-06,4.8728887368410494e-05,6.997917820290624e-05,tokens/kWh,900267.7884745953,,s,629,12.200843275070193,0.019397207114579002,0.005667996487495601,0.018815999984741212,0.019077529525756837,0.019170470046997068,0.06604796112060547,"[0.018860031127929687, 0.01820159912109375, 0.018263103485107422, 0.018254783630371092, 0.018280448913574218, 0.01824665641784668, 0.018233343124389647, 0.018058240890502928, 0.01822412872314453, 0.018313215255737304, 0.018745344161987306, 0.01906585693359375, 0.018976768493652343, 0.01902796745300293, 0.018961408615112304, 0.01888768005371094, 0.018465791702270508, 0.01899519920349121, 0.018923519134521484, 0.01904025650024414, 0.018963455200195312, 0.018961408615112304, 0.01903104019165039, 0.018988031387329102, 0.018931711196899414, 0.018941951751708985, 0.01897881507873535, 0.018932735443115235, 0.0192491512298584, 0.018677759170532226, 0.01861631965637207, 0.018945024490356444, 0.01905971145629883, 0.018983936309814452, 0.019283967971801756, 0.01957683181762695, 0.018371583938598633, 0.018293760299682618, 0.018929664611816405, 0.019009536743164062, 0.018663423538208008, 0.018997247695922852, 0.018293792724609376, 0.01886409568786621, 0.019144704818725586, 0.018929664611816405, 0.018981887817382814, 0.018988031387329102, 0.019079168319702147, 0.01901055908203125, 0.018824192047119142, 0.018156543731689453, 0.018318336486816408, 0.01808793640136719, 0.018749439239501953, 0.018118656158447266, 0.01802444839477539, 0.01808896064758301, 0.018191360473632814, 0.01827020835876465, 0.018667520523071288, 0.018153472900390624, 0.06681088256835938, 0.0190515193939209, 0.01866547203063965, 0.018137088775634767, 0.01803775978088379, 0.01803980827331543, 0.01862348747253418, 0.01824870491027832, 0.018101247787475586, 0.018316287994384766, 0.018667520523071288, 0.018697216033935548, 0.018601984024047852, 0.01819443130493164, 0.0182609920501709, 0.018258943557739257, 0.01820569610595703, 0.018892831802368164, 0.018950111389160158, 0.018149375915527344, 0.01804083251953125, 0.018135040283203126, 0.018309120178222657, 0.01881088066101074, 0.01863065528869629, 0.018777088165283205, 0.01865830421447754, 0.018637823104858398, 0.018715648651123046, 0.018753536224365236, 0.01864089584350586, 0.018405376434326173, 0.01865216064453125, 0.01880681610107422, 0.019593183517456054, 0.01900851249694824, 0.018219039916992186, 0.018270175933837892, 0.01845145606994629, 0.01905356788635254, 0.018687999725341797, 0.018817024230957033, 0.018799615859985352, 0.018713600158691408, 0.018571264266967775, 0.018708480834960937, 0.018928640365600585, 0.019119104385375976, 0.018889728546142577, 0.01868083190917969, 0.018626560211181642, 0.018927616119384767, 0.019162111282348633, 0.01904435157775879, 0.0184770565032959, 0.018840576171875, 0.01865011215209961, 0.01862246322631836, 0.018678783416748047, 0.018657312393188477, 0.01845756721496582, 0.01865216064453125, 0.018538496017456055, 0.06657126617431641, 0.018985984802246093, 0.018959360122680666, 0.018860031127929687, 0.01867263984680176, 0.018735103607177735, 0.018662399291992187, 0.018573312759399413, 0.018103296279907227, 0.018061311721801757, 0.01818623924255371, 0.018722816467285155, 0.018343936920166014, 0.01844633674621582, 0.01801318359375, 0.01801523208618164, 0.018548736572265623, 0.01869824028015137, 0.018249727249145507, 0.01810534477233887, 0.018167808532714845, 0.01859071922302246, 0.018520063400268554, 0.01808076858520508, 0.018033664703369142, 0.01803980827331543, 0.018043903350830077, 0.01865727996826172, 0.01861427116394043, 0.018669567108154296, 0.018922496795654296, 0.018719743728637696, 0.01884979248046875, 0.019076095581054688, 0.01896348762512207, 0.018963424682617188, 0.018662399291992187, 0.01864396858215332, 0.018752511978149415, 0.01906483268737793, 0.0190699520111084, 0.01903001594543457, 0.01820364761352539, 0.01881395149230957, 0.018949119567871094, 0.01860915184020996, 0.01899622344970703, 0.01905971145629883, 0.018913280487060546, 0.019083263397216797, 0.018952192306518553, 0.01900441551208496, 0.018869247436523438, 0.018991104125976564, 0.018751487731933594, 0.01859993553161621, 0.01860710334777832, 0.01862758445739746, 0.018217983245849608, 0.018728960037231446, 0.018471935272216796, 0.018025472640991212, 0.018077695846557617, 0.06666547393798829, 0.018998271942138673, 0.018953216552734374, 0.01898700714111328, 0.018997312545776367, 0.0189869441986084, 0.01901055908203125, 0.018976768493652343, 0.01903001594543457, 0.01905766487121582, 0.018745344161987306, 0.018581504821777343, 0.01905971145629883, 0.018914304733276367, 0.018335744857788085, 0.018735103607177735, 0.018723840713500976, 0.01833580780029297, 0.018297792434692383, 0.019166208267211913, 0.019772415161132813, 0.019360767364501954, 0.019019775390625, 0.01900748825073242, 0.019087360382080077, 0.019136512756347656, 0.019054655075073243, 0.018967487335205077, 0.019082239151000976, 0.018661376953125, 0.018249727249145507, 0.018970624923706055, 0.018962432861328125, 0.018990079879760743, 0.018723840713500976, 0.018736127853393555, 0.018704383850097657, 0.01864806365966797, 0.018122751235961913, 0.0182609920501709, 0.01825484848022461, 0.019062784194946288, 0.01905971145629883, 0.018982912063598634, 0.019099647521972657, 0.01906892776489258, 0.01903206443786621, 0.0190382080078125, 0.0190382080078125, 0.01888768005371094, 0.018906112670898437, 0.018915327072143554, 0.019072000503540038, 0.019013631820678712, 0.01903206443786621, 0.01883647918701172, 0.018559999465942383, 0.01859584045410156, 0.019309568405151366, 0.0184453125, 0.018461696624755858, 0.018941951751708985, 0.018948095321655273, 0.06608179473876953, 0.018217983245849608, 0.018239488601684572, 0.018945024490356444, 0.01887539291381836, 0.018920448303222655, 0.018355199813842774, 0.018497535705566406, 0.018817024230957033, 0.019091455459594727, 0.018965503692626954, 0.02003455924987793, 0.019146751403808594, 0.0192491512298584, 0.01925632095336914, 0.01902387237548828, 0.019338239669799806, 0.018962432861328125, 0.01880268859863281, 0.018778112411499022, 0.018699264526367186, 0.018287616729736327, 0.018184192657470705, 0.01823641586303711, 0.01885593605041504, 0.018897920608520507, 0.018953216552734374, 0.018189311981201172, 0.018888704299926756, 0.018701311111450195, 0.01868191909790039, 0.01860908889770508, 0.01937411117553711, 0.01910780715942383, 0.01886720085144043, 0.019124223709106446, 0.01897881507873535, 0.018874368667602538, 0.018968576431274413, 0.018908159255981445, 0.019173376083374022, 0.01899929618835449, 0.018988031387329102, 0.018882560729980468, 0.01900032043457031, 0.019013696670532227, 0.019054527282714843, 0.019087360382080077, 0.018229248046875, 0.0186429443359375, 0.019054592132568358, 0.019001344680786132, 0.018860031127929687, 0.019013631820678712, 0.018942975997924806, 0.018856960296630858, 0.0189901123046875, 0.01891017532348633, 0.01904947280883789, 0.01904025650024414, 0.018720767974853517, 0.01887539291381836, 0.019080192565917968, 0.06615142059326172, 0.018611200332641603, 0.018726911544799805, 0.018984960556030273, 0.018958335876464845, 0.018914304733276367, 0.018965503692626954, 0.018952192306518553, 0.018900991439819336, 0.018972671508789063, 0.018924543380737305, 0.018249727249145507, 0.018700288772583007, 0.018601984024047852, 0.01881497573852539, 0.018936832427978514, 0.018952192306518553, 0.018966527938842775, 0.018907136917114258, 0.018965503692626954, 0.018965503692626954, 0.018143232345581056, 0.018321407318115233, 0.019087360382080077, 0.018951168060302736, 0.018892799377441406, 0.018966527938842775, 0.018964479446411133, 0.018994176864624023, 0.01901468849182129, 0.0189071044921875, 0.018916351318359375, 0.018897983551025392, 0.019040191650390625, 0.018933759689331055, 0.018259967803955078, 0.018447359085083007, 0.018910207748413087, 0.01903923225402832, 0.01903513526916504, 0.01906175994873047, 0.018586624145507814, 0.01802444839477539, 0.01800704002380371, 0.0185784969329834, 0.018645952224731446, 0.018989055633544923, 0.018934783935546876, 0.018888704299926756, 0.018950143814086915, 0.018685951232910156, 0.018691072463989256, 0.01862041664123535, 0.018086912155151368, 0.01806540870666504, 0.018337791442871093, 0.018601984024047852, 0.019013631820678712, 0.018922496795654296, 0.0190699520111084, 0.018918399810791017, 0.018991104125976564, 0.018928640365600585, 0.06671462249755859, 0.018685951232910156, 0.018954240798950195, 0.018382848739624022, 0.019070976257324217, 0.01903206443786621, 0.01901670455932617, 0.019183616638183593, 0.018979839324951172, 0.018979839324951172, 0.018966527938842775, 0.01902694320678711, 0.018280448913574218, 0.01823744010925293, 0.018257919311523436, 0.01826201629638672, 0.018911231994628908, 0.019140607833862306, 0.018893823623657227, 0.019054592132568358, 0.018990079879760743, 0.01943449592590332, 0.01904742431640625, 0.018984960556030273, 0.018976768493652343, 0.018956287384033203, 0.019194879531860352, 0.019194879531860352, 0.018782207489013672, 0.018817024230957033, 0.01907711982727051, 0.01903001594543457, 0.01901875114440918, 0.01923891258239746, 0.01904537582397461, 0.01900748825073242, 0.01906790351867676, 0.019137535095214844, 0.0190382080078125, 0.019074047088623047, 0.018944000244140623, 0.019075071334838867, 0.01898700714111328, 0.019161088943481445, 0.019050495147705078, 0.018089984893798827, 0.018110464096069336, 0.018667520523071288, 0.018697216033935548, 0.018670623779296874, 0.018024415969848634, 0.018111488342285157, 0.018076671600341796, 0.01868185615539551, 0.01863987159729004, 0.01863270378112793, 0.018149375915527344, 0.0181790714263916, 0.01808076858520508, 0.018505727767944336, 0.01869004821777344, 0.01815657615661621, 0.01865519905090332, 0.0659609603881836, 0.01808896064758301, 0.01805619239807129, 0.01802239990234375, 0.018561023712158203, 0.01865011215209961, 0.018059263229370116, 0.018069503784179687, 0.018182144165039063, 0.018750463485717773, 0.018792448043823243, 0.018719743728637696, 0.018736127853393555, 0.018737152099609376, 0.018964479446411133, 0.018994176864624023, 0.019086336135864256, 0.019072063446044923, 0.019002304077148438, 0.01903615951538086, 0.019057695388793944, 0.018808799743652342, 0.018982912063598634, 0.0186429443359375, 0.01821900749206543, 0.018387968063354493, 0.018792448043823243, 0.018677759170532226, 0.01863372802734375, 0.01869004821777344, 0.018272256851196288, 0.018753536224365236, 0.01869004821777344, 0.018692096710205077, 0.018588672637939452, 0.01878118324279785, 0.01863065528869629, 0.018646015167236327, 0.01863065528869629, 0.01863577651977539, 0.01839308738708496, 0.01812384033203125, 0.018821056365966798, 0.018264064788818358, 0.018118656158447266, 0.018898944854736328, 0.018976768493652343, 0.0192675838470459, 0.019065887451171874, 0.018428895950317385, 0.018131967544555663, 0.018748416900634765, 0.018918399810791017, 0.01904844856262207, 0.018873344421386717, 0.01802342414855957, 0.01803980827331543, 0.018035711288452147, 0.018663423538208008, 0.018504703521728515, 0.018051071166992186, 0.018264064788818358, 0.018684928894042968, 0.06591385650634765, 0.018900991439819336, 0.018989055633544923, 0.020364288330078126, 0.019544063568115236, 0.019111936569213867, 0.018975744247436522, 0.018963455200195312, 0.019095552444458007, 0.018962432861328125, 0.018954240798950195, 0.018692096710205077, 0.018127872467041017, 0.01810534477233887, 0.018061311721801757, 0.01816268730163574, 0.01823744010925293, 0.018326528549194337, 0.01825484848022461, 0.018310144424438478, 0.018256895065307616, 0.018966527938842775, 0.01843404769897461, 0.018304000854492186, 0.018268159866333008, 0.018316287994384766, 0.018229280471801758, 0.018230239868164064, 0.01821286392211914, 0.01821286392211914, 0.018307071685791015, 0.01821388816833496, 0.01826201629638672, 0.018505727767944336, 0.01804287910461426, 0.018156543731689453, 0.01820159912109375, 0.018751487731933594, 0.018697216033935548, 0.018619392395019533, 0.01816268730163574, 0.01845043182373047, 0.019002368927001953, 0.018946048736572265, 0.019096576690673828, 0.019124223709106446, 0.019729408264160156, 0.01916214370727539, 0.019052511215209962, 0.019096576690673828, 0.0190515193939209, 0.018974720001220705, 0.01887129592895508, 0.01865216064453125, 0.01819647979736328, 0.018336767196655272, 0.01823641586303711, 0.01865318489074707, 0.018937856674194335, 0.018952192306518553, 0.01818726348876953, 0.018258943557739257, 0.01822412872314453, 0.06621593475341797, 0.01827020835876465, 0.01817087936401367, 0.018356224060058594, 0.018256895065307616, 0.01828556823730469, 0.018249727249145507, 0.018239488601684572, 0.018508800506591795, 0.01907302474975586, 0.019140607833862306, 0.018952192306518553, 0.018944000244140623, 0.01900339126586914, 0.019013631820678712, 0.019014656066894533, 0.01903411293029785, 0.018811904907226562, 0.018576383590698242, 0.01826304054260254, 0.018288639068603514, 0.018239488601684572, 0.018215936660766603, 0.01822822380065918, 0.018255872726440428, 0.018815999984741212, 0.018967552185058592, 0.018936832427978514, 0.01903513526916504, 0.01903411293029785, 0.01899212837219238, 0.018981887817382814, 0.01904025650024414, 0.019005504608154297, 0.019173311233520507, 0.01900851249694824, 0.018981887817382814, 0.019019775390625, 0.01888768005371094, 0.01864908790588379, 0.018773056030273436, 0.018710464477539063, 0.018217983245849608, 0.01883443260192871, 0.018584575653076172, 0.018268159866333008, 0.018309120178222657, 0.01819647979736328, 0.01826918411254883, 0.01820364761352539, 0.01822003173828125, 0.018199552536010744, 0.01909350395202637, 0.01904947280883789, 0.01898089599609375, 0.01892246437072754, 0.018953216552734374, 0.01902694320678711, 0.018982912063598634, 0.018974720001220705, 0.019098623275756836, 0.019100671768188478, 0.019013631820678712]",tokens/s,51.55381360280454,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-1.4b,EleutherAI/pythia-1.4b,cuda,0,42,,,True,True,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1140.355072,6809.976832,0.0,6169.821184,6038.342656,s,10,0.47966221237182616,0.04796622123718262,0.0007691327798441897,0.04779440116882324,0.0483730770111084,0.04923428325653076,0.04992324825286865,"[0.050095489501953125, 0.04746470260620117, 0.04725696182250977, 0.04799817657470703, 0.04747948837280273, 0.04759062576293945, 0.04818169784545898, 0.048028766632080076, 0.04752729415893555, 0.048039009094238284]",tokens/s,5337.089172276783,kWh,5.837309862918781e-07,3.1982100842241164e-07,4.071502983527083e-06,4.975054978241371e-06,tokens/kWh,51456717.78897472,MB,1140.355072,6809.976832,0.0,6169.821184,6138.651648,s,10,12.003288085937498,1.2003288085937498,0.014982751552041281,1.2024702758789063,1.21298896484375,1.2213482421875,1.2280356640624999,"[1.22970751953125, 1.19094677734375, 1.16836669921875, 1.2027806396484375, 1.189383056640625, 1.2023114013671874, 1.21113134765625, 1.202629150390625, 1.2047396240234376, 1.2012918701171875]",tokens/s,52.48561856463973,kWh,1.3905145442577158e-05,7.617866285702359e-06,4.862481695127448e-05,7.014782867955399e-05,tokens/kWh,898103.3509646269,,s,629,12.41336014938355,0.019735071779624073,0.005729428753984044,0.019175424575805664,0.019344998931884766,0.019746816635131836,0.06659940368652345,"[0.019190784454345702, 0.01903206443786621, 0.01924710464477539, 0.019734527587890623, 0.01991993522644043, 0.020001728057861327, 0.019952640533447266, 0.019772415161132813, 0.019746816635131836, 0.01982361602783203, 0.020435968399047853, 0.019950592041015625, 0.019336191177368164, 0.01921433639526367, 0.01925632095336914, 0.019985408782958985, 0.01987583923339844, 0.019915775299072267, 0.020152351379394532, 0.01988809585571289, 0.01982464027404785, 0.019672063827514647, 0.019746816635131836, 0.01985843276977539, 0.019722240447998047, 0.019893247604370116, 0.019744768142700195, 0.019619903564453124, 0.019469247817993165, 0.019283967971801756, 0.019358720779418945, 0.019043392181396484, 0.01917535972595215, 0.018775039672851563, 0.019187711715698243, 0.019219455718994142, 0.019171327590942384, 0.019117055892944337, 0.019396608352661132, 0.0191016960144043, 0.019147775650024415, 0.01923072052001953, 0.019106815338134766, 0.019176448822021484, 0.01922662353515625, 0.019144704818725586, 0.0193832950592041, 0.019203071594238282, 0.019137535095214844, 0.019120128631591796, 0.01924505615234375, 0.02048409652709961, 0.020222976684570314, 0.019615743637084963, 0.019252223968505858, 0.01924505615234375, 0.019174400329589843, 0.019109888076782225, 0.019150848388671874, 0.019487775802612305, 0.020645856857299805, 0.019575807571411134, 0.06711910247802734, 0.019195903778076173, 0.019215360641479492, 0.019172351837158205, 0.019143680572509765, 0.01944063949584961, 0.019182592391967773, 0.019156991958618166, 0.019131391525268555, 0.019547136306762695, 0.019174400329589843, 0.019290111541748048, 0.019175424575805664, 0.019135488510131835, 0.019293184280395507, 0.019290111541748048, 0.019173376083374022, 0.01921433639526367, 0.019150848388671874, 0.019113983154296875, 0.019171327590942384, 0.019188735961914064, 0.01908940887451172, 0.019182592391967773, 0.019183647155761718, 0.019126239776611327, 0.01919692802429199, 0.01918976020812988, 0.019108863830566408, 0.019219455718994142, 0.01920102310180664, 0.019150848388671874, 0.019169279098510742, 0.019178495407104493, 0.019175424575805664, 0.01864806365966797, 0.01846272087097168, 0.0186060791015625, 0.018534400939941405, 0.01856716728210449, 0.018527231216430663, 0.018584640502929687, 0.01856812858581543, 0.018501632690429686, 0.01858358383178711, 0.018563039779663087, 0.018520063400268554, 0.01855897521972656, 0.018503679275512695, 0.018491392135620118, 0.0186060791015625, 0.018506752014160157, 0.018490367889404297, 0.018549760818481444, 0.01844223976135254, 0.018413568496704103, 0.018508832931518556, 0.0185067195892334, 0.018455551147460936, 0.018541568756103514, 0.01846681594848633, 0.01843507194519043, 0.018479103088378905, 0.0661934051513672, 0.018535423278808593, 0.0184586238861084, 0.018482175827026368, 0.018380800247192384, 0.01858252716064453, 0.018457599639892578, 0.018480127334594726, 0.01841049575805664, 0.018535423278808593, 0.018331647872924805, 0.01838387107849121, 0.018325504302978517, 0.018293760299682618, 0.018292800903320312, 0.018249664306640625, 0.018300928115844727, 0.018390016555786134, 0.01848320007324219, 0.018455551147460936, 0.018495487213134765, 0.018491392135620118, 0.0184453125, 0.018482175827026368, 0.018465791702270508, 0.018487295150756835, 0.0184453125, 0.01845248031616211, 0.0184453125, 0.018499584197998048, 0.01845145606994629, 0.018481151580810547, 0.018472959518432617, 0.018455551147460936, 0.018421760559082033, 0.01850060844421387, 0.018530303955078126, 0.018450464248657227, 0.01846063995361328, 0.018481151580810547, 0.018471935272216796, 0.018496511459350586, 0.018497535705566406, 0.018381824493408205, 0.018480127334594726, 0.01849241638183594, 0.018412544250488282, 0.018562047958374024, 0.019050495147705078, 0.018512895584106445, 0.018498559951782227, 0.018498559951782227, 0.018526208877563476, 0.01945088005065918, 0.01924198341369629, 0.019125247955322267, 0.019204095840454103, 0.019188735961914064, 0.019111936569213867, 0.018516992568969725, 0.01847500801086426, 0.018479103088378905, 0.0185262393951416, 0.06707913970947266, 0.018968576431274413, 0.019175424575805664, 0.019307519912719725, 0.01920512008666992, 0.018541568756103514, 0.018550783157348632, 0.018515968322753908, 0.018501632690429686, 0.018508800506591795, 0.018482175827026368, 0.018752511978149415, 0.01922969627380371, 0.019151872634887695, 0.019171327590942384, 0.019207168579101562, 0.019224576950073242, 0.01914886474609375, 0.01892755126953125, 0.01941094398498535, 0.019192832946777344, 0.01919696044921875, 0.01914467239379883, 0.018750463485717773, 0.019192832946777344, 0.019351551055908203, 0.0192174072265625, 0.019324928283691405, 0.019170303344726563, 0.019178495407104493, 0.01923072052001953, 0.019147775650024415, 0.0184770565032959, 0.018513919830322266, 0.01929113578796387, 0.020198400497436524, 0.019627071380615233, 0.019247039794921875, 0.019216384887695313, 0.01904640007019043, 0.019088384628295898, 0.01923379135131836, 0.018512895584106445, 0.01849241638183594, 0.018593791961669923, 0.01927475166320801, 0.019358720779418945, 0.01917750358581543, 0.019226591110229493, 0.019138559341430664, 0.01916624069213867, 0.019272672653198243, 0.019141632080078123, 0.019184640884399414, 0.01924710464477539, 0.019449920654296876, 0.019213247299194335, 0.01929523277282715, 0.019184640884399414, 0.019170303344726563, 0.019172351837158205, 0.019185663223266602, 0.01922969627380371, 0.0664268798828125, 0.018554880142211915, 0.018568191528320312, 0.018553855895996094, 0.018498559951782227, 0.018561023712158203, 0.01929523277282715, 0.019198976516723632, 0.019094528198242186, 0.019132415771484376, 0.019208192825317383, 0.01923276710510254, 0.0192993278503418, 0.019224576950073242, 0.019160064697265625, 0.01927884864807129, 0.019165184020996092, 0.019311616897583008, 0.01850060844421387, 0.018518016815185546, 0.018497535705566406, 0.01904128074645996, 0.01919795227050781, 0.019317760467529296, 0.019212287902832033, 0.019500032424926757, 0.01963520050048828, 0.019307519912719725, 0.019137535095214844, 0.019377151489257814, 0.01923379135131836, 0.018913280487060546, 0.019182592391967773, 0.018577407836914063, 0.018481151580810547, 0.018524160385131837, 0.018522111892700196, 0.01850060844421387, 0.018515968322753908, 0.018530303955078126, 0.0184586238861084, 0.018530303955078126, 0.018511871337890624, 0.01848422431945801, 0.01849456024169922, 0.018499488830566405, 0.01846784019470215, 0.01999667167663574, 0.019162111282348633, 0.019125247955322267, 0.019148799896240236, 0.019140607833862306, 0.019143680572509765, 0.01923481559753418, 0.018471935272216796, 0.018449407577514648, 0.0184453125, 0.018507776260375978, 0.01859686470031738, 0.018580480575561522, 0.018486272811889647, 0.01849241638183594, 0.018508800506591795, 0.06666649627685547, 0.01880985641479492, 0.018586624145507814, 0.018561023712158203, 0.018482175827026368, 0.018540544509887694, 0.018499584197998048, 0.01850060844421387, 0.018518016815185546, 0.018547712326049806, 0.018464767456054687, 0.018611200332641603, 0.018494464874267577, 0.019162111282348633, 0.019224576950073242, 0.019195999145507812, 0.01922960090637207, 0.01924608039855957, 0.019219455718994142, 0.01906073570251465, 0.019216384887695313, 0.01925632095336914, 0.01927475166320801, 0.01921331214904785, 0.019219455718994142, 0.019163135528564454, 0.019169279098510742, 0.01927884864807129, 0.019163135528564454, 0.01928294372558594, 0.01916524887084961, 0.019212223052978514, 0.01918976020812988, 0.019293184280395507, 0.019162111282348633, 0.019296255111694336, 0.0192174072265625, 0.019174400329589843, 0.01923583984375, 0.01920614433288574, 0.01924710464477539, 0.01923174476623535, 0.01924608039855957, 0.019120128631591796, 0.01923583984375, 0.01924198341369629, 0.01937919998168945, 0.01927168083190918, 0.01922150421142578, 0.019169279098510742, 0.019158016204833983, 0.01920102310180664, 0.019215360641479492, 0.01925324821472168, 0.019123199462890626, 0.019168256759643554, 0.019212287902832033, 0.01928294372558594, 0.01922559928894043, 0.019162111282348633, 0.019215360641479492, 0.019108928680419923, 0.019227584838867186, 0.06794547271728515, 0.019092479705810548, 0.019204095840454103, 0.019195903778076173, 0.019207168579101562, 0.019124223709106446, 0.019212287902832033, 0.019188735961914064, 0.019142656326293944, 0.0192174072265625, 0.019116031646728517, 0.01922764778137207, 0.019160064697265625, 0.01923072052001953, 0.019190784454345702, 0.01926144027709961, 0.019156991958618166, 0.019284000396728517, 0.0191866569519043, 0.01927065658569336, 0.01932089614868164, 0.019302335739135743, 0.019302400588989257, 0.01921843147277832, 0.01923379135131836, 0.019169279098510742, 0.019268608093261717, 0.019203071594238282, 0.01924403190612793, 0.019154943466186524, 0.0192675838470459, 0.019260416030883788, 0.019184640884399414, 0.01924710464477539, 0.019322879791259767, 0.019310592651367187, 0.019179519653320314, 0.01923788833618164, 0.019107839584350587, 0.0192993278503418, 0.01923891258239746, 0.01918156814575195, 0.019293184280395507, 0.01918976020812988, 0.0192225284576416, 0.01920614433288574, 0.0192542724609375, 0.019185663223266602, 0.01927577590942383, 0.019176448822021484, 0.01924198341369629, 0.01923379135131836, 0.01923583984375, 0.0192194881439209, 0.01924300765991211, 0.01991881561279297, 0.01925836753845215, 0.019343360900878907, 0.01923993682861328, 0.0192225284576416, 0.019110912322998046, 0.018712575912475587, 0.018685951232910156, 0.06779904174804688, 0.01921433639526367, 0.01925119972229004, 0.01922047996520996, 0.01887027168273926, 0.019135488510131835, 0.019141632080078123, 0.019074047088623047, 0.01928294372558594, 0.01924812889099121, 0.01920204734802246, 0.01925529670715332, 0.019225664138793945, 0.019174335479736328, 0.01925939178466797, 0.01922867202758789, 0.01918976020812988, 0.01919491195678711, 0.01916208076477051, 0.019372032165527343, 0.01922764778137207, 0.01925734329223633, 0.019122175216674805, 0.01921843147277832, 0.01922559928894043, 0.01914886474609375, 0.019252159118652343, 0.019169279098510742, 0.019147775650024415, 0.019211263656616212, 0.019178495407104493, 0.01918976020812988, 0.019216384887695313, 0.019208192825317383, 0.019006464004516603, 0.018542591094970702, 0.018714624404907225, 0.01920204734802246, 0.019180543899536134, 0.019195903778076173, 0.019150848388671874, 0.019155967712402345, 0.01924198341369629, 0.019088384628295898, 0.019159040451049804, 0.019155967712402345, 0.01924608039855957, 0.019307519912719725, 0.019112960815429687, 0.018725887298583984, 0.019111936569213867, 0.019170303344726563, 0.019165184020996092, 0.01924505615234375, 0.01901670455932617, 0.018856960296630858, 0.018390016555786134, 0.018386943817138672, 0.018903039932250978, 0.018559999465942383, 0.018363391876220703, 0.018537471771240235, 0.018541568756103514, 0.06720716857910156, 0.0192675838470459, 0.019211263656616212, 0.019331071853637697, 0.019002368927001953, 0.018947071075439453, 0.019384319305419923, 0.01923379135131836, 0.019186752319335938, 0.019213247299194335, 0.019169279098510742, 0.01926655960083008, 0.019194879531860352, 0.01920512008666992, 0.019183616638183593, 0.019175424575805664, 0.01924300765991211, 0.019277824401855468, 0.0192491512298584, 0.019160064697265625, 0.01903718376159668, 0.018949119567871094, 0.019116031646728517, 0.018591808319091796, 0.0185599365234375, 0.018569215774536133, 0.019075071334838867, 0.019233888626098632, 0.01920604705810547, 0.01923788833618164, 0.019155967712402345, 0.019166208267211913, 0.019192832946777344, 0.019204095840454103, 0.019149824142456053, 0.01922764778137207, 0.019188735961914064, 0.019192832946777344, 0.019186687469482423, 0.01922867202758789, 0.019179519653320314, 0.019260416030883788, 0.01920614433288574, 0.01917344093322754, 0.019166208267211913, 0.01926652717590332, 0.019009504318237305, 0.019173376083374022, 0.019134464263916014, 0.019088384628295898, 0.018495487213134765, 0.018493440628051756, 0.01847305679321289, 0.018502559661865235, 0.018965503692626954, 0.01887846374511719, 0.020139072418212892, 0.01914566421508789, 0.0192225284576416, 0.01921023941040039, 0.019177471160888672, 0.019117055892944337, 0.01921843147277832, 0.0681707534790039, 0.01927475166320801, 0.019087360382080077, 0.019156991958618166, 0.019145727157592773, 0.019141632080078123, 0.0192542724609375, 0.019323904037475585, 0.01926348876953125, 0.01922047996520996, 0.019136512756347656, 0.019175424575805664, 0.01921023941040039, 0.01922150421142578, 0.019194879531860352, 0.019208192825317383, 0.019159040451049804, 0.019177471160888672, 0.01924300765991211, 0.019292160034179686, 0.019139583587646485, 0.01927577590942383, 0.019110912322998046, 0.01923174476623535, 0.01904025650024414, 0.018912256240844725, 0.018592767715454102, 0.018565120697021483, 0.01862451171875, 0.018610176086425782, 0.018569215774536133, 0.01864192008972168, 0.018528255462646484, 0.018564096450805666, 0.018548736572265623, 0.01863680076599121, 0.018507776260375978, 0.018538496017456055, 0.018569215774536133, 0.018565120697021483, 0.018548736572265623, 0.018577472686767578, 0.018594751358032225, 0.01922559928894043, 0.01920204734802246, 0.019272703170776367, 0.01968332862854004, 0.01919388771057129, 0.01923683166503906, 0.019126304626464842, 0.019258335113525392, 0.019292192459106447, 0.019409887313842772, 0.019215360641479492, 0.019388416290283202, 0.019564544677734375, 0.01923993682861328, 0.01926246452331543, 0.019219455718994142, 0.019192832946777344, 0.01918976020812988, 0.01930342483520508, 0.019342336654663086]",tokens/s,50.67121169695835,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,x,x,cuda,0,42,,,True,True,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/x/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66932454-69e11d2f3f19a13a5de2e970;aa63d1fa-6e0b-461a-bc9f-5091b063f7c1) Repository Not Found for url: https://huggingface.co/x/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: x is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,01-ai/Yi-34B,01-ai/Yi-34B,cuda,0,42,,,True,,,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",llama,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/xglm-4.5B,,cuda,0,42,,,,,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,d35829e539df8480b726c647eeabf91e41eae047,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,,MB,1267.949568,20984.6272,0.0,20344.471552,18704.26112,s,10,1.5023166809082031,0.1502316680908203,0.0006441345506977998,0.1502587661743164,0.1509025588989258,0.1511650703430176,0.15137507949829102,"[0.15142758178710938, 0.149262939453125, 0.150144287109375, 0.15084422302246095, 0.15062396240234374, 0.15017861938476562, 0.1504398651123047, 0.14921498107910156, 0.14984130859375, 0.1503389129638672]",tokens/s,1704.0348633101712,kWh,1.8083386599816858e-06,9.908841437209047e-07,1.2620569860757291e-05,1.541979266445988e-05,tokens/kWh,16602039.052706491,MB,1267.949568,20984.6272,0.0,20344.471552,18903.142912,s,10,28.97810498046875,2.897810498046875,0.003902678652760611,2.89655810546875,2.9020618408203127,2.9033882202148438,2.9044493237304687,"[2.9010439453125, 2.90176708984375, 2.895592041015625, 2.904714599609375, 2.89720263671875, 2.8955634765625, 2.901038330078125, 2.892088623046875, 2.89591357421875, 2.8931806640625]",tokens/s,21.740552062483733,kWh,3.416213252678865e-05,1.8722522664320418e-05,0.00013508754830564908,0.0001879722034967582,tokens/kWh,335155.9370377148,,s,629,30.282588092803962,0.04814401922544349,0.01791632267039625,0.04593664169311523,0.04626657180786133,0.04642242584228516,0.19630895141601562,"[0.04616089630126953, 0.04587724685668945, 0.045914112091064455, 0.0458260498046875, 0.04592950439453125, 0.04617212677001953, 0.04592435073852539, 0.04583628845214844, 0.04604620742797851, 0.04650598526000976, 0.04631449508666992, 0.04587417602539062, 0.04608412933349609, 0.04583427047729492, 0.04582393646240234, 0.045778942108154294, 0.04592947387695313, 0.04581171035766601, 0.045723648071289064, 0.04582092666625977, 0.04605542373657227, 0.045808639526367184, 0.04585683059692383, 0.046160831451416015, 0.04600934219360352, 0.04582092666625977, 0.04588236618041992, 0.046080001831054686, 0.04582809448242187, 0.0458158073425293, 0.04588956832885742, 0.04581372833251953, 0.04582297515869141, 0.04576051330566406, 0.04592844772338867, 0.045875297546386716, 0.045821857452392575, 0.04591820907592774, 0.04594790267944336, 0.04638412857055664, 0.04794572830200195, 0.046532608032226565, 0.04618239974975586, 0.046050304412841796, 0.04600012969970703, 0.04581171035766601, 0.046173183441162106, 0.046159870147705076, 0.046453758239746096, 0.04617830276489258, 0.046148609161376954, 0.04589977645874024, 0.04619059371948242, 0.04604828643798828, 0.04619055938720703, 0.046086143493652344, 0.04618137741088867, 0.046069759368896485, 0.04595609664916992, 0.04615884780883789, 0.046036991119384765, 0.046203903198242184, 0.19600076293945312, 0.045813793182373046, 0.045989856719970704, 0.046086143493652344, 0.0458342399597168, 0.04574105453491211, 0.04579942321777344, 0.04579328155517578, 0.04576563262939453, 0.04601958465576172, 0.046115840911865234, 0.046088191986083986, 0.04600934219360352, 0.046023681640625, 0.04577280044555664, 0.045963264465332034, 0.04582912063598633, 0.04586598587036133, 0.04577280044555664, 0.04592127990722656, 0.04582608032226562, 0.045817825317382814, 0.04579635238647461, 0.045810688018798826, 0.04576768112182617, 0.04607078552246094, 0.045965312957763675, 0.045818878173828126, 0.045884414672851564, 0.04588137435913086, 0.045801441192626954, 0.04593971252441406, 0.045908992767333984, 0.046099456787109375, 0.0459233283996582, 0.04597862243652344, 0.046360576629638675, 0.04672614288330078, 0.046279678344726564, 0.04608204650878906, 0.045878273010253906, 0.046274559020996094, 0.04586086273193359, 0.04608512115478516, 0.046569503784179685, 0.046736351013183595, 0.04613836669921875, 0.04693401718139648, 0.046301185607910154, 0.046235649108886716, 0.046002174377441404, 0.046061569213867185, 0.04605132675170898, 0.04594278335571289, 0.04609638214111328, 0.04616191864013672, 0.046281726837158206, 0.04628070449829102, 0.04618137741088867, 0.04613631820678711, 0.04631654357910156, 0.04640460968017578, 0.04618547058105469, 0.1966991424560547, 0.04572159957885742, 0.04575948715209961, 0.04588748931884765, 0.04585779190063476, 0.04602982330322265, 0.045825023651123044, 0.04605542373657227, 0.04582912063598633, 0.045870079040527346, 0.045801536560058594, 0.04579935836791992, 0.045774848937988284, 0.045810688018798826, 0.04591820907592774, 0.045840385437011716, 0.04585881423950195, 0.045871105194091794, 0.0460840950012207, 0.04587724685668945, 0.04588345718383789, 0.045697982788085935, 0.04579020690917969, 0.04593459320068359, 0.045995105743408204, 0.0458177604675293, 0.04616806411743164, 0.04595916748046875, 0.04584454345703125, 0.04604512023925781, 0.045897727966308595, 0.046004222869873046, 0.045830142974853515, 0.04605440139770508, 0.04591923141479492, 0.04604620742797851, 0.046036991119384765, 0.04597350311279297, 0.0459048957824707, 0.04589977645874024, 0.046091262817382815, 0.045868030548095705, 0.046298110961914066, 0.04594483184814453, 0.045932544708251956, 0.04589468765258789, 0.0460513916015625, 0.046010272979736325, 0.046089214324951173, 0.045881343841552735, 0.04601241683959961, 0.046083072662353515, 0.04585369491577149, 0.04609535980224609, 0.04593766403198242, 0.046306304931640625, 0.046018558502197264, 0.046053375244140625, 0.04596633529663086, 0.046148609161376954, 0.04614553451538086, 0.046099456787109375, 0.046124031066894534, 0.19773234558105468, 0.04665446472167969, 0.04600831985473633, 0.04610662460327149, 0.04598067092895508, 0.04581171035766601, 0.04583935928344727, 0.045840385437011716, 0.04583628845214844, 0.04604313659667969, 0.04591308975219727, 0.046064640045166014, 0.04603801727294922, 0.04599603271484375, 0.04634009552001953, 0.04600627136230469, 0.04581683349609375, 0.046063617706298826, 0.045910079956054686, 0.046046142578125, 0.045778942108154294, 0.045908031463623045, 0.045967296600341795, 0.04611481475830078, 0.04583321762084961, 0.04594483184814453, 0.04599193572998047, 0.04584550476074219, 0.0458076171875, 0.045810688018798826, 0.04601651382446289, 0.04591001510620117, 0.04591001510620117, 0.04594790267944336, 0.04598681640625, 0.045932544708251956, 0.04623769760131836, 0.04611072158813476, 0.04617216110229492, 0.046209022521972655, 0.04632064056396484, 0.04634009552001953, 0.04623155212402344, 0.046252033233642575, 0.04630527877807617, 0.046241790771484374, 0.04642406463623047, 0.04644454574584961, 0.04624281692504883, 0.04640563201904297, 0.04616191864013672, 0.04613631820678711, 0.045985790252685545, 0.04613529586791992, 0.04598988723754883, 0.046048255920410154, 0.046004222869873046, 0.046118911743164064, 0.04607385635375977, 0.047001598358154296, 0.04673126220703125, 0.046260223388671876, 0.04624281692504883, 0.1972418518066406, 0.045770751953125, 0.0457625617980957, 0.04593664169311523, 0.04602470397949219, 0.045813758850097655, 0.0458076171875, 0.045846527099609374, 0.04585884857177734, 0.045829086303710936, 0.045808639526367184, 0.04580556869506836, 0.04575027084350586, 0.045871105194091794, 0.046042110443115236, 0.045900798797607424, 0.04591206359863281, 0.04601446533203125, 0.04594073486328125, 0.04590694427490234, 0.045825088500976566, 0.045830078125, 0.045808639526367184, 0.046209022521972655, 0.04587417602539062, 0.0459233283996582, 0.046153728485107424, 0.04601036834716797, 0.046002174377441404, 0.045960193634033204, 0.04588544082641602, 0.046219264984130856, 0.0460134391784668, 0.045884414672851564, 0.04610460662841797, 0.046079967498779295, 0.04592639923095703, 0.04597043228149414, 0.04596428680419922, 0.04616089630126953, 0.046140415191650394, 0.04633599853515625, 0.046004222869873046, 0.04652134323120117, 0.04618547058105469, 0.04593459320068359, 0.045946880340576174, 0.04594790267944336, 0.04597862243652344, 0.045878273010253906, 0.046036991119384765, 0.04589363098144531, 0.046069759368896485, 0.04593561553955078, 0.046094337463378904, 0.04591513442993164, 0.045868030548095705, 0.04636876678466797, 0.046058494567871096, 0.04594585418701172, 0.046002174377441404, 0.04607590484619141, 0.04605952072143555, 0.19731149291992187, 0.046287872314453124, 0.04581785583496094, 0.04574310302734375, 0.0456734733581543, 0.04575027084350586, 0.045927486419677734, 0.04577990341186523, 0.04575743865966797, 0.0459233283996582, 0.04576768112182617, 0.04574515151977539, 0.04578201675415039, 0.04583833694458008, 0.04574105453491211, 0.045808639526367184, 0.045753345489501954, 0.04578406524658203, 0.04575948715209961, 0.046148609161376954, 0.046096446990966794, 0.04599599838256836, 0.0458004150390625, 0.04583833694458008, 0.04585369491577149, 0.04581683349609375, 0.04576665496826172, 0.04581478500366211, 0.04593971252441406, 0.045732929229736326, 0.045806526184082035, 0.04574105453491211, 0.04577177429199219, 0.04583935928344727, 0.045992961883544924, 0.04597862243652344, 0.0459315185546875, 0.04606473541259765, 0.04617718505859375, 0.04617523193359375, 0.04585779190063476, 0.04624895858764649, 0.04622438430786133, 0.04620697784423828, 0.04597555160522461, 0.04604313659667969, 0.046015487670898435, 0.046235649108886716, 0.04595097732543945, 0.04591308975219727, 0.04605440139770508, 0.04595199966430664, 0.046004222869873046, 0.04599193572998047, 0.04612505722045898, 0.04601651382446289, 0.046173183441162106, 0.046295040130615236, 0.046071807861328126, 0.0460769271850586, 0.046043167114257814, 0.04610351943969727, 0.04597760009765625, 0.1973544921875, 0.04614144134521484, 0.045917182922363284, 0.046023681640625, 0.04575033569335937, 0.04632160186767578, 0.04609638214111328, 0.04627558517456055, 0.04617728042602539, 0.04604006576538086, 0.046124031066894534, 0.04621209716796875, 0.046271488189697264, 0.04617216110229492, 0.046219264984130856, 0.046104576110839846, 0.047162368774414064, 0.04659814453125, 0.04620083236694336, 0.04656947326660156, 0.046018558502197264, 0.045843456268310545, 0.045827072143554685, 0.04584960174560547, 0.04578713607788086, 0.046086143493652344, 0.045868030548095705, 0.0459950065612793, 0.04579840087890625, 0.046048255920410154, 0.04586809539794922, 0.045886398315429684, 0.04582195281982422, 0.04611174392700195, 0.04590694427490234, 0.04581171035766601, 0.04590591812133789, 0.0459315185546875, 0.04585580825805664, 0.04592022323608398, 0.04586288070678711, 0.04592025756835937, 0.04589567947387695, 0.045908992767333984, 0.045846527099609374, 0.04594073486328125, 0.04584755325317383, 0.04592537689208984, 0.045917182922363284, 0.04594585418701172, 0.045889537811279295, 0.04598988723754883, 0.04584447860717773, 0.046069759368896485, 0.04592127990722656, 0.04601036834716797, 0.04593971252441406, 0.046083072662353515, 0.046448638916015625, 0.04599398422241211, 0.04592844772338867, 0.04598476791381836, 0.046513153076171876, 0.1957171173095703, 0.045778942108154294, 0.0456703987121582, 0.045795326232910154, 0.04570111846923828, 0.045695999145507815, 0.045687808990478515, 0.04568371200561523, 0.04569804763793945, 0.04594483184814453, 0.04580249786376953, 0.04641996765136719, 0.04698521423339844, 0.045908992767333984, 0.04579942321777344, 0.045963264465332034, 0.04580044937133789, 0.04583731079101563, 0.0457625617980957, 0.045897727966308595, 0.04575539016723633, 0.04580556869506836, 0.04575955200195313, 0.045784000396728516, 0.04574617767333984, 0.04586703872680664, 0.045763553619384764, 0.0458342399597168, 0.04577587127685547, 0.045774848937988284, 0.04583935928344727, 0.04587724685668945, 0.045982719421386715, 0.04593766403198242, 0.045916160583496096, 0.04592230224609375, 0.04590182495117188, 0.045870079040527346, 0.045894657135009766, 0.04587417602539062, 0.04576979064941406, 0.04584134292602539, 0.045846527099609374, 0.04578303909301758, 0.04582710266113281, 0.04589667129516602, 0.045851646423339845, 0.045889537811279295, 0.045878273010253906, 0.04594489669799805, 0.04615366363525391, 0.04611072158813476, 0.045889537811279295, 0.04595199966430664, 0.04599814224243164, 0.046034881591796875, 0.04594278335571289, 0.04601139068603516, 0.045946880340576174, 0.04598681640625, 0.04595916748046875, 0.046107646942138675, 0.04624281692504883, 0.19642880249023437, 0.045797374725341795, 0.04583833694458008, 0.04583833694458008, 0.04597452926635742, 0.04625305557250976, 0.0458076171875, 0.04576768112182617, 0.04578713607788086, 0.04581478500366211, 0.045764606475830076, 0.04578924942016602, 0.04576147079467773, 0.04572467041015625, 0.04575539016723633, 0.04588137435913086, 0.045980640411376957, 0.04705279922485352, 0.04674355316162109, 0.04630323028564453, 0.04589875030517578, 0.045739009857177736, 0.04578307342529297, 0.04593660736083984, 0.04589670562744141, 0.045856769561767576, 0.0457891845703125, 0.04592230224609375, 0.04583116912841797, 0.045846527099609374, 0.045841407775878903, 0.045827072143554685, 0.045886463165283206, 0.04583833694458008, 0.04581683349609375, 0.04579328155517578, 0.046015487670898435, 0.04581990432739258, 0.04590182495117188, 0.04586393737792969, 0.046063617706298826, 0.045976577758789064, 0.045889537811279295, 0.04587519836425781, 0.04592435073852539, 0.045930496215820314, 0.04589977645874024, 0.04591308975219727, 0.0459048957824707, 0.04593561553955078, 0.04588851165771484, 0.0459417610168457, 0.04599705505371094, 0.04648550415039063, 0.04619980621337891, 0.04613836669921875, 0.04597145462036133, 0.04631449508666992, 0.046298110961914066, 0.046050304412841796, 0.04593868637084961, 0.045954078674316404, 0.04616188812255859, 0.19703602600097656, 0.04583628845214844, 0.04582092666625977, 0.04577177429199219, 0.045728767395019534, 0.04566527938842774, 0.04571136093139649, 0.04575641632080078, 0.045712383270263675, 0.045870079040527346, 0.04600012969970703, 0.045753345489501954, 0.04571648025512695, 0.04587212753295898, 0.045722625732421876, 0.0457891845703125, 0.045712383270263675, 0.04588851165771484, 0.04592947387695313, 0.045835262298583986, 0.04598886489868164, 0.04639539337158203, 0.04599193572998047, 0.045843456268310545, 0.04579840087890625, 0.04579020690917969, 0.04589875030517578, 0.04587212753295898, 0.04578201675415039, 0.04582297515869141, 0.045878273010253906, 0.04591414260864258, 0.04586697769165039, 0.04586703872680664, 0.04582499313354492, 0.04634726333618164, 0.046339073181152345, 0.04588851165771484, 0.045985790252685545, 0.04598476791381836, 0.04588339233398438, 0.046045185089111325, 0.04593868637084961, 0.04605440139770508, 0.04590182495117188, 0.04598886489868164, 0.04585779190063476, 0.04600320053100586, 0.0459315185546875, 0.04594483184814453, 0.0459048957824707, 0.04600735855102539, 0.0458771858215332, 0.04599398422241211, 0.04591820907592774, 0.0459048957824707, 0.04596121597290039, 0.045954113006591794, 0.045866943359375, 0.045999103546142575, 0.046265342712402346, 0.04604415893554688, 0.04611174392700195]",tokens/s,20.77101197798444,,,main,False,False,,,,,,,,,,,,,,,,,,,,,,, float32-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/xglm-7.5B,facebook/xglm-7.5B,cuda,0,42,,,True,,,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",xglm,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-14B,Qwen/Qwen1.5-14B,cuda,0,42,,,True,,,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",qwen2,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,meta-llama/Llama-2-70b-hf,meta-llama/Llama-2-70b-hf,cuda,0,42,,,True,True,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ self.load_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 250, in load_model_with_no_weights self.pretrained_model = self.automodel_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3710, in from_pretrained model = cls(config, *model_args, **model_kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 1101, in __init__ self.model = LlamaModel(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 884, in __init__ [LlamaDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 884, in [LlamaDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 679, in __init__ self.mlp = LlamaMLP(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 193, in __init__ self.up_proj = nn.Linear(self.hidden_size, self.intermediate_size, bias=config.mlp_bias) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/linear.py"", line 98, in __init__ self.weight = Parameter(torch.empty((out_features, in_features), **factory_kwargs)) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_device.py"", line 78, in __torch_function__ return func(*args, **kwargs) torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 896.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Deci/DeciCoder-1b,Deci/DeciCoder-1b,cuda,0,42,,,True,,,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",llama,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,meta-llama/Meta-Llama-3-70B,meta-llama/Meta-Llama-3-70B,cuda,0,42,,,True,True,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ self.load_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 250, in load_model_with_no_weights self.pretrained_model = self.automodel_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3710, in from_pretrained model = cls(config, *model_args, **model_kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 1101, in __init__ self.model = LlamaModel(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 884, in __init__ [LlamaDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 884, in [LlamaDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 679, in __init__ self.mlp = LlamaMLP(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 193, in __init__ self.up_proj = nn.Linear(self.hidden_size, self.intermediate_size, bias=config.mlp_bias) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/linear.py"", line 98, in __init__ self.weight = Parameter(torch.empty((out_features, in_features), **factory_kwargs)) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_device.py"", line 78, in __torch_function__ return func(*args, **kwargs) torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 896.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,r,r,cuda,0,42,,,True,True,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/r/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669473c4-470cf07d6836099f2ae90965;34e8fe25-e865-4606-bc46-314e988603d6) Repository Not Found for url: https://huggingface.co/r/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: r is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,togethercomputer/RedPajama-INCITE-Base-7B-v0.1,togethercomputer/RedPajama-INCITE-Base-7B-v0.1,cuda,0,42,,,True,True,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ self.load_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 250, in load_model_with_no_weights self.pretrained_model = self.automodel_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3710, in from_pretrained model = cls(config, *model_args, **model_kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 1028, in __init__ self.gpt_neox = GPTNeoXModel(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 853, in __init__ self.layers = nn.ModuleList([GPTNeoXLayer(config) for _ in range(config.num_hidden_layers)]) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 853, in self.layers = nn.ModuleList([GPTNeoXLayer(config) for _ in range(config.num_hidden_layers)]) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 741, in __init__ self.mlp = GPTNeoXMLP(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 715, in __init__ self.dense_4h_to_h = nn.Linear(config.intermediate_size, config.hidden_size) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/linear.py"", line 98, in __init__ self.weight = Parameter(torch.empty((out_features, in_features), **factory_kwargs)) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_device.py"", line 78, in __torch_function__ return func(*args, **kwargs) torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 256.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-72B,Qwen/Qwen1.5-72B,cuda,0,42,,,True,,,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",qwen2,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,internlm/internlm-20b,internlm/internlm-20b,cuda,0,42,,,True,,,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",internlm,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-110B,Qwen/Qwen1.5-110B,cuda,0,42,,,True,,,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",qwen2,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Salesforce/codegen-16B-nl,Salesforce/codegen-16B-nl,cuda,0,42,,,True,,,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",codegen,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,m,m,cuda,0,42,,,True,True,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/m/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669472c6-2b77031e786b9f78671a3548;252aa2e9-2958-42e9-b788-bff74573d83f) Repository Not Found for url: https://huggingface.co/m/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: m is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,openai-community/gpt2-large,openai-community/gpt2-large,cuda,0,42,,,True,True,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1096.851456,4232.577024,0.0,3592.421376,3412.881408,s,10,0.3649756813049317,0.036497568130493166,0.0003309469851754426,0.0364003849029541,0.036552003860473636,0.03701843357086182,0.037391577339172366,"[0.03748486328125, 0.036448352813720705, 0.03641116714477539, 0.03640460968017578, 0.036356414794921875, 0.03635260772705078, 0.036411777496337894, 0.03639139175415039, 0.036318336486816406, 0.03639616012573242]",tokens/s,7014.165959899007,kWh,4.343153494927619e-07,2.3798318424510127e-07,2.9822920100147437e-06,3.6545905437526076e-06,tokens/kWh,70048886.99162835,MB,1096.851456,4232.577024,0.0,3592.421376,3508.287488,s,10,14.515923828125,1.4515923828125001,0.013427151349847092,1.4556751098632814,1.4639001831054688,1.464859454345703,1.4656268713378906,"[1.4658187255859374, 1.461234619140625, 1.461124755859375, 1.4601092529296875, 1.4474161376953125, 1.4246217041015625, 1.4503206787109375, 1.451240966796875, 1.46368701171875, 1.4303499755859375]",tokens/s,43.40061352342988,kWh,1.722910017189052e-05,9.440607064832288e-06,4.424721799578537e-05,7.091692523250816e-05,tokens/kWh,888363.3884781138,,s,629,14.818543619155893,0.02355889287624147,0.0043467260687054966,0.023227392196655275,0.02349322166442871,0.023884391021728516,0.05934923919677735,"[0.02288128089904785, 0.02275225639343262, 0.022681600570678712, 0.02254643249511719, 0.022425600051879883, 0.023121952056884765, 0.02273072052001953, 0.023595008850097656, 0.023983104705810547, 0.02451148796081543, 0.02410393524169922, 0.024000511169433594, 0.023888896942138672, 0.024184831619262694, 0.02412748718261719, 0.02414899253845215, 0.02435686492919922, 0.02390220832824707, 0.024078336715698243, 0.023766048431396486, 0.024072160720825197, 0.02410393524169922, 0.023751743316650392, 0.023645120620727537, 0.02348441505432129, 0.02325299263000488, 0.023020544052124024, 0.02308608055114746, 0.02326425552368164, 0.023013376235961915, 0.023015424728393553, 0.022996992111206056, 0.022244352340698242, 0.02234880065917969, 0.023019519805908203, 0.022960128784179686, 0.023390207290649414, 0.023043071746826172, 0.02304614448547363, 0.023006240844726564, 0.023035871505737306, 0.02307174491882324, 0.02304412841796875, 0.022934560775756837, 0.023048128128051758, 0.0225218563079834, 0.022150144577026368, 0.02203647994995117, 0.0222740478515625, 0.02216048049926758, 0.02222480010986328, 0.022338560104370117, 0.02251468849182129, 0.02445516777038574, 0.02387763214111328, 0.023282688140869142, 0.023373823165893554, 0.02313113594055176, 0.023215103149414062, 0.023570432662963867, 0.02335436820983887, 0.02344960021972656, 0.05952614212036133, 0.022562816619873048, 0.022751232147216797, 0.023327743530273438, 0.02329190444946289, 0.02326425552368164, 0.023327743530273438, 0.02329497528076172, 0.023360544204711915, 0.02251055908203125, 0.02356332778930664, 0.023563199996948243, 0.023540735244750977, 0.023344127655029297, 0.02329702377319336, 0.023359487533569336, 0.023068672180175782, 0.02351103973388672, 0.02348236846923828, 0.023430143356323242, 0.02347417640686035, 0.02249728012084961, 0.02327859115600586, 0.023326719284057617, 0.02350694465637207, 0.023010303497314453, 0.022588415145874022, 0.022502399444580077, 0.022743040084838868, 0.023343103408813477, 0.023480319976806642, 0.023508991241455078, 0.02346598434448242, 0.023375871658325196, 0.023396352767944335, 0.02329599952697754, 0.023555103302001952, 0.023255008697509766, 0.023524351119995117, 0.023367679595947266, 0.023319583892822265, 0.023275487899780272, 0.02246659278869629, 0.02255356788635254, 0.02235699272155762, 0.02285977554321289, 0.02346905517578125, 0.023374847412109375, 0.023560192108154295, 0.02329292869567871, 0.023379968643188476, 0.023383039474487305, 0.02330828857421875, 0.02328166389465332, 0.023441408157348635, 0.023188480377197264, 0.023159807205200195, 0.02328678321838379, 0.02342092704772949, 0.023648351669311524, 0.023137311935424804, 0.022210432052612306, 0.0224901123046875, 0.059510784149169924, 0.022571008682250978, 0.022483968734741212, 0.02286796760559082, 0.022509567260742186, 0.022591487884521484, 0.022525951385498046, 0.022518783569335937, 0.022478847503662108, 0.022837247848510742, 0.023627775192260742, 0.023614463806152345, 0.023334911346435547, 0.023438335418701172, 0.023254016876220703, 0.023340032577514647, 0.023358495712280273, 0.023459808349609375, 0.02332057571411133, 0.023230464935302734, 0.023164928436279295, 0.02326016044616699, 0.02305023956298828, 0.02330419158935547, 0.022966272354125978, 0.023034879684448242, 0.023166976928710937, 0.023307264328002928, 0.02239897537231445, 0.02326016044616699, 0.023249919891357423, 0.023378944396972655, 0.023540735244750977, 0.023436288833618164, 0.023203840255737306, 0.023385087966918947, 0.023258111953735353, 0.02335436820983887, 0.023851007461547852, 0.023214080810546874, 0.023274496078491212, 0.023455743789672853, 0.02347929573059082, 0.02325299263000488, 0.022434816360473633, 0.023219232559204103, 0.02328265571594238, 0.023315456390380858, 0.023307264328002928, 0.023427072525024413, 0.023250944137573244, 0.0233175048828125, 0.023371776580810546, 0.023331840515136718, 0.023333887100219726, 0.023388160705566406, 0.02346188735961914, 0.023319551467895508, 0.02322332763671875, 0.023641056060791015, 0.02329804801940918, 0.023224319458007812, 0.02327244758605957, 0.05942988967895508, 0.022459392547607423, 0.02329395294189453, 0.023249919891357423, 0.02330624008178711, 0.023212032318115236, 0.023221248626708983, 0.023216127395629883, 0.023397375106811523, 0.023451648712158202, 0.023327743530273438, 0.024129568099975587, 0.023498720169067382, 0.023340032577514647, 0.023241727828979493, 0.023318527221679687, 0.023330848693847658, 0.023636959075927735, 0.02325299263000488, 0.023266304016113282, 0.023370752334594725, 0.022541311264038084, 0.022608959197998046, 0.023261119842529297, 0.023544832229614256, 0.023975936889648438, 0.024400896072387695, 0.02347315216064453, 0.023459840774536132, 0.02332569694519043, 0.023267328262329103, 0.02338822364807129, 0.023450559616088867, 0.02334617614746094, 0.023158784866333007, 0.02307072067260742, 0.023355392456054686, 0.023394304275512694, 0.02326425552368164, 0.023235584259033205, 0.022496288299560546, 0.022566879272460937, 0.022466560363769532, 0.02253926467895508, 0.02255462455749512, 0.022611967086791994, 0.02246143913269043, 0.022441984176635742, 0.02290278434753418, 0.02334617614746094, 0.023208959579467774, 0.02332057571411133, 0.023237632751464843, 0.023250944137573244, 0.023451648712158202, 0.022961151123046874, 0.023068672180175782, 0.023021568298339845, 0.02304614448547363, 0.023053312301635744, 0.0228853759765625, 0.023130111694335938, 0.02306252861022949, 0.05904383850097656, 0.02242355155944824, 0.022207584381103516, 0.02213177680969238, 0.02237628746032715, 0.022096000671386718, 0.022404991149902343, 0.022170719146728517, 0.022246400833129884, 0.022127519607543944, 0.02279542350769043, 0.022848480224609374, 0.023214975357055665, 0.02305023956298828, 0.023005184173583985, 0.022940736770629883, 0.02300921630859375, 0.023068672180175782, 0.022994943618774414, 0.02293769645690918, 0.022969280242919922, 0.02298569679260254, 0.02333497619628906, 0.02337273597717285, 0.02307072067260742, 0.023340032577514647, 0.023343103408813477, 0.023374847412109375, 0.023335935592651368, 0.023249919891357423, 0.023405567169189453, 0.023350271224975586, 0.022451200485229493, 0.02288332748413086, 0.023266304016113282, 0.023228416442871092, 0.024285184860229493, 0.023181312561035155, 0.02322329521179199, 0.023397375106811523, 0.023203840255737306, 0.023237632751464843, 0.02332057571411133, 0.023216127395629883, 0.02325196838378906, 0.02349567985534668, 0.02325503921508789, 0.022435840606689454, 0.02246143913269043, 0.022390783309936522, 0.022553600311279298, 0.023172096252441408, 0.022529024124145508, 0.022641664505004884, 0.0224849910736084, 0.023221248626708983, 0.023225343704223633, 0.023177215576171875, 0.02327859115600586, 0.023206911087036132, 0.023234560012817384, 0.02333798408508301, 0.023323648452758788, 0.05924454498291016, 0.022536191940307617, 0.022380544662475587, 0.022583295822143554, 0.022562816619873048, 0.022594560623168947, 0.02230988883972168, 0.022129663467407225, 0.022397951126098634, 0.02249318313598633, 0.022549503326416014, 0.023227392196655275, 0.023350271224975586, 0.02335955238342285, 0.023177152633666993, 0.022551551818847656, 0.022416383743286132, 0.023818239212036133, 0.023230464935302734, 0.02305536079406738, 0.02229862403869629, 0.022541311264038084, 0.022366207122802736, 0.022153215408325197, 0.022152191162109376, 0.022214656829833986, 0.022157312393188477, 0.022167552947998048, 0.022320127487182616, 0.022459392547607423, 0.02249318313598633, 0.02245427131652832, 0.022360063552856444, 0.022466560363769532, 0.02269593620300293, 0.023018495559692383, 0.023190528869628906, 0.022244352340698242, 0.022203392028808593, 0.022404096603393556, 0.022373376846313478, 0.02247065544128418, 0.022358015060424806, 0.022452224731445314, 0.022193151473999022, 0.02228121566772461, 0.022413312911987306, 0.02239897537231445, 0.022424575805664062, 0.022395904541015626, 0.022410240173339844, 0.022399999618530272, 0.022382591247558595, 0.022420480728149415, 0.02248806381225586, 0.022451200485229493, 0.022353919982910156, 0.02347520065307617, 0.023407615661621094, 0.02328166389465332, 0.023193599700927735, 0.02329395294189453, 0.023342111587524413, 0.0593919677734375, 0.02245427131652832, 0.022592544555664062, 0.022461408615112304, 0.022537216186523438, 0.022466560363769532, 0.022375423431396483, 0.022362112045288086, 0.022477823257446287, 0.02242355155944824, 0.022374399185180666, 0.022681600570678712, 0.02254643249511719, 0.022707199096679686, 0.02343731117248535, 0.023394304275512694, 0.023352319717407227, 0.023385087966918947, 0.024175615310668946, 0.023923776626586915, 0.0233768310546875, 0.02329497528076172, 0.02332569694519043, 0.023352319717407227, 0.02328166389465332, 0.023377920150756838, 0.023350271224975586, 0.023371776580810546, 0.02334617614746094, 0.023439359664916993, 0.023444480895996093, 0.02392678451538086, 0.02374963188171387, 0.023362560272216795, 0.023367679595947266, 0.02354380798339844, 0.02328473663330078, 0.023433216094970705, 0.022550527572631835, 0.022517759323120116, 0.02247372817993164, 0.022569984436035157, 0.022561792373657227, 0.02294169616699219, 0.02345369529724121, 0.023561216354370116, 0.02329599952697754, 0.02327859115600586, 0.022492160797119142, 0.022572032928466795, 0.02243071937561035, 0.022605823516845702, 0.022433792114257813, 0.022541311264038084, 0.022586368560791017, 0.022592512130737305, 0.02251263999938965, 0.02308403205871582, 0.023407615661621094, 0.023379968643188476, 0.023235584259033205, 0.0233175048828125, 0.02323865509033203, 0.059737087249755856, 0.023373823165893554, 0.023387136459350585, 0.023330816268920897, 0.02328678321838379, 0.023358463287353515, 0.023469087600708007, 0.023321567535400392, 0.02332057571411133, 0.022898687362670898, 0.023198720932006835, 0.02285875129699707, 0.02232524871826172, 0.02244812774658203, 0.02326527976989746, 0.023330816268920897, 0.02326835250854492, 0.02327347183227539, 0.02346905517578125, 0.023371776580810546, 0.023331872940063475, 0.023452640533447266, 0.02345779228210449, 0.0233175048828125, 0.02349260711669922, 0.0233123836517334, 0.02337696075439453, 0.023374784469604493, 0.023343103408813477, 0.023323648452758788, 0.023326719284057617, 0.022623231887817383, 0.023302143096923827, 0.023359487533569336, 0.023408639907836915, 0.02347007942199707, 0.024239103317260743, 0.02292531204223633, 0.02261299133300781, 0.0224901123046875, 0.02247475242614746, 0.022410240173339844, 0.022607872009277344, 0.02244915199279785, 0.02244095993041992, 0.022619136810302733, 0.022544384002685547, 0.022450176239013672, 0.02243174362182617, 0.022424575805664062, 0.022529024124145508, 0.02247372817993164, 0.02249932861328125, 0.022384639739990234, 0.022845439910888672, 0.023162879943847657, 0.02305843162536621, 0.022917247772216796, 0.023056255340576173, 0.022999040603637694, 0.02330316734313965, 0.02324787139892578, 0.02251468849182129, 0.05938995361328125, 0.022578176498413087, 0.02250547218322754, 0.022890495300292968, 0.02264678382873535, 0.02244095993041992, 0.023334911346435547, 0.023129087448120117, 0.02305433654785156, 0.022922239303588866, 0.022981632232666017, 0.023180288314819338, 0.023242752075195314, 0.023364608764648437, 0.02330112075805664, 0.023562240600585937, 0.023383039474487305, 0.023371776580810546, 0.023258111953735353, 0.023217151641845703, 0.02334617614746094, 0.023407615661621094, 0.023344127655029297, 0.023360544204711915, 0.023250911712646486, 0.02330931282043457, 0.023378944396972655, 0.0232806396484375, 0.023273536682128906, 0.02333791923522949, 0.023228416442871092, 0.023237632751464843, 0.02327244758605957, 0.023369728088378908, 0.02325606346130371, 0.02346188735961914, 0.023410688400268553, 0.023331840515136718, 0.023330816268920897, 0.023342079162597656, 0.02327244758605957, 0.023393280029296876, 0.023209983825683594, 0.023190528869628906, 0.02304614448547363, 0.023415807723999024, 0.023358463287353515, 0.023282688140869142, 0.02332876777648926, 0.02344550323486328, 0.02327756881713867, 0.023282688140869142, 0.02328371238708496, 0.023376895904541017, 0.023267328262329103, 0.02326937675476074, 0.023311359405517578, 0.023069696426391603, 0.023537664413452147, 0.023304224014282227, 0.023196640014648436, 0.023369728088378908, 0.02325503921508789, 0.05949849700927735, 0.022597631454467772, 0.022517759323120116, 0.022434816360473633, 0.022519807815551757, 0.02245427131652832, 0.022428672790527345, 0.02254643249511719, 0.02249728012084961, 0.02245734405517578, 0.02244710350036621, 0.022556671142578123, 0.022494207382202147, 0.02253926467895508, 0.02253209686279297, 0.02310860824584961, 0.02330828857421875, 0.023352319717407227, 0.023373823165893554, 0.023418880462646483, 0.023426048278808592, 0.023385087966918947, 0.02331648063659668, 0.023435264587402343, 0.02326118469238281, 0.02332569694519043, 0.02328883171081543, 0.023433216094970705, 0.02247065544128418, 0.022541311264038084, 0.022541311264038084, 0.022486015319824217, 0.0224399356842041, 0.022565887451171874, 0.02250649642944336, 0.022583295822143554, 0.022391807556152343, 0.023299072265625, 0.02320604705810547, 0.023158624649047853, 0.024070144653320313, 0.023150592803955077, 0.022124544143676757, 0.022082559585571288, 0.022459392547607423, 0.02225663948059082, 0.02215936088562012, 0.022177791595458983, 0.022200319290161134, 0.022305791854858398, 0.022238208770751954, 0.022232063293457033, 0.02217692756652832, 0.02234761619567871, 0.022189056396484375, 0.022125568389892578, 0.022130687713623046, 0.02227097511291504, 0.0221409912109375, 0.022715328216552734, 0.02311577606201172, 0.023028736114501954, 0.022943744659423827]",tokens/s,42.446816378560555,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,huggyllama/llama-65b,huggyllama/llama-65b,cuda,0,42,,,True,,,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",llama,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,l,l,cuda,0,42,,,True,True,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/l/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6694743a-0bc35fda54c4d66a36e78186;6440c2b8-0774-4472-a7ba-faa47f03beae) Repository Not Found for url: https://huggingface.co/l/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: l is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,B,B,cuda,0,42,,,True,True,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/B/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669324ee-3eaaff71125624b317e58bc4;d1eb9337-feff-434a-83b2-114c3ced91db) Repository Not Found for url: https://huggingface.co/B/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: B is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,huggyllama/llama-30b,huggyllama/llama-30b,cuda,0,42,,,True,,,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",llama,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,meta-llama/Llama-2-13b-hf,meta-llama/Llama-2-13b-hf,cuda,0,42,,,True,True,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ self.load_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 250, in load_model_with_no_weights self.pretrained_model = self.automodel_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3710, in from_pretrained model = cls(config, *model_args, **model_kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 1101, in __init__ self.model = LlamaModel(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 884, in __init__ [LlamaDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 884, in [LlamaDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 677, in __init__ self.self_attn = LLAMA_ATTENTION_CLASSES[config._attn_implementation](config=config, layer_idx=layer_idx) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 262, in __init__ self.q_proj = nn.Linear(self.hidden_size, self.num_heads * self.head_dim, bias=config.attention_bias) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/linear.py"", line 98, in __init__ self.weight = Parameter(torch.empty((out_features, in_features), **factory_kwargs)) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_device.py"", line 78, in __torch_function__ return func(*args, **kwargs) torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 100.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-7B,Qwen/Qwen1.5-7B,cuda,0,42,,,True,,,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",qwen2,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,M,M,cuda,0,42,,,True,True,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/M/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66932282-01e1497b3c91b1bc33aeef0b;c846c8ad-8b37-4c31-b7cb-60ac2e5fa191) Repository Not Found for url: https://huggingface.co/M/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: M is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,stabilityai/stablelm-base-alpha-3b,stabilityai/stablelm-base-alpha-3b,cuda,0,42,,,True,True,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1134.391296,16014.37696,0.0,15374.221312,15168.105472,s,10,1.122411491394043,0.1122411491394043,0.0003948758294262617,0.11217387008666992,0.1127305793762207,0.11285148048400878,0.11294820137023925,"[0.11262064361572266, 0.11157008361816406, 0.11205372619628906, 0.1121588134765625, 0.11204422760009766, 0.11222179412841797, 0.11218892669677734, 0.11187718200683594, 0.1127037124633789, 0.11297238159179687]",tokens/s,2280.8034483150755,kWh,1.3559955256906425e-06,7.428953467751853e-07,9.514183747704476e-06,1.1613074620170305e-05,tokens/kWh,22044119.096192095,MB,1134.71488,16014.37696,0.0,15374.221312,15300.094464,s,10,20.30970324707031,2.030970324707031,0.0029506713849007135,2.0304553833007812,2.035147399902344,2.035291046142578,2.0354059631347656,"[2.0304290771484377, 2.033796630859375, 2.02862548828125, 2.0284708251953125, 2.0284163818359375, 2.030481689453125, 2.02641357421875, 2.035115478515625, 2.0325194091796877, 2.0354346923828124]",tokens/s,31.019655596931383,kWh,2.3825207097888595e-05,1.3054617339857938e-05,0.00010378676800409741,0.00014066659244184394,tokens/kWh,447867.5349020501,,s,629,21.28694987106322,0.03384252761695268,0.013402911117138075,0.032228351593017575,0.03236045379638672,0.032398950195312505,0.14482880981445315,"[0.03212492752075195, 0.03204915237426758, 0.03203891372680664, 0.032007167816162106, 0.032086017608642575, 0.0324659194946289, 0.03236662292480469, 0.03206857681274414, 0.032039936065673826, 0.032064510345458985, 0.03210137557983399, 0.032054271697998044, 0.032059391021728514, 0.0320706558227539, 0.03250688171386719, 0.03282944107055664, 0.03240140914916992, 0.03215052795410156, 0.032115711212158206, 0.0320819206237793, 0.03210854339599609, 0.03210137557983399, 0.03207167816162109, 0.03214748764038086, 0.03214742279052735, 0.032130046844482424, 0.032129024505615236, 0.032129024505615236, 0.03217203140258789, 0.032123905181884765, 0.032176128387451174, 0.03223961639404297, 0.03214233779907227, 0.032126976013183595, 0.03216588973999023, 0.0321003532409668, 0.0322529296875, 0.03216691207885742, 0.0321638412475586, 0.03222016143798828, 0.032159744262695314, 0.032148479461669925, 0.032115711212158206, 0.03217510223388672, 0.03222937774658203, 0.032254974365234376, 0.032210945129394535, 0.032290817260742184, 0.03232767868041992, 0.03225600051879883, 0.03223772811889648, 0.032376670837402345, 0.032176128387451174, 0.03225600051879883, 0.03230310440063477, 0.03253452682495117, 0.03239833450317383, 0.03294822311401367, 0.03236249542236328, 0.03233280181884766, 0.03235027313232422, 0.032409534454345704, 0.1444638671875, 0.03223056030273438, 0.032188255310058596, 0.032217086791992186, 0.0321341438293457, 0.03217510223388672, 0.032194561004638675, 0.03218329620361328, 0.03226521682739258, 0.03226726531982422, 0.032198654174804685, 0.03226217651367187, 0.032289760589599606, 0.03225395202636719, 0.03230515289306641, 0.03230822372436523, 0.03227852630615234, 0.032282623291015625, 0.03224063873291016, 0.03230310440063477, 0.03237171173095703, 0.03234406280517578, 0.0321976318359375, 0.032307201385498044, 0.03222118377685547, 0.032284671783447266, 0.03231129455566406, 0.03223759841918945, 0.032311264038085936, 0.032289886474609376, 0.03228355026245117, 0.03230310440063477, 0.032304126739501955, 0.03217715072631836, 0.03222016143798828, 0.03224063873291016, 0.032194561004638675, 0.03227238464355469, 0.0323768310546875, 0.032188480377197265, 0.03227436828613281, 0.03232665634155273, 0.03233792114257812, 0.032317440032958986, 0.03227852630615234, 0.03232876968383789, 0.03231123352050781, 0.03226521682739258, 0.03238195037841797, 0.03233385467529297, 0.032369663238525394, 0.032360416412353515, 0.03237785720825195, 0.03242496109008789, 0.032315391540527344, 0.03231948852539063, 0.03223961639404297, 0.03231235122680664, 0.03222320175170899, 0.03223551940917969, 0.03220889663696289, 0.032189441680908204, 0.03221913528442383, 0.14497074890136719, 0.03224576187133789, 0.032061439514160156, 0.03205836868286133, 0.03209625625610352, 0.032074752807617186, 0.032064510345458985, 0.03206041717529297, 0.032146430969238284, 0.03212799835205078, 0.032075775146484374, 0.032159744262695314, 0.03224892807006836, 0.0321995849609375, 0.032215038299560544, 0.03216486358642578, 0.032216064453125, 0.03235532760620117, 0.032312320709228515, 0.032210945129394535, 0.03228160095214844, 0.03223756790161133, 0.0321003532409668, 0.03215769577026367, 0.032094207763671875, 0.03206883239746094, 0.03206224060058594, 0.032092159271240234, 0.03209625625610352, 0.03209318542480469, 0.032156673431396485, 0.03216486358642578, 0.0321701431274414, 0.03212374496459961, 0.032123905181884765, 0.03213516616821289, 0.03215155029296875, 0.03213107299804688, 0.03212287902832031, 0.032146430969238284, 0.032092159271240234, 0.032215038299560544, 0.032568321228027344, 0.032312320709228515, 0.032123966217041014, 0.03214944076538086, 0.032189441680908204, 0.032215038299560544, 0.032233470916748046, 0.03213619232177734, 0.032231422424316404, 0.03215462493896484, 0.03218022537231445, 0.03224063873291016, 0.032233470916748046, 0.03246182250976563, 0.03237376022338867, 0.0323133430480957, 0.032261119842529294, 0.032279552459716795, 0.0323164176940918, 0.032315391540527344, 0.032315391540527344, 0.1451724853515625, 0.032418815612792966, 0.03222323226928711, 0.032302078247070314, 0.03228876876831055, 0.03225600051879883, 0.03223551940917969, 0.03220991897583008, 0.03218739318847656, 0.032130046844482424, 0.032143360137939454, 0.032097278594970705, 0.03220889663696289, 0.032132095336914065, 0.032205825805664064, 0.03215155029296875, 0.032132095336914065, 0.032121856689453124, 0.032205825805664064, 0.03214131164550781, 0.03218124771118164, 0.032194561004638675, 0.03212595367431641, 0.032148479461669925, 0.03219353485107422, 0.03216281509399414, 0.032130046844482424, 0.0321341438293457, 0.03212083053588867, 0.03215564727783203, 0.03216691207885742, 0.032145408630371096, 0.03215769577026367, 0.032143360137939454, 0.032105472564697264, 0.03211980819702148, 0.03209830474853516, 0.0321710090637207, 0.03212083053588867, 0.03218431854248047, 0.03214950561523437, 0.03216998291015625, 0.032198654174804685, 0.032198654174804685, 0.03223244857788086, 0.032189441680908204, 0.0321976318359375, 0.032290817260742184, 0.032246784210205076, 0.032178272247314454, 0.03213711929321289, 0.0321475830078125, 0.032159614562988284, 0.03223756790161133, 0.032215198516845706, 0.032248672485351564, 0.032178176879882815, 0.03218329620361328, 0.03218739318847656, 0.032228351593017575, 0.032261119842529294, 0.03220275115966797, 0.03229183959960937, 0.14486834716796876, 0.03218329620361328, 0.03208806228637695, 0.03209625625610352, 0.032102401733398435, 0.03208703994750976, 0.03203891372680664, 0.032320510864257815, 0.03220684814453125, 0.0320819206237793, 0.03208294296264649, 0.03214745712280274, 0.03205740737915039, 0.032132030487060544, 0.0321003532409668, 0.0321003532409668, 0.03216691207885742, 0.032150623321533206, 0.03205519866943359, 0.032099327087402346, 0.0320634880065918, 0.03208294296264649, 0.032099327087402346, 0.032102401733398435, 0.03206758499145508, 0.0320634880065918, 0.03213516616821289, 0.03208703994750976, 0.03206553649902344, 0.032061439514160156, 0.032110591888427735, 0.032107521057128906, 0.0321003532409668, 0.032118785858154295, 0.03211471939086914, 0.03214640045166016, 0.03212799835205078, 0.0322426872253418, 0.03234918212890625, 0.03214863967895508, 0.032174945831298825, 0.03213619232177734, 0.032347137451171876, 0.03211775970458984, 0.03210137557983399, 0.03213516616821289, 0.032115711212158206, 0.032261119842529294, 0.032165920257568356, 0.03214128112792969, 0.03239731216430664, 0.03268307113647461, 0.03255187225341797, 0.032399360656738284, 0.03232972717285156, 0.03237478256225586, 0.03237283325195312, 0.032323486328125, 0.03236556625366211, 0.032314369201660156, 0.03236767959594727, 0.03237267303466797, 0.032353279113769534, 0.14510592651367188, 0.03230515289306641, 0.032146430969238284, 0.032315391540527344, 0.03225702285766602, 0.032254974365234376, 0.03218739318847656, 0.03228979110717774, 0.03226726531982422, 0.0322979850769043, 0.0322949104309082, 0.03234201431274414, 0.03218636703491211, 0.032317440032958986, 0.03222220611572266, 0.032246784210205076, 0.03223859024047852, 0.03230822372436523, 0.03222220611572266, 0.032244735717773435, 0.03223244857788086, 0.03230310440063477, 0.03236249542236328, 0.03225804901123047, 0.032254974365234376, 0.03232563018798828, 0.03231846237182617, 0.032307201385498044, 0.032363521575927735, 0.03227340698242188, 0.03225804901123047, 0.03231846237182617, 0.03232153701782227, 0.0322949104309082, 0.03235027313232422, 0.03222828674316406, 0.03212083053588867, 0.03213107299804688, 0.03214131164550781, 0.03218431854248047, 0.0321525764465332, 0.032173057556152344, 0.03211775970458984, 0.0321607666015625, 0.03213619232177734, 0.03215155029296875, 0.03218022537231445, 0.03215564727783203, 0.032143360137939454, 0.0321710090637207, 0.03212595367431641, 0.03214233779907227, 0.032161792755126956, 0.032161792755126956, 0.03213721466064453, 0.03211775970458984, 0.032115711212158206, 0.03211673736572265, 0.032176128387451174, 0.03214438247680664, 0.032143360137939454, 0.03211264038085938, 0.03218534469604492, 0.1450250244140625, 0.03220377731323242, 0.03203583908081055, 0.03205836868286133, 0.03205017471313477, 0.03208499145507813, 0.032054271697998044, 0.03204915237426758, 0.03204915237426758, 0.032099327087402346, 0.03209318542480469, 0.03209830474853516, 0.03207987213134766, 0.032080894470214845, 0.03208806228637695, 0.03205836868286133, 0.032039936065673826, 0.032110591888427735, 0.032102401733398435, 0.03205120086669922, 0.03208806228637695, 0.03211468887329102, 0.032075775146484374, 0.032091136932373046, 0.03210137557983399, 0.032056320190429685, 0.03207884979248047, 0.032080894470214845, 0.032140289306640625, 0.03212799835205078, 0.03224166488647461, 0.03229183959960937, 0.03211468887329102, 0.03212799835205078, 0.03223654556274414, 0.03215462493896484, 0.03211264038085938, 0.03215052795410156, 0.03212287902832031, 0.03214950561523437, 0.032102401733398435, 0.03209625625610352, 0.03215155029296875, 0.03215871810913086, 0.0324771842956543, 0.03237580871582031, 0.03221299362182617, 0.03215769577026367, 0.03219148635864258, 0.03217919921875, 0.032142398834228515, 0.032127937316894534, 0.03214233779907227, 0.03218636703491211, 0.03222732925415039, 0.032210945129394535, 0.032312320709228515, 0.03234204864501953, 0.032376800537109375, 0.03233187103271484, 0.032290721893310545, 0.03231948852539063, 0.03233792114257812, 0.14472714233398437, 0.032285598754882815, 0.03213824081420898, 0.03219353485107422, 0.03212492752075195, 0.032205825805664064, 0.03214438247680664, 0.032210945129394535, 0.03226214218139648, 0.03217715072631836, 0.032230400085449216, 0.032224353790283204, 0.03225078582763672, 0.03219968032836914, 0.032159744262695314, 0.03215462493896484, 0.03226521682739258, 0.032241825103759766, 0.03219030380249024, 0.03236044692993164, 0.032508926391601564, 0.03236249542236328, 0.03223756790161133, 0.03226828765869141, 0.032301055908203126, 0.032304126739501955, 0.0323164176940918, 0.032330753326416016, 0.032312320709228515, 0.03225907135009766, 0.03234304046630859, 0.03225600051879883, 0.03229695892333984, 0.032276481628417966, 0.03225190353393555, 0.03225190353393555, 0.03237376022338867, 0.03234611129760742, 0.03235334396362305, 0.032376766204833984, 0.032350208282470705, 0.03227238464355469, 0.032287776947021486, 0.03231638336181641, 0.032353279113769534, 0.032307201385498044, 0.032216064453125, 0.03233792114257812, 0.03235430526733398, 0.032322559356689456, 0.032510974884033206, 0.03268505477905274, 0.03233996963500976, 0.03237478256225586, 0.03232153701782227, 0.0322949104309082, 0.03237273788452148, 0.032355457305908206, 0.032306049346923826, 0.032331775665283204, 0.03231129455566406, 0.032304126739501955, 0.032330753326416016, 0.14555238342285157, 0.03220479965209961, 0.03216998291015625, 0.032198654174804685, 0.03216998291015625, 0.0321525764465332, 0.03218022537231445, 0.032046112060546875, 0.03206550216674805, 0.03216486358642578, 0.032078880310058594, 0.03220169448852539, 0.03213836669921875, 0.03225075149536133, 0.03226214218139648, 0.032129024505615236, 0.03225190353393555, 0.03230003356933594, 0.03220172882080078, 0.032254974365234376, 0.032317440032958986, 0.03227545547485351, 0.03220076751708984, 0.03221907043457031, 0.0321607666015625, 0.03222732925415039, 0.032295936584472655, 0.032312320709228515, 0.03226931381225586, 0.0322949104309082, 0.032290817260742184, 0.03223654556274414, 0.032239646911621095, 0.032292831420898435, 0.032295936584472655, 0.03226931381225586, 0.03226214218139648, 0.03232767868041992, 0.032234527587890624, 0.032204769134521485, 0.032200702667236326, 0.03230620956420899, 0.03230511856079102, 0.03230003356933594, 0.03234611129760742, 0.03230515289306641, 0.03232563018798828, 0.032312320709228515, 0.032284671783447266, 0.03230003356933594, 0.03234611129760742, 0.03227958297729492, 0.03225289535522461, 0.03236454391479492, 0.03229695892333984, 0.03232972717285156, 0.03235635375976562, 0.03233484649658203, 0.032277503967285154, 0.03235942459106445, 0.032284671783447266, 0.032279552459716795, 0.03230822372436523, 0.14584626770019532, 0.032304126739501955, 0.03225395202636719, 0.032246784210205076, 0.03227443313598633, 0.03227033615112305, 0.03213721466064453, 0.032200702667236326, 0.03223244857788086, 0.03220275115966797, 0.03218739318847656, 0.0322529296875, 0.032215038299560544, 0.03219558334350586, 0.03229388809204101, 0.03226009750366211, 0.0324136962890625, 0.032379905700683595, 0.032335968017578126, 0.03225487899780274, 0.0322979850769043, 0.03225395202636719, 0.032307201385498044, 0.032271358489990236, 0.03224787139892578, 0.032250816345214844, 0.03225600051879883, 0.03218534469604492, 0.03250688171386719, 0.032553985595703126, 0.032290817260742184, 0.0323061752319336, 0.032285694122314454, 0.03227545547485351, 0.03227238464355469, 0.03227443313598633, 0.032317440032958986, 0.03228876876831055, 0.032328704833984374, 0.032263168334960936, 0.03234918212890625, 0.03225804901123047, 0.032382976531982424, 0.03232153701782227, 0.03232767868041992, 0.032328704833984374, 0.03226214218139648, 0.03236048126220703, 0.03241571044921875, 0.03238092803955078, 0.032314369201660156, 0.032379905700683595, 0.032350208282470705, 0.032287742614746096, 0.03232972717285156, 0.03230003356933594, 0.03229388809204101, 0.03240345764160156, 0.03233894348144531, 0.032322559356689456, 0.03235737609863281, 0.032334880828857424, 0.03226723098754883]",tokens/s,29.548620342975553,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,google/recurrentgemma-9b,google/recurrentgemma-9b,cuda,0,42,,,True,,,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",recurrent_gemma,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,mistralai/Mixtral-8x22B-v0.1,mistralai/Mixtral-8x22B-v0.1,cuda,0,42,,,True,True,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/base.py"", line 43, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 54, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1021, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 403 Client Error: Forbidden for url: https://huggingface.co/mistralai/Mixtral-8x22B-v0.1/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 398, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 321, in hf_raise_for_status raise GatedRepoError(message, response) from e huggingface_hub.utils._errors.GatedRepoError: 403 Client Error. (Request ID: Root=1-66475068-1135bdb9071d6a0f06e52ac2;d08b57b2-cc1c-4195-85ab-6c20c3f43c7e) Cannot access gated repo for url https://huggingface.co/mistralai/Mixtral-8x22B-v0.1/resolve/main/config.json. Access to model mistralai/Mixtral-8x22B-v0.1 is restricted and you are not in the authorized list. Visit https://huggingface.co/mistralai/Mixtral-8x22B-v0.1 to ask for access. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 87, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/base.py"", line 56, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 37, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 62, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 928, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 631, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 686, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 416, in cached_file raise EnvironmentError( OSError: You are trying to access a gated repo. Make sure to have access to it at https://huggingface.co/mistralai/Mixtral-8x22B-v0.1. 403 Client Error. (Request ID: Root=1-66475068-1135bdb9071d6a0f06e52ac2;d08b57b2-cc1c-4195-85ab-6c20c3f43c7e) Cannot access gated repo for url https://huggingface.co/mistralai/Mixtral-8x22B-v0.1/resolve/main/config.json. Access to model mistralai/Mixtral-8x22B-v0.1 is restricted and you are not in the authorized list. Visit https://huggingface.co/mistralai/Mixtral-8x22B-v0.1 to ask for access. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,01-ai/Yi-6B,01-ai/Yi-6B,cuda,0,42,,,True,,,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking backend.load() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load self.load_transformers_model() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 151, in load_transformers_model self.load_transformers_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 126, in load_transformers_model_with_no_weights self.pretrained_model = self.automodel_loader.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3832, in from_pretrained model = cls(config, *model_args, **model_kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 1118, in __init__ self.lm_head = nn.Linear(config.hidden_size, config.vocab_size, bias=False) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/linear.py"", line 99, in __init__ self.weight = Parameter(torch.empty((out_features, in_features), **factory_kwargs)) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_device.py"", line 79, in __torch_function__ return func(*args, **kwargs) torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 1000.00 MiB. GPU 0 has a total capacity of 22.18 GiB of which 170.50 MiB is free. Process 19965 has 22.01 GiB memory in use. Of the allocated memory 21.60 GiB is allocated by PyTorch, and 128.97 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) ",llama,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,/,/,cuda,0,42,,,True,True,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 373, in cached_file raise EnvironmentError( OSError: / does not appear to have a file named config.json. Checkout 'https://huggingface.co///tree/None' for available files. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,TencentARC/Mistral_Pro_8B_v0.1,TencentARC/Mistral_Pro_8B_v0.1,cuda,0,42,,,True,,,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",mistral,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-160m,EleutherAI/pythia-160m,cuda,0,42,,,True,True,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1146.056704,1544.02816,0.0,903.872512,809.756672,s,16,0.18308800220489505,0.011443000137805939,0.00042258731502986183,0.011342448234558105,0.011977904319763184,0.012453376293182373,0.012548070430755615,"[0.012413920402526856, 0.011084768295288086, 0.01131868839263916, 0.011341567993164062, 0.011343328475952149, 0.011293824195861817, 0.01143932819366455, 0.010996895790100098, 0.011011327743530273, 0.011301952362060547, 0.01138758373260498, 0.011541888236999512, 0.011395872116088867, 0.012571743965148926, 0.011372544288635255, 0.011272768020629882]",tokens/s,22371.755389062244,kWh,1.329379381350973e-07,7.284346626370683e-08,6.004332387514018e-07,8.062146431502059e-07,tokens/kWh,317533304.77812296,MB,1146.056704,1544.02816,0.0,903.872512,828.946432,s,16,10.01527850341797,0.6259549064636231,0.007251981997644968,0.6277228698730468,0.6305034179687501,0.6340920562744141,0.6402375518798828,"[0.64177392578125, 0.6235167846679688, 0.6279869384765625, 0.6288497924804688, 0.628356201171875, 0.6281631469726563, 0.6274588012695312, 0.6217149658203125, 0.6065394287109375, 0.61642919921875, 0.6211452026367188, 0.6315314331054688, 0.626811767578125, 0.6294754028320313, 0.6273834838867187, 0.6281420288085937]",tokens/s,100.64622762671995,kWh,7.163375407297186e-06,3.924160038128976e-06,1.3517895049366748e-05,2.4605430494792912e-05,tokens/kWh,2560410.3944993885,,s,1007,10.17525146198273,0.010104519823220186,0.0014488776123283977,0.009955327987670898,0.010065101051330566,0.010168115139007568,0.021698180274963376,"[0.010592255592346191, 0.010226688385009765, 0.010169343948364258, 0.010157055854797363, 0.01012019157409668, 0.010145792007446289, 0.010078207969665527, 0.010091520309448243, 0.010112000465393066, 0.010189824104309082, 0.010500096321105956, 0.010062848091125488, 0.010157055854797363, 0.010165247917175293, 0.01022054386138916, 0.010160127639770507, 0.010142720222473145, 0.010193920135498047, 0.010002431869506835, 0.010025983810424804, 0.010088447570800782, 0.010348544120788575, 0.010058752059936523, 0.010405983924865723, 0.010373023986816407, 0.010239999771118164, 0.01011302375793457, 0.010247167587280273, 0.010119168281555176, 0.010106880187988282, 0.01034444808959961, 0.010066944122314453, 0.010233856201171876, 0.010059776306152344, 0.010084351539611817, 0.0101212158203125, 0.010158080101013184, 0.010066944122314453, 0.0101396484375, 0.010617856025695802, 0.010247167587280273, 0.010148863792419433, 0.010108927726745605, 0.010098688125610352, 0.010005503654479981, 0.009948160171508789, 0.009999360084533691, 0.01011404800415039, 0.009771007537841797, 0.009834495544433594, 0.01017958354949951, 0.01042739200592041, 0.010338303565979003, 0.010529791831970215, 0.010503168106079102, 0.010090496063232422, 0.010199040412902831, 0.010131456375122071, 0.010027008056640625, 0.010265600204467774, 0.01002086353302002, 0.010093567848205566, 0.021198848724365234, 0.009622528076171874, 0.009660415649414063, 0.009620479583740234, 0.009620479583740234, 0.009613311767578125, 0.009613311767578125, 0.009645055770874024, 0.009606143951416016, 0.00961023998260498, 0.009866239547729493, 0.00961945629119873, 0.009673760414123536, 0.009620448112487794, 0.00960102367401123, 0.00963276767730713, 0.009628671646118164, 0.009835519790649415, 0.009671680450439453, 0.00962559986114502, 0.00985804843902588, 0.010056703567504884, 0.010015744209289551, 0.01006387233734131, 0.010035200119018555, 0.010081279754638671, 0.010004480361938477, 0.01004032039642334, 0.01002393627166748, 0.010051584243774414, 0.01004032039642334, 0.01004851245880127, 0.01003929615020752, 0.010100735664367675, 0.010011648178100586, 0.010042367935180664, 0.01003929615020752, 0.010273792266845704, 0.009970687866210937, 0.009925632476806641, 0.009934880256652833, 0.00997987174987793, 0.009971712112426758, 0.009993215560913087, 0.009942015647888184, 0.009956352233886719, 0.009981951713562011, 0.010044416427612305, 0.010016768455505372, 0.00995638370513916, 0.009950176239013672, 0.009975808143615723, 0.009955327987670898, 0.010007552146911621, 0.009955327987670898, 0.010007552146911621, 0.009951231956481933, 0.009997311592102051, 0.009968640327453614, 0.009966591835021972, 0.009900032043457031, 0.009971712112426758, 0.009943039894104003, 0.021794815063476563, 0.009986111640930176, 0.009985983848571777, 0.009967616081237793, 0.009945088386535645, 0.009977855682373048, 0.009974783897399902, 0.009960448265075684, 0.00991641616821289, 0.009935872077941894, 0.00994918441772461, 0.009952256202697754, 0.009956352233886719, 0.009997311592102051, 0.009982975959777832, 0.010001407623291016, 0.009989119529724122, 0.009968640327453614, 0.009981951713562011, 0.009961471557617188, 0.009986047744750976, 0.009953280448913575, 0.01004851245880127, 0.010000384330749512, 0.009968640327453614, 0.00994611167907715, 0.00994611167907715, 0.009966591835021972, 0.00993280029296875, 0.009994239807128906, 0.009943039894104003, 0.009955327987670898, 0.009933823585510254, 0.009968640327453614, 0.009902079582214356, 0.009958399772644042, 0.00991641616821289, 0.009955327987670898, 0.009905152320861817, 0.009962495803833007, 0.00993280029296875, 0.009963520050048828, 0.009934847831726074, 0.009966591835021972, 0.009910271644592286, 0.009982975959777832, 0.009933823585510254, 0.009991168022155762, 0.009920512199401856, 0.009955327987670898, 0.00993791961669922, 0.009977855682373048, 0.009966591835021972, 0.009955327987670898, 0.00992460823059082, 0.00992460823059082, 0.009938943862915038, 0.009947135925292968, 0.009979904174804688, 0.01004032039642334, 0.009996288299560547, 0.009979904174804688, 0.009944064140319824, 0.021762048721313477, 0.009943039894104003, 0.009930751800537109, 0.009966591835021972, 0.009925632476806641, 0.009973759651184083, 0.00992972755432129, 0.009973759651184083, 0.009905152320861817, 0.009967616081237793, 0.009953280448913575, 0.0100065279006958, 0.009923583984375, 0.009957375526428223, 0.009918463706970216, 0.009969663619995118, 0.009953280448913575, 0.009966591835021972, 0.009972736358642579, 0.010060799598693848, 0.009956352233886719, 0.009938943862915038, 0.00994099235534668, 0.009995264053344726, 0.009954303741455077, 0.010035200119018555, 0.010070015907287597, 0.009989119529724122, 0.009988096237182617, 0.01004748821258545, 0.010070015907287597, 0.010009599685668944, 0.009972736358642579, 0.009976832389831543, 0.009994239807128906, 0.010000384330749512, 0.009958399772644042, 0.010016768455505372, 0.009984000205993653, 0.009991168022155762, 0.009964544296264649, 0.010030079841613769, 0.009953344345092773, 0.010006464004516602, 0.009988096237182617, 0.009975808143615723, 0.009955327987670898, 0.010019840240478516, 0.009948160171508789, 0.009986047744750976, 0.009964544296264649, 0.010028032302856446, 0.009986047744750976, 0.010087424278259278, 0.009951231956481933, 0.010001407623291016, 0.009862144470214844, 0.009874431610107422, 0.009845760345458985, 0.01004748821258545, 0.009976832389831543, 0.010011648178100586, 0.009912320137023926, 0.021744640350341796, 0.009973759651184083, 0.009935872077941894, 0.009976832389831543, 0.009991168022155762, 0.009945088386535645, 0.009977855682373048, 0.010057727813720703, 0.009981951713562011, 0.009960448265075684, 0.009999360084533691, 0.009959424018859863, 0.009972736358642579, 0.009971776008605958, 0.009999296188354493, 0.010001407623291016, 0.009994239807128906, 0.009967616081237793, 0.009930751800537109, 0.009975808143615723, 0.009938943862915038, 0.010068991661071777, 0.00994099235534668, 0.009996288299560547, 0.00992972755432129, 0.010002431869506835, 0.009930751800537109, 0.009971712112426758, 0.009927680015563965, 0.010036224365234376, 0.009968640327453614, 0.009977855682373048, 0.009943039894104003, 0.009999360084533691, 0.009926655769348144, 0.009956352233886719, 0.009933823585510254, 0.009960448265075684, 0.009944064140319824, 0.009943039894104003, 0.009921536445617676, 0.009954303741455077, 0.009923583984375, 0.009957375526428223, 0.00994918441772461, 0.01002188777923584, 0.009961471557617188, 0.009964544296264649, 0.009919487953186035, 0.009953280448913575, 0.009944064140319824, 0.00990822410583496, 0.009921536445617676, 0.00993791961669922, 0.00992972755432129, 0.010008576393127442, 0.009921536445617676, 0.009936896324157715, 0.010414079666137695, 0.009938943862915038, 0.009917440414428711, 0.009890815734863282, 0.00992460823059082, 0.02181427192687988, 0.009957375526428223, 0.010034175872802734, 0.009966591835021972, 0.009893888473510743, 0.009810943603515625, 0.009844736099243164, 0.009834495544433594, 0.009838591575622559, 0.009952256202697754, 0.009918463706970216, 0.010038271903991699, 0.009948160171508789, 0.009962495803833007, 0.009959424018859863, 0.009974783897399902, 0.009893888473510743, 0.00994819164276123, 0.009922528266906738, 0.00994611167907715, 0.009913344383239747, 0.009985024452209473, 0.009963520050048828, 0.009937952041625977, 0.009933792114257812, 0.009920512199401856, 0.009936896324157715, 0.00993388843536377, 0.009936832427978515, 0.009977855682373048, 0.009936896324157715, 0.00994918441772461, 0.00992972755432129, 0.009911295890808105, 0.009947135925292968, 0.009971712112426758, 0.009977855682373048, 0.009944064140319824, 0.009964544296264649, 0.009910271644592286, 0.009934847831726074, 0.009913375854492187, 0.00996348762512207, 0.00991436767578125, 0.010049535751342773, 0.009953280448913575, 0.009943039894104003, 0.009920512199401856, 0.009978879928588867, 0.010135552406311036, 0.010016768455505372, 0.010256383895874023, 0.010431488037109376, 0.009973759651184083, 0.009953280448913575, 0.009989119529724122, 0.009959487915039063, 0.009936832427978515, 0.009979904174804688, 0.009978879928588867, 0.010002431869506835, 0.00993177604675293, 0.01002086353302002, 0.021839872360229492, 0.010004480361938477, 0.00992460823059082, 0.009988096237182617, 0.009926655769348144, 0.009972736358642579, 0.009910271644592286, 0.009972736358642579, 0.00994611167907715, 0.009973759651184083, 0.00994099235534668, 0.010000384330749512, 0.009927680015563965, 0.009979904174804688, 0.009943039894104003, 0.009958399772644042, 0.009925632476806641, 0.010027008056640625, 0.00990822410583496, 0.009948160171508789, 0.009941056251525878, 0.00994707202911377, 0.009927680015563965, 0.009934847831726074, 0.009934847831726074, 0.009927680015563965, 0.009903103828430175, 0.009927680015563965, 0.009911295890808105, 0.009938943862915038, 0.00989900779724121, 0.009921536445617676, 0.009966591835021972, 0.009959424018859863, 0.010169343948364258, 0.009973759651184083, 0.009958399772644042, 0.00991648006439209, 0.010036160469055175, 0.009925632476806641, 0.009969663619995118, 0.00992255973815918, 0.009985024452209473, 0.009917440414428711, 0.010004480361938477, 0.009934847831726074, 0.009962495803833007, 0.009881600379943848, 0.009955360412597656, 0.009935839653015137, 0.00998300838470459, 0.009967583656311035, 0.009948160171508789, 0.009934847831726074, 0.009967616081237793, 0.009955327987670898, 0.009934847831726074, 0.00993996810913086, 0.010018815994262695, 0.009935872077941894, 0.009943039894104003, 0.009942015647888184, 0.00994918441772461, 0.021093408584594728, 0.009599967956542969, 0.00963276767730713, 0.009680895805358887, 0.009602047920227052, 0.009634816169738769, 0.009578495979309083, 0.009637887954711915, 0.009576448440551758, 0.00960921573638916, 0.009613311767578125, 0.009637887954711915, 0.009629695892333985, 0.009809920310974121, 0.009975808143615723, 0.009944064140319824, 0.009972736358642579, 0.009925632476806641, 0.009968640327453614, 0.009921536445617676, 0.009910271644592286, 0.009927680015563965, 0.00993791961669922, 0.00993996810913086, 0.01002188777923584, 0.009947135925292968, 0.00994611167907715, 0.00995126438140869, 0.009960415840148926, 0.009926655769348144, 0.009952256202697754, 0.009951231956481933, 0.009960448265075684, 0.00999833583831787, 0.009950207710266114, 0.009956352233886719, 0.0100065279006958, 0.009954303741455077, 0.009911295890808105, 0.009963583946228028, 0.00991430377960205, 0.009981951713562011, 0.009935935974121094, 0.009965503692626953, 0.009903103828430175, 0.00994918441772461, 0.009934847831726074, 0.01034348773956299, 0.010104767799377442, 0.00994918441772461, 0.009954303741455077, 0.00994918441772461, 0.009910271644592286, 0.00993996810913086, 0.010098688125610352, 0.00994918441772461, 0.009902144432067871, 0.009944031715393067, 0.009612256050109864, 0.009592831611633301, 0.009595904350280762, 0.009596960067749023, 0.009608160018920899, 0.021207040786743164, 0.009608192443847656, 0.009592896461486816, 0.009610176086425781, 0.009639936447143555, 0.009570367813110351, 0.009607104301452637, 0.009634816169738769, 0.00960102367401123, 0.009631744384765625, 0.009604096412658691, 0.009661439895629884, 0.009623552322387695, 0.009578495979309083, 0.00962764835357666, 0.009571328163146972, 0.009508864402770996, 0.00951193618774414, 0.009490431785583496, 0.00955084800720215, 0.009614336013793945, 0.009585663795471192, 0.009605119705200196, 0.009588735580444336, 0.009631744384765625, 0.009602047920227052, 0.009648127555847168, 0.009592831611633301, 0.009614336013793945, 0.00962559986114502, 0.009646080017089843, 0.009669631958007812, 0.009629695892333985, 0.009637887954711915, 0.009641983985900878, 0.009508864402770996, 0.009537535667419434, 0.00960921573638916, 0.009593855857849122, 0.009636863708496094, 0.00961638355255127, 0.009691136360168457, 0.00962662410736084, 0.009634816169738769, 0.00972697639465332, 0.009605119705200196, 0.00962764835357666, 0.00960921573638916, 0.009631744384765625, 0.009643008232116699, 0.00963276767730713, 0.009653247833251954, 0.009704447746276856, 0.009640959739685059, 0.009676799774169922, 0.00960921573638916, 0.009597951889038087, 0.009598976135253906, 0.009597951889038087, 0.009666560173034668, 0.00959388828277588, 0.009622495651245117, 0.009921536445617676, 0.02171392059326172, 0.009935903549194335, 0.009928671836853028, 0.009960448265075684, 0.009919487953186035, 0.00989900779724121, 0.0098088960647583, 0.009794560432434082, 0.009810943603515625, 0.009974783897399902, 0.009935872077941894, 0.009970687866210937, 0.009923583984375, 0.009965567588806153, 0.009942015647888184, 0.009978879928588867, 0.00994611167907715, 0.009977855682373048, 0.009801728248596191, 0.009629695892333985, 0.009584639549255371, 0.009630720138549804, 0.009613311767578125, 0.00962559986114502, 0.009631744384765625, 0.009656319618225098, 0.009675775527954102, 0.009645055770874024, 0.009605119705200196, 0.00962764835357666, 0.009595904350280762, 0.009597951889038087, 0.00961843204498291, 0.009677824020385742, 0.00962662410736084, 0.009657343864440919, 0.009777152061462402, 0.00960307216644287, 0.009702400207519531, 0.00961023998260498, 0.00961945629119873, 0.009658368110656738, 0.00961740779876709, 0.00971673583984375, 0.009621503829956055, 0.009620479583740234, 0.00963584041595459, 0.00961638355255127, 0.009683967590332031, 0.009719807624816895, 0.009666560173034668, 0.009623552322387695, 0.009597951889038087, 0.010090496063232422, 0.010071040153503418, 0.00994099235534668, 0.009945088386535645, 0.009975808143615723, 0.009953280448913575, 0.009962495803833007, 0.009938943862915038, 0.009957375526428223, 0.00993996810913086, 0.021659648895263672, 0.009827327728271485, 0.009830400466918946, 0.009877504348754883, 0.009993215560913087, 0.010036224365234376, 0.009959424018859863, 0.010038271903991699, 0.00999014377593994, 0.009988096237182617, 0.009919487953186035, 0.009963520050048828, 0.009923583984375, 0.010042367935180664, 0.01002291202545166, 0.0100065279006958, 0.009967616081237793, 0.010012672424316407, 0.010003456115722656, 0.009868288040161132, 0.009911295890808105, 0.009981951713562011, 0.00999833583831787, 0.010009599685668944, 0.009955327987670898, 0.01005568027496338, 0.010112000465393066, 0.010081279754638671, 0.009960448265075684, 0.009996288299560547, 0.009942015647888184, 0.009952256202697754, 0.009971712112426758, 0.009971712112426758, 0.00991539192199707, 0.009964544296264649, 0.009699328422546387, 0.009689087867736817, 0.00961023998260498, 0.009653280258178711, 0.009812959671020508, 0.009643008232116699, 0.009655296325683594, 0.00963584041595459, 0.009746432304382324, 0.009721856117248535, 0.009650176048278808, 0.009621503829956055, 0.009623616218566895, 0.0097074556350708, 0.009736191749572755, 0.009653247833251954, 0.009640959739685059, 0.009630720138549804, 0.009634816169738769, 0.009654272079467773, 0.009692192077636719, 0.009628640174865723, 0.00963276767730713, 0.009820159912109374, 0.00959488010406494, 0.009680928230285645, 0.010348511695861816, 0.022608896255493165, 0.010483712196350097, 0.009994239807128906, 0.010162176132202149, 0.009969696044921875, 0.009989088058471679, 0.009982975959777832, 0.009995264053344726, 0.009980928421020508, 0.00994918441772461, 0.00994918441772461, 0.009960448265075684, 0.009948160171508789, 0.01010483169555664, 0.009955327987670898, 0.009973823547363281, 0.009965503692626953, 0.009976832389831543, 0.00994918441772461, 0.009975808143615723, 0.00994918441772461, 0.010001407623291016, 0.009997344017028809, 0.010043359756469727, 0.009962495803833007, 0.010005503654479981, 0.010210304260253907, 0.009986047744750976, 0.009965567588806153, 0.009977855682373048, 0.00994099235534668, 0.010046463966369629, 0.009947135925292968, 0.009980928421020508, 0.009972736358642579, 0.009958399772644042, 0.00994611167907715, 0.009948160171508789, 0.010053631782531738, 0.010046463966369629, 0.009993215560913087, 0.010125311851501465, 0.010076160430908204, 0.009973759651184083, 0.009962495803833007, 0.009965567588806153, 0.009973823547363281, 0.00997164821624756, 0.009965567588806153, 0.009975808143615723, 0.009944064140319824, 0.009960448265075684, 0.010173439979553223, 0.010049535751342773, 0.009991168022155762, 0.009986047744750976, 0.009966591835021972, 0.009974783897399902, 0.009972736358642579, 0.010027008056640625, 0.009977855682373048, 0.010095616340637208, 0.010213376045227051, 0.021833728790283204, 0.010000384330749512, 0.010132479667663574, 0.010088447570800782, 0.010053631782531738, 0.010033151626586915, 0.010038271903991699, 0.009970687866210937, 0.010466303825378418, 0.010128383636474609, 0.01004032039642334, 0.010041343688964843, 0.009985024452209473, 0.009964544296264649, 0.009959424018859863, 0.009930751800537109, 0.009985024452209473, 0.01004748821258545, 0.010019840240478516, 0.009968640327453614, 0.009982975959777832, 0.01004748821258545, 0.009991168022155762, 0.009977855682373048, 0.009968640327453614, 0.009953344345092773, 0.009959360122680665, 0.009591808319091797, 0.00963276767730713, 0.009637887954711915, 0.009612288475036621, 0.0096245756149292, 0.00960102367401123, 0.009641983985900878, 0.009641983985900878, 0.009593855857849122, 0.009758720397949219, 0.010061823844909668, 0.009988096237182617, 0.009948224067687987, 0.009979840278625488, 0.009952256202697754, 0.009978879928588867, 0.009925632476806641, 0.009982975959777832, 0.009930751800537109, 0.009953280448913575, 0.009942015647888184, 0.01003110408782959, 0.009942015647888184, 0.009996288299560547, 0.00992972755432129, 0.010081279754638671, 0.00993791961669922, 0.009973759651184083, 0.009981951713562011, 0.009956352233886719, 0.009951231956481933, 0.009982975959777832, 0.01002393627166748, 0.009952256202697754, 0.009923583984375, 0.009944064140319824, 0.0233123836517334, 0.010062848091125488, 0.0100065279006958, 0.010028032302856446, 0.009979904174804688, 0.009987071990966797, 0.00994918441772461, 0.010005503654479981, 0.00993177604675293, 0.010889216423034668, 0.00999014377593994, 0.00994611167907715, 0.009950207710266114, 0.010024959564208985, 0.009953280448913575, 0.009945088386535645, 0.009965567588806153, 0.009947168350219726, 0.009943008422851562, 0.009927680015563965, 0.00992972755432129, 0.009903136253356934, 0.00995529556274414, 0.0100065279006958, 0.009953280448913575, 0.009967647552490234, 0.00995734405517578, 0.009976832389831543, 0.009926655769348144, 0.009981951713562011, 0.009928704261779785, 0.01003110408782959, 0.009935872077941894, 0.009976832389831543, 0.009933823585510254, 0.00999014377593994, 0.009923583984375, 0.009986047744750976, 0.009966591835021972, 0.010016768455505372, 0.009933823585510254, 0.010024959564208985, 0.009954303741455077, 0.009969663619995118, 0.009942015647888184, 0.009961471557617188, 0.009925632476806641, 0.009992192268371582, 0.009986047744750976, 0.009969663619995118, 0.00993280029296875, 0.010008576393127442, 0.009957375526428223, 0.009955327987670898, 0.009955327987670898, 0.009970687866210937, 0.009974783897399902, 0.009967616081237793, 0.00994918441772461, 0.009965567588806153, 0.009942015647888184, 0.009956352233886719, 0.009976832389831543, 0.021768224716186522, 0.009967583656311035, 0.009926655769348144, 0.009971712112426758, 0.009942015647888184, 0.00993177604675293, 0.009955327987670898, 0.009964544296264649, 0.00992255973815918, 0.009921536445617676, 0.009920512199401856, 0.009925663948059082, 0.009924575805664063, 0.009966591835021972, 0.009920512199401856, 0.009976832389831543, 0.009965567588806153, 0.009933823585510254, 0.009933823585510254, 0.010015744209289551, 0.009958399772644042, 0.009963520050048828, 0.009928704261779785, 0.009978879928588867, 0.00994099235534668, 0.009945119857788086, 0.009916383743286133, 0.009953311920166016, 0.009929696083068847, 0.009963520050048828, 0.009935872077941894, 0.010036224365234376, 0.00994918441772461, 0.009971712112426758, 0.00994611167907715, 0.009969663619995118, 0.00994099235534668, 0.009987071990966797, 0.009925632476806641, 0.010054656028747559, 0.009963520050048828, 0.01002905559539795, 0.009917440414428711, 0.009958399772644042, 0.00994099235534668, 0.009959424018859863, 0.00992972755432129, 0.009950207710266114, 0.009942015647888184, 0.009934847831726074, 0.00994921588897705, 0.009943008422851562, 0.009947135925292968, 0.00994918441772461, 0.009935872077941894, 0.009950207710266114, 0.009957375526428223, 0.009956352233886719, 0.009953280448913575, 0.009917440414428711, 0.009967616081237793, 0.009944064140319824, 0.00994611167907715, 0.021700639724731446, 0.009991200447082519, 0.009917375564575195, 0.009985024452209473, 0.00991436767578125, 0.009987071990966797, 0.009921536445617676, 0.009975808143615723, 0.009948160171508789, 0.010015744209289551, 0.00999014377593994, 0.009976832389831543, 0.009981951713562011, 0.00994099235534668, 0.009967616081237793, 0.009989119529724122, 0.010002495765686036, 0.010007488250732421, 0.009992192268371582, 0.00999014377593994, 0.009956352233886719, 0.010000384330749512, 0.009986047744750976, 0.009969663619995118, 0.009970687866210937, 0.00995126438140869, 0.009924575805664063, 0.009913344383239747, 0.009966591835021972, 0.009827327728271485, 0.009925632476806641, 0.009958399772644042, 0.009918463706970216, 0.00993177604675293, 0.009959424018859863, 0.009917440414428711, 0.009951231956481933, 0.010044447898864746, 0.009986016273498535, 0.009911359786987305, 0.01005356788635254, 0.009954303741455077, 0.009986047744750976, 0.00992460823059082, 0.009997311592102051, 0.00991436767578125, 0.009982975959777832, 0.009947135925292968, 0.009961471557617188, 0.00992972755432129, 0.009962495803833007, 0.009919487953186035, 0.00999833583831787, 0.00994611167907715, 0.010126367568969726, 0.010008543968200683, 0.009963520050048828, 0.009928704261779785, 0.010015744209289551, 0.00989900779724121, 0.009956352233886719, 0.00993177604675293, 0.009973759651184083]",tokens/s,98.96561316075605,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-6.7b,EleutherAI/pythia-6.7b,cuda,0,42,,,True,,,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",gpt_neox,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-13b,facebook/opt-13b,cuda,0,42,,,True,,,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",opt,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-6.7b,facebook/opt-6.7b,cuda,0,42,,,True,,,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",opt,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,microsoft/rho-math-1b-v0.1,microsoft/rho-math-1b-v0.1,cuda,0,42,,,True,True,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1130.88512,5172.10112,0.0,4531.945472,4478.990336,s,10,0.42166092681884765,0.042166092681884766,0.0005249509913911508,0.0420281925201416,0.04232391090393066,0.0430108190536499,0.043560345573425294,"[0.04369772720336914, 0.042159454345703125, 0.041919521331787106, 0.0421712646484375, 0.04211916732788086, 0.04204617691040039, 0.04201020812988281, 0.04183244705200195, 0.04185776138305664, 0.041847198486328126]",tokens/s,6071.228888371289,kWh,5.16059840697215e-07,2.8277386164204495e-07,3.61884568983405e-06,4.4176793921733096e-06,tokens/kWh,57948976.66262262,MB,1130.88512,5172.10112,0.0,4531.945472,4492.920832,s,10,13.891888793945311,1.3891888793945313,0.023342960546545024,1.3916763305664062,1.42256240234375,1.4231272094726561,1.4235790551757812,"[1.4224368896484374, 1.4236920166015625, 1.3934423828125, 1.3551798095703125, 1.381556396484375, 1.3899102783203126, 1.349650390625, 1.3772557373046874, 1.401121337890625, 1.3976435546875]",tokens/s,45.35020466580335,kWh,1.6424137673465907e-05,9.000330906268072e-06,4.601633290716629e-05,7.144080148690028e-05,tokens/kWh,881849.0090925419,,s,629,14.2456176738739,0.022648040816969636,0.005014991498451189,0.022172672271728516,0.02259086685180664,0.02295295944213867,0.06372139022827149,"[0.023433216094970705, 0.02329599952697754, 0.023392255783081056, 0.02315673637390137, 0.023367679595947266, 0.02327859115600586, 0.022940671920776368, 0.021988351821899413, 0.02204979133605957, 0.022413375854492188, 0.021933023452758788, 0.021866464614868165, 0.021921791076660157, 0.021787647247314454, 0.021924863815307616, 0.02206719970703125, 0.02194534492492676, 0.02182143974304199, 0.021769216537475586, 0.021787647247314454, 0.02168320083618164, 0.023205888748168944, 0.02326527976989746, 0.023174144744873046, 0.022991872787475585, 0.022542335510253905, 0.022572032928466795, 0.022590463638305663, 0.02248089599609375, 0.022627328872680662, 0.022587392807006838, 0.022598655700683593, 0.022595584869384764, 0.022563840866088865, 0.023376895904541017, 0.022808576583862306, 0.0226375675201416, 0.02262118339538574, 0.022571008682250978, 0.0225167350769043, 0.0228351993560791, 0.0225218563079834, 0.02248089599609375, 0.022601728439331056, 0.02248908805847168, 0.022535167694091796, 0.022483968734741212, 0.022526975631713866, 0.0224901123046875, 0.022560768127441407, 0.022564863204956053, 0.02252390480041504, 0.022427648544311524, 0.02250547218322754, 0.02255564880371094, 0.022435840606689454, 0.022697984695434572, 0.02259660720825195, 0.02248089599609375, 0.02245529556274414, 0.02246348762512207, 0.022494207382202147, 0.06446387481689453, 0.022527999877929687, 0.02255564880371094, 0.022553600311279298, 0.022487039566040038, 0.022534143447875975, 0.022207487106323243, 0.02268876838684082, 0.02250444793701172, 0.022905855178833007, 0.02249728012084961, 0.022769664764404295, 0.02271129608154297, 0.02248089599609375, 0.022451200485229493, 0.023824384689331055, 0.02342710494995117, 0.02288739204406738, 0.02251263999938965, 0.02246246337890625, 0.02249625587463379, 0.02255462455749512, 0.022475776672363282, 0.022518783569335937, 0.022562816619873048, 0.022486015319824217, 0.02252288055419922, 0.022609920501708985, 0.02249113655090332, 0.02251263999938965, 0.022071296691894532, 0.02248806381225586, 0.022527999877929687, 0.022623231887817383, 0.022558719635009765, 0.02242870330810547, 0.02251158332824707, 0.022393856048583984, 0.02251263999938965, 0.022587392807006838, 0.02242355155944824, 0.022402048110961914, 0.02260688018798828, 0.022592479705810548, 0.022432767868041992, 0.022462495803833006, 0.022575071334838867, 0.022599679946899414, 0.022401023864746093, 0.022397951126098634, 0.022389759063720704, 0.022466560363769532, 0.022452224731445314, 0.022417407989501953, 0.022502399444580077, 0.022445056915283205, 0.022498367309570312, 0.022459327697753908, 0.022780927658081054, 0.023590911865234376, 0.02341379165649414, 0.023615455627441405, 0.022580223083496095, 0.06491545867919922, 0.02242252731323242, 0.022483968734741212, 0.022397951126098634, 0.022509567260742186, 0.02246348762512207, 0.02244095993041992, 0.022428735733032227, 0.022558656692504883, 0.022467584609985353, 0.022509567260742186, 0.02244710350036621, 0.022419456481933595, 0.02247372817993164, 0.022503423690795898, 0.022433792114257813, 0.02249318313598633, 0.02247987174987793, 0.02245734405517578, 0.022510591506958007, 0.022436864852905275, 0.022433792114257813, 0.022566911697387695, 0.022502399444580077, 0.022500352859497072, 0.022477888107299806, 0.022497215270996095, 0.022535167694091796, 0.022508544921875, 0.022589439392089843, 0.022608896255493165, 0.022737920761108397, 0.022572032928466795, 0.02273587226867676, 0.022585344314575196, 0.023017471313476562, 0.022789119720458984, 0.02249728012084961, 0.02284236717224121, 0.022635520935058592, 0.022640640258789063, 0.02270412826538086, 0.021995519638061522, 0.02122854423522949, 0.02125312042236328, 0.02122035217285156, 0.021156864166259767, 0.021190656661987304, 0.021332992553710937, 0.021195775985717775, 0.021193727493286133, 0.021159936904907226, 0.02123776054382324, 0.02105446434020996, 0.02104729652404785, 0.0211824951171875, 0.021224416732788087, 0.02126438331604004, 0.021192703247070312, 0.021243904113769533, 0.021335039138793945, 0.02126643180847168, 0.021214208602905273, 0.06389555358886718, 0.02123676872253418, 0.021221343994140624, 0.021242879867553712, 0.02125004768371582, 0.02122137641906738, 0.02127359962463379, 0.021167104721069335, 0.0212541446685791, 0.021274688720703126, 0.021263296127319337, 0.02125721549987793, 0.02127257537841797, 0.021226495742797852, 0.021227519989013673, 0.021332992553710937, 0.02123161506652832, 0.0212490234375, 0.02119987106323242, 0.021219327926635743, 0.02125721549987793, 0.021238784790039062, 0.021269535064697264, 0.021258207321166993, 0.02125823974609375, 0.02125004768371582, 0.02128486442565918, 0.021296127319335938, 0.021197887420654298, 0.021256128311157228, 0.02126131248474121, 0.0212541446685791, 0.02128486442565918, 0.021259263992309572, 0.021404672622680664, 0.022982656478881838, 0.022576128005981445, 0.02231705665588379, 0.02217065620422363, 0.02207535934448242, 0.0220897274017334, 0.02207027244567871, 0.022012928009033202, 0.02207744026184082, 0.022168575286865236, 0.02206208038330078, 0.022099967956542968, 0.022071296691894532, 0.022128639221191407, 0.022072351455688477, 0.02212451171875, 0.0220579833984375, 0.02125312042236328, 0.021207040786743164, 0.021200895309448242, 0.021202943801879884, 0.02124083137512207, 0.021191680908203125, 0.021329919815063478, 0.02123676872253418, 0.021241823196411134, 0.02119987106323242, 0.021359615325927735, 0.06389759826660156, 0.021198848724365234, 0.021219327926635743, 0.021164031982421876, 0.02119987106323242, 0.021203968048095705, 0.021217279434204102, 0.021166080474853514, 0.021175296783447265, 0.020980735778808594, 0.02104012870788574, 0.021206016540527343, 0.021214208602905273, 0.021130239486694336, 0.021212160110473634, 0.021157888412475585, 0.022006784439086914, 0.021816320419311523, 0.021991424560546875, 0.021747711181640626, 0.021974016189575195, 0.022148096084594726, 0.021772287368774415, 0.021735424041748046, 0.021770240783691407, 0.02209382438659668, 0.02211123275756836, 0.022115327835083007, 0.022162431716918944, 0.022156288146972656, 0.022164480209350586, 0.022132736206054687, 0.022128639221191407, 0.02211737632751465, 0.022167552947998048, 0.022164480209350586, 0.022165504455566407, 0.022138879776000975, 0.022198272705078126, 0.022200351715087892, 0.022202335357666015, 0.022195199966430663, 0.02253209686279297, 0.022961151123046874, 0.02304614448547363, 0.0224716796875, 0.022180864334106445, 0.02217068862915039, 0.022125503540039063, 0.022129663467407225, 0.022178815841674804, 0.022114368438720704, 0.022077375411987305, 0.022152191162109376, 0.02226790428161621, 0.022255615234375, 0.022260736465454102, 0.022194175720214843, 0.022205440521240235, 0.02226585578918457, 0.022222848892211915, 0.02265088081359863, 0.022260736465454102, 0.06377062225341797, 0.021202943801879884, 0.021223424911499023, 0.02125721549987793, 0.021204992294311522, 0.021222400665283202, 0.021181440353393553, 0.021941247940063476, 0.02229043197631836, 0.022040576934814454, 0.02212761688232422, 0.02223411178588867, 0.022210559844970702, 0.02227609634399414, 0.022223871231079103, 0.022189056396484375, 0.02226585578918457, 0.022099967956542968, 0.02228121566772461, 0.022188032150268554, 0.022153215408325197, 0.02211123275756836, 0.022214656829833986, 0.022185983657836913, 0.022136831283569337, 0.022477823257446287, 0.022183935165405275, 0.02211123275756836, 0.022304800033569334, 0.022066144943237304, 0.022098943710327147, 0.02223411178588867, 0.022170623779296874, 0.02215020751953125, 0.022253503799438478, 0.022131711959838866, 0.022137855529785155, 0.022365184783935548, 0.02209689521789551, 0.022013952255249023, 0.02204979133605957, 0.022167552947998048, 0.021985279083251954, 0.02211123275756836, 0.022173696517944336, 0.022013952255249023, 0.022131711959838866, 0.022112255096435548, 0.022167552947998048, 0.02210508728027344, 0.02209791946411133, 0.022198272705078126, 0.02208665657043457, 0.02215116882324219, 0.02220134353637695, 0.022133760452270508, 0.022160383224487306, 0.022168575286865236, 0.022137855529785155, 0.022143999099731446, 0.022147071838378905, 0.022152191162109376, 0.022178815841674804, 0.06375321578979493, 0.02152448081970215, 0.02125721549987793, 0.02152448081970215, 0.02128281593322754, 0.021408767700195314, 0.021296127319335938, 0.02122547149658203, 0.021251071929931642, 0.021207040786743164, 0.021218303680419923, 0.0212807674407959, 0.0219238395690918, 0.021343231201171875, 0.021255168914794922, 0.021151744842529296, 0.021201919555664063, 0.021195775985717775, 0.02124595260620117, 0.021204992294311522, 0.0212992000579834, 0.02130227279663086, 0.02122035217285156, 0.02128691291809082, 0.021212160110473634, 0.02126438331604004, 0.02128895950317383, 0.021345279693603517, 0.021293088912963866, 0.02132169532775879, 0.021303295135498047, 0.02126950454711914, 0.02128281593322754, 0.02129715156555176, 0.021126144409179686, 0.021585920333862304, 0.022161407470703123, 0.022082559585571288, 0.022195199966430663, 0.02208460807800293, 0.022185983657836913, 0.022150144577026368, 0.022120447158813478, 0.02208051109313965, 0.02208153533935547, 0.02205286407470703, 0.021357568740844726, 0.021193727493286133, 0.02126131248474121, 0.021262367248535155, 0.02118550491333008, 0.021201919555664063, 0.021315584182739256, 0.02126643180847168, 0.021283840179443358, 0.021272640228271484, 0.021241792678833006, 0.021191680908203125, 0.021230592727661132, 0.021186559677124024, 0.021215232849121093, 0.021168127059936523, 0.021204992294311522, 0.0636119041442871, 0.021196800231933592, 0.021167104721069335, 0.021202943801879884, 0.0211712646484375, 0.02117318344116211, 0.02127667236328125, 0.021226495742797852, 0.0212807674407959, 0.021295103073120117, 0.021334016799926758, 0.02145280075073242, 0.021219327926635743, 0.021193727493286133, 0.021218303680419923, 0.021234687805175782, 0.02123263931274414, 0.02121625518798828, 0.02124492835998535, 0.02122854423522949, 0.021222400665283202, 0.021212160110473634, 0.021174272537231444, 0.02131046485900879, 0.021389312744140625, 0.022411264419555665, 0.022288383483886717, 0.022079488754272462, 0.02206003189086914, 0.022288383483886717, 0.02230169677734375, 0.022982656478881838, 0.022477823257446287, 0.022369279861450195, 0.0223242244720459, 0.022389759063720704, 0.02229248046875, 0.022296640396118166, 0.022214591979980467, 0.022177791595458983, 0.022210559844970702, 0.02208153533935547, 0.022189056396484375, 0.02230067253112793, 0.02185932731628418, 0.021811199188232423, 0.022211584091186523, 0.022174720764160157, 0.02229350471496582, 0.022171648025512695, 0.022213632583618165, 0.02233344078063965, 0.02223308753967285, 0.022164480209350586, 0.022404096603393556, 0.022326271057128907, 0.022221824645996095, 0.022175743103027345, 0.02239897537231445, 0.0223242244720459, 0.022138879776000975, 0.022239231109619142, 0.022213632583618165, 0.06426214599609376, 0.021881856918334962, 0.021737472534179687, 0.022112255096435548, 0.022108160018920898, 0.022162431716918944, 0.02210508728027344, 0.02215936088562012, 0.022253568649291993, 0.022240255355834963, 0.022163455963134765, 0.022204416275024414, 0.022139904022216796, 0.022106111526489256, 0.022389759063720704, 0.02223308753967285, 0.022116352081298828, 0.02225049591064453, 0.022107135772705077, 0.02345881652832031, 0.022524927139282228, 0.022189056396484375, 0.02169856071472168, 0.022386688232421875, 0.022495264053344725, 0.022146015167236327, 0.022190080642700196, 0.022185983657836913, 0.02213580894470215, 0.022255615234375, 0.02234060859680176, 0.022230016708374024, 0.02209382438659668, 0.02226380729675293, 0.022303743362426756, 0.022344703674316405, 0.022222848892211915, 0.022303743362426756, 0.02228428840637207, 0.022107135772705077, 0.022152191162109376, 0.02205900764465332, 0.022172672271728516, 0.02209791946411133, 0.02211737632751465, 0.022273056030273436, 0.02215011215209961, 0.02224332809448242, 0.0221214714050293, 0.023085056304931642, 0.02227097511291504, 0.022254592895507814, 0.022328319549560546, 0.022204416275024414, 0.02224127960205078, 0.022228992462158204, 0.022433792114257813, 0.02225868797302246, 0.02230169677734375, 0.022394880294799805, 0.02233344078063965, 0.022226943969726562, 0.022157312393188477, 0.06363955307006836, 0.021313535690307618, 0.02150399971008301, 0.02230169677734375, 0.022204416275024414, 0.022148096084594726, 0.02211020851135254, 0.022236160278320313, 0.022192127227783204, 0.022130687713623046, 0.022344703674316405, 0.022268928527832032, 0.022185983657836913, 0.02221059226989746, 0.022237152099609376, 0.022334463119506837, 0.021828607559204103, 0.022014976501464844, 0.022213632583618165, 0.022231039047241212, 0.022199296951293947, 0.022237184524536133, 0.02227712059020996, 0.022168575286865236, 0.022215679168701173, 0.022239231109619142, 0.022149120330810547, 0.022130687713623046, 0.022120447158813478, 0.022199296951293947, 0.022175743103027345, 0.022183967590332032, 0.022308832168579103, 0.022339584350585938, 0.02228223991394043, 0.022328319549560546, 0.022294559478759766, 0.022185951232910155, 0.022208511352539064, 0.022426624298095704, 0.02245427131652832, 0.02226380729675293, 0.022167552947998048, 0.022177791595458983, 0.022152191162109376, 0.022106143951416017, 0.02216649627685547, 0.022374399185180666, 0.022219776153564453, 0.022164480209350586, 0.022149120330810547, 0.022133760452270508, 0.022179840087890625, 0.02211020851135254, 0.022221824645996095, 0.022182912826538087, 0.02225766372680664, 0.022106111526489256, 0.022401023864746093, 0.022260736465454102, 0.022359039306640623, 0.022246400833129884, 0.022360063552856444]",tokens/s,44.15392960837142,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-2.7b,EleutherAI/pythia-2.7b,cuda,0,42,,,True,True,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1146.540032,12579.241984,0.0,11939.086336,11714.427904,s,10,1.006503044128418,0.1006503044128418,0.0014761109927726888,0.10015551757812499,0.10116352462768555,0.10308752250671387,0.10462672080993653,"[0.10501152038574219, 0.10034623718261719, 0.100162109375, 0.10001036834716796, 0.09981382751464844, 0.10006797027587891, 0.09984611511230469, 0.10014892578125, 0.10036000061035157, 0.10073596954345704]",tokens/s,2543.459768884091,kWh,1.206257881896566e-06,6.609661215741058e-07,8.43877375781639e-06,1.0305997761287061e-05,tokens/kWh,24839904.483739138,MB,1146.540032,12579.241984,0.0,11939.086336,11881.618432,s,10,18.944419067382814,1.8944419067382814,0.005874336927729204,1.8952128906249999,1.9003347290039063,1.9019706481933594,1.9032793835449218,"[1.894547119140625, 1.8846368408203125, 1.886148681640625, 1.895878662109375, 1.9036065673828124, 1.8965111083984374, 1.894208984375, 1.8890150146484375, 1.89997119140625, 1.8998948974609375]",tokens/s,33.25517651183563,kWh,2.219931550273279e-05,1.2164041081180825e-05,8.769219203578269e-05,0.00012205554861969631,tokens/kWh,516158.42714612634,,s,629,19.815456756591768,0.031503110900781874,0.011961892980804514,0.030057472229003908,0.030304045104980467,0.03044679718017578,0.13060185424804688,"[0.029978624343872072, 0.02998476791381836, 0.0300001277923584, 0.029764608383178712, 0.029841407775878907, 0.029799423217773437, 0.02977484893798828, 0.02978816032409668, 0.02979635238647461, 0.029853696823120116, 0.029908992767333983, 0.029791231155395507, 0.029829120635986327, 0.02979020881652832, 0.02987014389038086, 0.029915071487426757, 0.030509056091308592, 0.03040870475769043, 0.030235647201538086, 0.03013324737548828, 0.030127103805541993, 0.030176256179809572, 0.030445568084716795, 0.03039129638671875, 0.030096384048461915, 0.030216192245483397, 0.03016089630126953, 0.03013734436035156, 0.030187519073486328, 0.030100479125976562, 0.030242816925048828, 0.030076927185058593, 0.03015475273132324, 0.030273536682128906, 0.030233600616455077, 0.030192640304565428, 0.02995199966430664, 0.029953023910522462, 0.030202880859375, 0.029867008209228517, 0.030085119247436523, 0.02996735954284668, 0.029906944274902345, 0.029858816146850587, 0.029901824951171874, 0.03000531196594238, 0.03026425552368164, 0.030286848068237306, 0.03002060890197754, 0.02995097541809082, 0.029936639785766602, 0.029880319595336914, 0.029921279907226563, 0.02994486427307129, 0.029961183547973634, 0.030107648849487304, 0.03036057662963867, 0.030501888275146483, 0.03057868766784668, 0.030091264724731445, 0.03002572822570801, 0.030067712783813476, 0.13073715209960937, 0.029765695571899415, 0.029743040084838867, 0.02997452735900879, 0.030110719680786133, 0.030036991119384765, 0.02981177520751953, 0.029830080032348633, 0.02976153564453125, 0.029823999404907226, 0.02976870346069336, 0.029879295349121093, 0.02977791976928711, 0.0297574405670166, 0.029775871276855468, 0.029839359283447265, 0.02979327964782715, 0.029967391967773437, 0.029869024276733398, 0.02975129508972168, 0.02994790458679199, 0.029709312438964845, 0.02979327964782715, 0.029896703720092774, 0.029882368087768556, 0.029807615280151366, 0.029880319595336914, 0.02993356704711914, 0.029826047897338868, 0.029840383529663086, 0.029821952819824218, 0.029833215713500977, 0.029850688934326172, 0.029847488403320313, 0.029826047897338868, 0.029929471969604493, 0.029916160583496092, 0.029930496215820314, 0.029920255661010742, 0.029881376266479492, 0.029929439544677736, 0.029944831848144532, 0.030050304412841795, 0.030027776718139648, 0.030212095260620117, 0.0301844482421875, 0.029954048156738283, 0.02993152046203613, 0.02996940803527832, 0.030042112350463866, 0.030015487670898438, 0.029965311050415038, 0.030015487670898438, 0.02997452735900879, 0.02995916748046875, 0.029973503112792968, 0.02999603271484375, 0.029978624343872072, 0.029970432281494142, 0.02997452735900879, 0.0299233283996582, 0.02995916748046875, 0.03000115203857422, 0.130845703125, 0.029944831848144532, 0.029833215713500977, 0.02979327964782715, 0.029833215713500977, 0.029821952819824218, 0.02981068801879883, 0.02979532814025879, 0.02978508758544922, 0.029842432022094727, 0.029856767654418945, 0.029877248764038085, 0.02983526420593262, 0.029823999404907226, 0.02979020881652832, 0.029911039352416992, 0.02992742347717285, 0.029937664031982423, 0.03018342399597168, 0.03017420768737793, 0.02977996826171875, 0.029846527099609374, 0.029900800704956054, 0.02995814323425293, 0.029863935470581054, 0.029896703720092774, 0.029901824951171874, 0.029891584396362306, 0.029853696823120116, 0.029862911224365234, 0.029878271102905272, 0.029894655227661132, 0.029921279907226563, 0.029889535903930665, 0.030102527618408204, 0.02992742347717285, 0.02996633529663086, 0.029872127532958984, 0.029933631896972655, 0.02987718391418457, 0.029938688278198244, 0.029925376892089843, 0.02994892883300781, 0.029962240219116212, 0.029937664031982423, 0.029945856094360353, 0.02992639923095703, 0.02996735954284668, 0.02995916748046875, 0.030113792419433592, 0.030280704498291015, 0.03016499137878418, 0.029962240219116212, 0.029899776458740233, 0.030013439178466796, 0.029929471969604493, 0.03001651191711426, 0.02995916748046875, 0.03000217628479004, 0.02995916748046875, 0.02999091148376465, 0.029917184829711913, 0.02998784065246582, 0.13061016845703124, 0.03016499137878418, 0.02983526420593262, 0.02998067283630371, 0.030138368606567382, 0.03013324737548828, 0.030097408294677733, 0.03011686325073242, 0.030065664291381834, 0.02998579216003418, 0.029834239959716798, 0.029853696823120116, 0.029867040634155274, 0.029844480514526366, 0.02983216094970703, 0.029858816146850587, 0.02994486427307129, 0.030125024795532227, 0.03013734436035156, 0.030144512176513674, 0.030113792419433592, 0.029829120635986327, 0.02980659294128418, 0.029863935470581054, 0.029894655227661132, 0.029907968521118163, 0.029873151779174805, 0.029925376892089843, 0.029872127532958984, 0.030090240478515624, 0.030204927444458008, 0.02993561553955078, 0.02994380760192871, 0.03017932891845703, 0.029865983963012696, 0.029856767654418945, 0.030028799057006835, 0.03021824073791504, 0.029944831848144532, 0.029907968521118163, 0.029954048156738283, 0.029887487411499023, 0.029911039352416992, 0.0299683837890625, 0.030538751602172853, 0.03097599983215332, 0.030530559539794923, 0.030280704498291015, 0.030288896560668944, 0.030272512435913085, 0.030460927963256838, 0.030239744186401366, 0.030265344619750976, 0.030255104064941408, 0.030307327270507813, 0.030318592071533205, 0.030271488189697264, 0.03032268714904785, 0.03033087921142578, 0.03037696075439453, 0.030303232192993163, 0.030002208709716798, 0.03001955223083496, 0.13050469970703124, 0.030180383682250976, 0.03005955123901367, 0.03061958312988281, 0.030086143493652344, 0.030023679733276368, 0.030071807861328126, 0.03011686325073242, 0.030117887496948242, 0.030096384048461915, 0.03012505531311035, 0.030079999923706056, 0.030111743927001954, 0.03011177635192871, 0.030183391571044924, 0.030135295867919923, 0.03018854331970215, 0.030288896560668944, 0.030303232192993163, 0.030144512176513674, 0.030093311309814453, 0.030151679992675783, 0.030102527618408204, 0.03014041519165039, 0.030076927185058593, 0.030166015625, 0.030103551864624024, 0.030126079559326172, 0.03015884780883789, 0.03021516799926758, 0.03018649673461914, 0.030176256179809572, 0.03015884780883789, 0.030277664184570313, 0.030210079193115233, 0.03016294479370117, 0.030209983825683594, 0.029880319595336914, 0.03016908836364746, 0.030240768432617186, 0.030219263076782226, 0.030219263076782226, 0.03021516799926758, 0.030252031326293945, 0.030213119506835938, 0.03021824073791504, 0.030202880859375, 0.03021824073791504, 0.030235647201538086, 0.030255104064941408, 0.030858240127563476, 0.030486528396606444, 0.030264320373535155, 0.030298112869262695, 0.03033497619628906, 0.03022233581542969, 0.030257152557373046, 0.03041996765136719, 0.030442527770996094, 0.030309343338012697, 0.03023155212402344, 0.030263296127319338, 0.030234624862670898, 0.13058047485351562, 0.029891584396362306, 0.03017932891845703, 0.030146623611450197, 0.029785024642944337, 0.02981068801879883, 0.029844480514526366, 0.030144512176513674, 0.030112768173217775, 0.030102527618408204, 0.029812736511230467, 0.03016499137878418, 0.030136320114135744, 0.030095359802246095, 0.030069759368896484, 0.030090240478515624, 0.030104576110839845, 0.03021516799926758, 0.030311424255371092, 0.030297088623046874, 0.030101503372192383, 0.030103551864624024, 0.030187519073486328, 0.03012505531311035, 0.03054899215698242, 0.03037593650817871, 0.0301977596282959, 0.029902847290039062, 0.029914112091064454, 0.029858816146850587, 0.029922304153442384, 0.029885440826416015, 0.029865983963012696, 0.029884416580200194, 0.02996019172668457, 0.029877248764038085, 0.0299683837890625, 0.0299233283996582, 0.029904895782470704, 0.029901824951171874, 0.030048320770263672, 0.03023251152038574, 0.030196735382080078, 0.03019161605834961, 0.030136320114135744, 0.030200832366943358, 0.030258176803588867, 0.030248960494995116, 0.030303232192993163, 0.030196735382080078, 0.030254079818725587, 0.030005247116088866, 0.02996633529663086, 0.02999603271484375, 0.02999295997619629, 0.02998067283630371, 0.03002572822570801, 0.02999193572998047, 0.030057472229003908, 0.030993408203125, 0.030447616577148437, 0.030279680252075194, 0.030340095520019532, 0.13070851135253905, 0.030311391830444335, 0.03015475273132324, 0.030101503372192383, 0.03001759910583496, 0.03007174491882324, 0.03014143943786621, 0.030125120162963866, 0.030121919631958007, 0.030091264724731445, 0.030076927185058593, 0.0301977596282959, 0.030107648849487304, 0.029829120635986327, 0.02982809638977051, 0.029864959716796875, 0.030187519073486328, 0.030264352798461912, 0.030161888122558593, 0.029851648330688478, 0.02994892883300781, 0.029853696823120116, 0.029860864639282225, 0.029894655227661132, 0.029905920028686524, 0.029837312698364257, 0.029860864639282225, 0.029865983963012696, 0.02981888008117676, 0.029815807342529296, 0.029879295349121093, 0.029902847290039062, 0.029905920028686524, 0.030094335556030274, 0.030224384307861327, 0.030198848724365235, 0.030155712127685547, 0.03016908836364746, 0.030211103439331054, 0.03024073600769043, 0.03023155212402344, 0.029940799713134767, 0.029922239303588866, 0.02992848014831543, 0.029909984588623047, 0.029886463165283202, 0.02993152046203613, 0.02991417694091797, 0.030027711868286133, 0.030088191986083986, 0.029936639785766602, 0.030110719680786133, 0.030238719940185548, 0.030225408554077147, 0.030309375762939454, 0.029945856094360353, 0.03038412857055664, 0.030298112869262695, 0.030292991638183595, 0.030281728744506835, 0.030239744186401366, 0.030275583267211914, 0.029917184829711913, 0.13073408508300782, 0.029876224517822264, 0.029859840393066408, 0.029792255401611328, 0.029832191467285156, 0.02977894401550293, 0.02978713607788086, 0.02978201675415039, 0.029799423217773437, 0.029830144882202148, 0.029783039093017577, 0.029826047897338868, 0.029838336944580077, 0.029825023651123047, 0.029800447463989257, 0.029877248764038085, 0.029861888885498046, 0.029892608642578124, 0.029954048156738283, 0.03019878387451172, 0.03019878387451172, 0.03018547248840332, 0.03014860725402832, 0.0301711368560791, 0.029879295349121093, 0.029848575592041016, 0.029851680755615236, 0.029930463790893556, 0.029877248764038085, 0.029898752212524415, 0.029889535903930665, 0.029932544708251952, 0.029911039352416992, 0.0299182071685791, 0.029944831848144532, 0.029901824951171874, 0.02992742347717285, 0.029914112091064454, 0.03021107292175293, 0.02994175910949707, 0.02993971252441406, 0.029925376892089843, 0.029911039352416992, 0.02999091148376465, 0.029897727966308595, 0.029925376892089843, 0.02999603271484375, 0.029981695175170898, 0.02995916748046875, 0.02994790458679199, 0.030107648849487304, 0.03035852813720703, 0.030232576370239257, 0.02998886489868164, 0.02994380760192871, 0.030064640045166017, 0.030262271881103517, 0.030261247634887696, 0.030287872314453124, 0.030321664810180664, 0.03013324737548828, 0.030310400009155275, 0.03002572822570801, 0.1309102020263672, 0.02998067283630371, 0.029817855834960938, 0.03002060890197754, 0.03015372848510742, 0.030138368606567382, 0.030254079818725587, 0.03015372848510742, 0.030114816665649413, 0.03018649673461914, 0.030086143493652344, 0.02996019172668457, 0.029797376632690428, 0.029851648330688478, 0.030028799057006835, 0.030102527618408204, 0.02980454444885254, 0.030031871795654298, 0.029996063232421873, 0.03030729675292969, 0.030134271621704102, 0.030081024169921877, 0.029820928573608397, 0.029830144882202148, 0.030049280166625978, 0.03015065574645996, 0.029854719161987304, 0.02981888008117676, 0.029825088500976562, 0.030040000915527342, 0.03020185661315918, 0.03015782356262207, 0.030061567306518554, 0.029915136337280275, 0.030023679733276368, 0.03017011260986328, 0.029944831848144532, 0.03016499137878418, 0.03017523193359375, 0.03022336006164551, 0.03019878387451172, 0.03017318344116211, 0.03021830368041992, 0.030189504623413087, 0.03030940818786621, 0.030727136611938478, 0.03075584030151367, 0.030487552642822265, 0.030320640563964843, 0.030309375762939454, 0.030317567825317384, 0.030608383178710938, 0.03037286376953125, 0.030500864028930662, 0.030321664810180664, 0.030278656005859376, 0.03022336006164551, 0.030245887756347657, 0.030270463943481447, 0.030130176544189452, 0.030299135208129883, 0.030283775329589844, 0.030453760147094725, 0.13132389831542968, 0.029904895782470704, 0.029965311050415038, 0.030178367614746095, 0.03011577606201172, 0.02997760009765625, 0.030064640045166017, 0.030078975677490235, 0.03015782356262207, 0.030139392852783203, 0.030113792419433592, 0.030087167739868165, 0.03013222312927246, 0.030165056228637695, 0.030129087448120116, 0.030071807861328126, 0.03018342399597168, 0.030198816299438477, 0.030581727981567382, 0.030224447250366212, 0.03013420867919922, 0.03013324737548828, 0.030120960235595705, 0.03015475273132324, 0.030104576110839845, 0.03015577507019043, 0.030143487930297853, 0.030134271621704102, 0.030100479125976562, 0.030192640304565428, 0.03020595169067383, 0.03019059181213379, 0.0301977596282959, 0.030177312850952147, 0.030157791137695313, 0.03016908836364746, 0.03017011260986328, 0.030154783248901366, 0.03024995231628418, 0.030154783248901366, 0.030143455505371095, 0.02993971252441406, 0.02994380760192871, 0.029937664031982423, 0.0299683837890625, 0.02994175910949707, 0.03023155212402344, 0.02997760009765625, 0.02997248077392578, 0.030139392852783203, 0.030229503631591798, 0.030508031845092775, 0.030713855743408205, 0.030452735900878908, 0.030265344619750976, 0.030195711135864257, 0.030221343994140625, 0.03004105567932129, 0.03031449508666992, 0.030327808380126952, 0.030243839263916016, 0.030029823303222656, 0.02996428871154785]",tokens/s,31.742896857058682,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen2-beta-72B,Qwen/Qwen2-beta-72B,cuda,0,42,,,True,,,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",qwen2,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-0.5B,,cuda,0,42,,,,,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,217063f5c507ed7cc255df7e1f64c4333a0b4dfe,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,,MB,1186.746368,3657.957376,0.0,3017.801728,2552.274944,s,10,0.24441414260864255,0.024441414260864255,0.00043251915634180297,0.024288496017456054,0.024610393714904782,0.02516721258163452,0.02561266767501831,"[0.025724031448364257, 0.024263999938964845, 0.024272928237915038, 0.024242431640625, 0.024300703048706053, 0.024286367416381835, 0.024290624618530272, 0.024486656188964843, 0.02425142478942871, 0.02429497528076172]",tokens/s,10474.025654477318,kWh,2.9139205421915977e-07,1.596689524313692e-07,1.9127485197979913e-06,2.3638095264485206e-06,tokens/kWh,108299758.13856049,MB,1186.746368,3657.957376,0.0,3017.801728,2600.838656,s,10,12.865958374023437,1.2865958374023436,0.0049996890914484355,1.28646435546875,1.2950384643554687,1.2950643371582031,1.2950850354003907,"[1.29503271484375, 1.28695263671875, 1.287338623046875, 1.2841790771484376, 1.2802489013671876, 1.2830169677734375, 1.2950902099609376, 1.288172119140625, 1.28597607421875, 1.2799510498046875]",tokens/s,48.96642610565098,kWh,1.5121064223124806e-05,8.285549490908029e-06,3.177475065240171e-05,5.5181364366434566e-05,tokens/kWh,1141689.784646233,,s,629,13.062800367355349,0.020767568151598328,0.0029595608638864606,0.020345855712890625,0.020733133316040042,0.020934041213989257,0.04508356506347656,"[0.02146201515197754, 0.02138630485534668, 0.02181427192687988, 0.02086809539794922, 0.02039084815979004, 0.020444160461425782, 0.020447231292724608, 0.02024448013305664, 0.020135936737060548, 0.020377664566040038, 0.02037139129638672, 0.020360191345214843, 0.020312063217163084, 0.02037555122375488, 0.020328447341918944, 0.020372480392456056, 0.02031820869445801, 0.020298751831054687, 0.02049228858947754, 0.020358144760131838, 0.020404224395751954, 0.02043084716796875, 0.020333568572998048, 0.02027519989013672, 0.020361215591430663, 0.020302911758422852, 0.020850624084472656, 0.02085990333557129, 0.02087424087524414, 0.020852800369262695, 0.02038777542114258, 0.021507072448730468, 0.02063974380493164, 0.02039193534851074, 0.020746240615844725, 0.020550655364990233, 0.02067865562438965, 0.020601856231689454, 0.02025369644165039, 0.020410367965698242, 0.02030284881591797, 0.02039091110229492, 0.020348928451538087, 0.02033363151550293, 0.020396991729736327, 0.020539392471313478, 0.020354047775268554, 0.020377599716186523, 0.020330495834350586, 0.020373504638671876, 0.020376575469970702, 0.020369407653808593, 0.020311040878295897, 0.02045337677001953, 0.020485120773315428, 0.02044313621520996, 0.0202926082611084, 0.020586496353149415, 0.02166169548034668, 0.020554752349853517, 0.020324352264404297, 0.020380672454833985, 0.045093887329101565, 0.020382720947265624, 0.020350976943969725, 0.020313087463378905, 0.020553728103637696, 0.020380672454833985, 0.020374528884887694, 0.02038374328613281, 0.020484128952026368, 0.020444128036499025, 0.020419584274291993, 0.020364288330078126, 0.02043497657775879, 0.020532192230224608, 0.020404224395751954, 0.02042470359802246, 0.020802560806274413, 0.020925439834594727, 0.020455423355102538, 0.020436992645263673, 0.02043391990661621, 0.020434944152832032, 0.020496383666992187, 0.02020457649230957, 0.020371423721313477, 0.020311040878295897, 0.020381696701049806, 0.020330495834350586, 0.02043801689147949, 0.02050668716430664, 0.020406208038330077, 0.020351999282836913, 0.02039091110229492, 0.020381696701049806, 0.020495359420776366, 0.02035916709899902, 0.020406272888183592, 0.020351999282836913, 0.020428800582885744, 0.0204769287109375, 0.02042163276672363, 0.020388927459716797, 0.020453311920166015, 0.020447231292724608, 0.020377599716186523, 0.020346879959106445, 0.020509695053100584, 0.020314111709594726, 0.020378623962402344, 0.020386816024780274, 0.02040115165710449, 0.02032640075683594, 0.020414464950561522, 0.020298751831054687, 0.020405248641967775, 0.02031718444824219, 0.02041548728942871, 0.020366336822509764, 0.02072985649108887, 0.020460544586181642, 0.020388864517211915, 0.020379648208618165, 0.020410367965698242, 0.045126655578613284, 0.021143552780151367, 0.02088140869140625, 0.020273151397705077, 0.020362239837646484, 0.020468799591064454, 0.020827167510986327, 0.020427679061889647, 0.020368383407592772, 0.020393983840942383, 0.02070425605773926, 0.0202936954498291, 0.02038675117492676, 0.020380672454833985, 0.020356096267700196, 0.020339712142944336, 0.020349952697753908, 0.020322303771972656, 0.020591615676879883, 0.020603904724121092, 0.020406272888183592, 0.020393983840942383, 0.02040729522705078, 0.020332544326782227, 0.020336639404296874, 0.020513792037963868, 0.020370431900024414, 0.020336639404296874, 0.020345855712890625, 0.020331520080566406, 0.020332576751708985, 0.02031510353088379, 0.020341760635375978, 0.020363264083862305, 0.020327423095703127, 0.02032537651062012, 0.02041548728942871, 0.02036735916137695, 0.020369407653808593, 0.020381696701049806, 0.020357120513916017, 0.020336639404296874, 0.020303871154785155, 0.020434944152832032, 0.020348928451538087, 0.020381696701049806, 0.020338687896728515, 0.020341760635375978, 0.020370431900024414, 0.020338687896728515, 0.020634624481201173, 0.020773887634277344, 0.020343807220458983, 0.020314111709594726, 0.020353023529052734, 0.020617216110229493, 0.02082713508605957, 0.020620288848876952, 0.020366336822509764, 0.020324352264404297, 0.020376575469970702, 0.02028646469116211, 0.020299776077270508, 0.045052928924560545, 0.02030284881591797, 0.02025881576538086, 0.02021887969970703, 0.020271104812622072, 0.02028544044494629, 0.020331520080566406, 0.0203305606842041, 0.020256736755371093, 0.020278303146362305, 0.020352960586547852, 0.020306943893432617, 0.020405248641967775, 0.02048307228088379, 0.020299776077270508, 0.020291584014892578, 0.02030182456970215, 0.02045747184753418, 0.021621759414672852, 0.021209087371826172, 0.02064793586730957, 0.020254783630371094, 0.020268991470336915, 0.02031001663208008, 0.020314111709594726, 0.020387840270996094, 0.020468736648559572, 0.02028544044494629, 0.020303871154785155, 0.02028646469116211, 0.020274175643920898, 0.02026393508911133, 0.02026700782775879, 0.02025267219543457, 0.02063052749633789, 0.02137196731567383, 0.020917184829711916, 0.02023423957824707, 0.02026700782775879, 0.020287488937377928, 0.020290559768676757, 0.020265983581542968, 0.020380672454833985, 0.020248575210571287, 0.020274175643920898, 0.020262912750244142, 0.02020147132873535, 0.0200263671875, 0.020163583755493163, 0.020327423095703127, 0.020298751831054687, 0.02027008056640625, 0.02027827262878418, 0.020266048431396483, 0.0203703670501709, 0.0202608642578125, 0.020296703338623046, 0.02030284881591797, 0.020274175643920898, 0.02039193534851074, 0.02036735916137695, 0.020603904724121092, 0.02045132827758789, 0.04511129760742188, 0.020338687896728515, 0.02026700782775879, 0.020246528625488282, 0.020297727584838866, 0.020298751831054687, 0.02032537651062012, 0.02027724838256836, 0.020312063217163084, 0.020389984130859375, 0.020294559478759765, 0.020257791519165038, 0.020288511276245116, 0.02030080032348633, 0.02033459281921387, 0.020360191345214843, 0.02032537651062012, 0.020372512817382813, 0.02033967971801758, 0.020381696701049806, 0.020727807998657227, 0.020724735260009765, 0.020338687896728515, 0.020296703338623046, 0.020118528366088868, 0.020164608001708984, 0.020380672454833985, 0.020335615158081053, 0.020347904205322266, 0.020145151138305666, 0.020123647689819335, 0.02021990394592285, 0.020316160202026368, 0.02039193534851074, 0.020260927200317382, 0.020243392944335938, 0.02026803207397461, 0.020282367706298828, 0.020331520080566406, 0.020274175643920898, 0.020303871154785155, 0.02029465675354004, 0.02035916709899902, 0.02025369644165039, 0.02040012741088867, 0.020325408935546876, 0.020369375228881836, 0.020330495834350586, 0.020246528625488282, 0.020312063217163084, 0.020312063217163084, 0.0202608642578125, 0.020265983581542968, 0.02027212715148926, 0.020344831466674804, 0.02029465675354004, 0.020329471588134765, 0.020333568572998048, 0.020323360443115234, 0.020355039596557618, 0.020320255279541014, 0.0202608642578125, 0.020296703338623046, 0.04512768173217773, 0.02045337677001953, 0.020770816802978515, 0.020380735397338867, 0.020542400360107422, 0.020521984100341797, 0.021292032241821288, 0.020287488937377928, 0.020395008087158203, 0.02037555122375488, 0.020366336822509764, 0.020295679092407228, 0.020356096267700196, 0.02029465675354004, 0.02028544044494629, 0.02031007957458496, 0.020321216583251953, 0.020306943893432617, 0.020914176940917968, 0.020504575729370117, 0.020099071502685546, 0.020304895401000975, 0.020364288330078126, 0.020384767532348632, 0.020290559768676757, 0.020294719696044922, 0.020091840744018555, 0.020106239318847655, 0.020320255279541014, 0.020397056579589845, 0.02038374328613281, 0.0202926082611084, 0.020344831466674804, 0.020341760635375978, 0.020368383407592772, 0.020320255279541014, 0.020336639404296874, 0.020288511276245116, 0.02031718444824219, 0.02009600067138672, 0.020426752090454102, 0.020287488937377928, 0.020273151397705077, 0.02025164794921875, 0.020312063217163084, 0.020234304428100584, 0.020290496826171876, 0.020296703338623046, 0.020315135955810547, 0.020332544326782227, 0.020313087463378905, 0.02025062370300293, 0.02025574493408203, 0.0202608642578125, 0.020331520080566406, 0.021341184616088867, 0.020591615676879883, 0.020273151397705077, 0.020092927932739257, 0.02022400093078613, 0.020190208435058594, 0.02002227210998535, 0.020315135955810547, 0.04602982330322265, 0.0211015682220459, 0.02091110420227051, 0.020339712142944336, 0.020327423095703127, 0.02024345588684082, 0.020316160202026368, 0.02082815933227539, 0.020351999282836913, 0.020313087463378905, 0.02042470359802246, 0.020414464950561522, 0.020460544586181642, 0.020321279525756835, 0.020497472763061523, 0.020328384399414062, 0.020365312576293947, 0.020298751831054687, 0.020369407653808593, 0.020363264083862305, 0.02037555122375488, 0.020337663650512695, 0.020295679092407228, 0.02036735916137695, 0.020370431900024414, 0.020090879440307616, 0.02010316848754883, 0.02025164794921875, 0.020355072021484375, 0.020332544326782227, 0.02045952033996582, 0.02150092887878418, 0.021793792724609375, 0.021149696350097655, 0.02044825553894043, 0.020340736389160157, 0.020387840270996094, 0.020320255279541014, 0.020801536560058592, 0.020807680130004884, 0.021638208389282227, 0.020918207168579103, 0.02088755226135254, 0.020493312835693358, 0.020374528884887694, 0.020329471588134765, 0.02065407943725586, 0.020783103942871094, 0.020486143112182616, 0.020281343460083007, 0.020346879959106445, 0.02030284881591797, 0.020344831466674804, 0.020336639404296874, 0.02087116813659668, 0.02089369583129883, 0.02045747184753418, 0.020296703338623046, 0.02040012741088867, 0.020814847946166993, 0.02041753578186035, 0.02061311912536621, 0.020971519470214844, 0.04546867370605469, 0.020369407653808593, 0.02028544044494629, 0.02028441619873047, 0.020360191345214843, 0.02082713508605957, 0.020569087982177735, 0.02032537651062012, 0.020264959335327147, 0.020486143112182616, 0.020323328018188477, 0.0202608642578125, 0.02071244812011719, 0.020916223526000977, 0.020449312210083007, 0.02035708808898926, 0.020387840270996094, 0.02026803207397461, 0.020315135955810547, 0.020402175903320312, 0.02083839988708496, 0.020747264862060546, 0.020321279525756835, 0.021279743194580078, 0.02246246337890625, 0.02101759910583496, 0.020360191345214843, 0.020327423095703127, 0.02032537651062012, 0.020335615158081053, 0.020314111709594726, 0.020357120513916017, 0.020311040878295897, 0.020340736389160157, 0.02036735916137695, 0.020336639404296874, 0.02042572784423828, 0.020366336822509764, 0.020316160202026368, 0.02035305595397949, 0.020404191970825197, 0.020126720428466797, 0.020131839752197265, 0.020338687896728515, 0.020302911758422852, 0.02016864013671875, 0.02018716812133789, 0.020397024154663088, 0.02046771240234375, 0.02040729522705078, 0.020385791778564453, 0.020382720947265624, 0.020347904205322266, 0.020350976943969725, 0.020339712142944336, 0.020341760635375978, 0.020329471588134765, 0.020385791778564453, 0.020337663650512695, 0.02033459281921387, 0.020378623962402344, 0.020362239837646484, 0.020360191345214843, 0.04505702209472656, 0.020365312576293947, 0.02032137680053711, 0.020279199600219726, 0.02036735916137695, 0.020369407653808593, 0.020380672454833985, 0.02050048065185547, 0.020518911361694335, 0.0212674560546875, 0.020571136474609376, 0.020559871673583984, 0.020406272888183592, 0.02026803207397461, 0.02025881576538086, 0.02026393508911133, 0.02023526382446289, 0.02023321533203125, 0.02030182456970215, 0.020289535522460937, 0.020327423095703127, 0.020372480392456056, 0.020495359420776366, 0.020288511276245116, 0.020389888763427736, 0.020297727584838866, 0.020337663650512695, 0.020351999282836913, 0.020336639404296874, 0.020364288330078126, 0.020371456146240235, 0.020197439193725585, 0.020444095611572264, 0.02040934371948242, 0.020354047775268554, 0.020405248641967775, 0.02045337677001953, 0.02085683250427246, 0.020566015243530272, 0.020305919647216796, 0.02025574493408203, 0.02032537651062012, 0.020343807220458983, 0.020290559768676757, 0.020295679092407228, 0.02028339195251465, 0.020366336822509764, 0.020290559768676757, 0.020354047775268554, 0.020323328018188477, 0.020378623962402344, 0.020319232940673827, 0.0202926082611084, 0.02031718444824219, 0.020329471588134765, 0.020299776077270508, 0.02033667182922363, 0.02028028869628906, 0.021098495483398438, 0.020939775466918945, 0.020904960632324218, 0.020561920166015626, 0.020336639404296874, 0.04512153625488281, 0.020376575469970702, 0.020330495834350586, 0.02023628807067871, 0.02033459281921387, 0.020289535522460937, 0.020127775192260743, 0.020048864364624025, 0.020296703338623046, 0.02026905632019043, 0.02025881576538086, 0.02031718444824219, 0.02026188850402832, 0.020257791519165038, 0.020275295257568358, 0.02027408027648926, 0.020364288330078126, 0.02032640075683594, 0.0202926082611084, 0.020290559768676757, 0.02031718444824219, 0.020257791519165038, 0.020358144760131838, 0.02025881576538086, 0.020327423095703127, 0.0202936954498291, 0.020338623046875, 0.020320255279541014, 0.02035923194885254, 0.02013587188720703, 0.020135936737060548, 0.020214784622192384, 0.020341791152954102, 0.02029257583618164, 0.020303871154785155, 0.020257791519165038, 0.02035916709899902, 0.0202926082611084, 0.020344831466674804, 0.02027622413635254, 0.020295679092407228, 0.020324352264404297, 0.020221952438354493, 0.02063974380493164, 0.020699136734008788, 0.02037555122375488, 0.020320255279541014, 0.020536319732666015, 0.020312063217163084, 0.020404224395751954, 0.020328447341918944, 0.020385791778564453, 0.020296703338623046, 0.020271104812622072, 0.020249599456787108, 0.020307968139648438, 0.020291584014892578, 0.020289535522460937, 0.020281343460083007, 0.020324352264404297, 0.0202926082611084, 0.02038374328613281, 0.020331520080566406]",tokens/s,48.15200281035492,,,main,False,False,,,,,,,,,,,,,,,,,,,,,,, float32-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-1.8B,,cuda,0,42,,,,,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,217063f5c507ed7cc255df7e1f64c4333a0b4dfe,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,,MB,1176.375296,9272.03328,0.0,8631.877632,8498.684928,s,10,0.5381008338928224,0.053810083389282226,0.00035037334611295,0.05368016052246094,0.05422293357849121,0.05444493808746338,0.05462254169464112,"[0.05466694259643555, 0.05365244674682617, 0.05390115356445312, 0.05355408096313476, 0.05359164810180664, 0.05367539215087891, 0.053816032409667966, 0.05338460922241211, 0.05417359924316406, 0.05368492889404297]",tokens/s,4757.472649651933,kWh,6.675134018316108e-07,3.6573236569800707e-07,4.647660654723297e-06,5.680906422252914e-06,tokens/kWh,45063231.28246081,MB,1176.702976,9272.03328,0.0,8631.877632,8597.292544,s,10,13.016972412109373,1.3016972412109373,0.008260526575922323,1.2994389038085936,1.31171298828125,1.3148390380859374,1.3173398779296874,"[1.3035977783203125, 1.296099609375, 1.2949403076171875, 1.3008309326171874, 1.311018310546875, 1.317965087890625, 1.3095673828125, 1.298046875, 1.291166259765625, 1.2937398681640624]",tokens/s,48.39835101854609,kWh,1.532481990768636e-05,8.397759716100169e-06,5.2373523851075e-05,7.609610347486153e-05,tokens/kWh,827900.4722076493,,s,629,13.478948848724393,0.021429171460611074,0.006443288456210155,0.020567039489746093,0.021075148391723635,0.021187609863281248,0.0746680322265625,"[0.020566015243530272, 0.020549631118774413, 0.020551679611206054, 0.020537343978881836, 0.020470783233642577, 0.02045644760131836, 0.02043084716796875, 0.020504575729370117, 0.020477951049804686, 0.020528127670288086, 0.02048409652709961, 0.020271104812622072, 0.020376575469970702, 0.020439071655273436, 0.02044108772277832, 0.02051171112060547, 0.020512767791748047, 0.020567039489746093, 0.020534271240234374, 0.02043187141418457, 0.02045747184753418, 0.020431936264038084, 0.020485055923461913, 0.021149696350097655, 0.021112831115722656, 0.02450432014465332, 0.022615039825439453, 0.021727231979370116, 0.02063667106628418, 0.02065407943725586, 0.020559871673583984, 0.02058956718444824, 0.020658176422119142, 0.02061516761779785, 0.020625408172607423, 0.020585472106933594, 0.020552703857421875, 0.020546560287475587, 0.020584447860717774, 0.020582399368286132, 0.020556800842285155, 0.020527103424072265, 0.020550655364990233, 0.020534271240234374, 0.02058956718444824, 0.020599807739257812, 0.020612096786499022, 0.020561920166015626, 0.020567039489746093, 0.020618240356445314, 0.020618240356445314, 0.020588544845581053, 0.020548608779907225, 0.020584447860717774, 0.020509695053100584, 0.020578336715698243, 0.02053424072265625, 0.02061516761779785, 0.020540416717529295, 0.020595712661743162, 0.020567039489746093, 0.020543487548828124, 0.07469465637207032, 0.02064691162109375, 0.020553728103637696, 0.020607999801635742, 0.020572160720825194, 0.02058956718444824, 0.020641792297363282, 0.020555776596069338, 0.02062335968017578, 0.02040934371948242, 0.020392959594726562, 0.020546560287475587, 0.020552703857421875, 0.020520959854125977, 0.020706304550170897, 0.020552703857421875, 0.02057318305969238, 0.020582399368286132, 0.020609024047851563, 0.020525056838989256, 0.020570144653320313, 0.02054550361633301, 0.020600831985473633, 0.020516864776611327, 0.020562944412231447, 0.020488191604614257, 0.020559871673583984, 0.02057222366333008, 0.020487104415893555, 0.020580352783203124, 0.02108006477355957, 0.020548608779907225, 0.02049843215942383, 0.02057219123840332, 0.020572128295898436, 0.020519935607910156, 0.020564992904663085, 0.020548608779907225, 0.02052403259277344, 0.02050662422180176, 0.020568063735961914, 0.020541439056396483, 0.020559871673583984, 0.02062335968017578, 0.020572160720825194, 0.020575231552124023, 0.020634624481201173, 0.020550655364990233, 0.020582496643066408, 0.020598751068115234, 0.02036934471130371, 0.020374528884887694, 0.020591615676879883, 0.02053222465515137, 0.020577280044555665, 0.020587520599365236, 0.020571168899536134, 0.020640735626220704, 0.020626432418823244, 0.020609024047851563, 0.020547615051269532, 0.020557792663574218, 0.020529151916503906, 0.07494348907470703, 0.020534271240234374, 0.020545536041259766, 0.020559871673583984, 0.02051481628417969, 0.020557823181152343, 0.020520959854125977, 0.020517887115478514, 0.020529151916503906, 0.020557823181152343, 0.020545536041259766, 0.020582399368286132, 0.02062848091125488, 0.020505599975585938, 0.020634624481201173, 0.020554752349853517, 0.020562944412231447, 0.020658176422119142, 0.020600831985473633, 0.020579328536987306, 0.020617216110229493, 0.0206561279296875, 0.020562944412231447, 0.020582399368286132, 0.020591615676879883, 0.02060697555541992, 0.020595712661743162, 0.020576255798339844, 0.020572160720825194, 0.020534271240234374, 0.020519935607910156, 0.020592639923095703, 0.020397056579589845, 0.020350976943969725, 0.020366336822509764, 0.0204400634765625, 0.020536319732666015, 0.02059775924682617, 0.020561920166015626, 0.02057318305969238, 0.020591615676879883, 0.020641792297363282, 0.020561920166015626, 0.020542463302612304, 0.02052511978149414, 0.02057209587097168, 0.020528127670288086, 0.020555776596069338, 0.020601856231689454, 0.020525056838989256, 0.02054969596862793, 0.020471744537353516, 0.020505599975585938, 0.02050048065185547, 0.0204902400970459, 0.020550655364990233, 0.020518911361694335, 0.020539392471313478, 0.020551679611206054, 0.020489248275756836, 0.020499423980712892, 0.02048204803466797, 0.020580352783203124, 0.07464959716796875, 0.020530176162719727, 0.020580352783203124, 0.020523008346557618, 0.020505599975585938, 0.020578304290771485, 0.02047488021850586, 0.020584447860717774, 0.020505599975585938, 0.020541439056396483, 0.020504575729370117, 0.020620288848876952, 0.020545536041259766, 0.020521984100341797, 0.02052403259277344, 0.020555776596069338, 0.020548608779907225, 0.020973567962646485, 0.02058137512207031, 0.020742143630981445, 0.02045952033996582, 0.020552736282348633, 0.020530208587646485, 0.02051068878173828, 0.02051375961303711, 0.020632575988769532, 0.020578304290771485, 0.020504575729370117, 0.020550655364990233, 0.020550655364990233, 0.020540416717529295, 0.02058038330078125, 0.020603872299194335, 0.020570175170898437, 0.02053215980529785, 0.020548608779907225, 0.020898815155029296, 0.020549631118774413, 0.020494335174560546, 0.020764671325683593, 0.020560895919799805, 0.02067865562438965, 0.020518911361694335, 0.020534271240234374, 0.021013504028320314, 0.020619264602661135, 0.02058963203430176, 0.02100524711608887, 0.020634624481201173, 0.020925439834594727, 0.02063155174255371, 0.020593664169311524, 0.020974592208862306, 0.02051481628417969, 0.020537343978881836, 0.02088960075378418, 0.020540416717529295, 0.02067865562438965, 0.021784576416015625, 0.021168127059936523, 0.02091929626464844, 0.020542463302612304, 0.020528127670288086, 0.07467520141601562, 0.020539392471313478, 0.02046771240234375, 0.020999168395996092, 0.02051481628417969, 0.020676607131958007, 0.020575231552124023, 0.02050048065185547, 0.020880384445190428, 0.020543487548828124, 0.020577280044555665, 0.021061632156372072, 0.02063871955871582, 0.020769792556762694, 0.020593664169311524, 0.020521984100341797, 0.020977664947509765, 0.021168127059936523, 0.02085785675048828, 0.02107596778869629, 0.020574207305908202, 0.02063155174255371, 0.021082111358642578, 0.021113855361938477, 0.020913152694702147, 0.020544511795043945, 0.020685823440551757, 0.021234687805175782, 0.021046272277832033, 0.020544511795043945, 0.02065715217590332, 0.02063667106628418, 0.020624383926391602, 0.020548608779907225, 0.020567039489746093, 0.020832256317138673, 0.021118976593017577, 0.021170175552368165, 0.02109644889831543, 0.020596736907958983, 0.020917247772216797, 0.020561920166015626, 0.02053222465515137, 0.02099404716491699, 0.020546560287475587, 0.02049849510192871, 0.021067712783813478, 0.020572160720825194, 0.020786176681518553, 0.02050662422180176, 0.02066739273071289, 0.021170175552368165, 0.020701183319091796, 0.020668415069580077, 0.021178367614746094, 0.021135360717773437, 0.021204992294311522, 0.020900863647460938, 0.021169151306152344, 0.02084966468811035, 0.020819007873535158, 0.02115987205505371, 0.021189632415771483, 0.07523020935058594, 0.021137407302856445, 0.02123673629760742, 0.021105663299560547, 0.02109951972961426, 0.02107494354248047, 0.021194751739501954, 0.02106470489501953, 0.021117952346801756, 0.02072985649108887, 0.021176319122314453, 0.021184576034545897, 0.021138368606567384, 0.02123776054382324, 0.021234687805175782, 0.02108006477355957, 0.020992000579833983, 0.021158912658691405, 0.02087120056152344, 0.020984800338745117, 0.02116204833984375, 0.020683712005615234, 0.02109951972961426, 0.021008384704589843, 0.02205081558227539, 0.02228428840637207, 0.020999168395996092, 0.02060492706298828, 0.020634624481201173, 0.020538368225097657, 0.020548608779907225, 0.02057318305969238, 0.020527103424072265, 0.02051481628417969, 0.020764671325683593, 0.020509695053100584, 0.02066329574584961, 0.02053222465515137, 0.020508672714233397, 0.020891647338867187, 0.020736000061035157, 0.020557823181152343, 0.021029888153076173, 0.020518911361694335, 0.020738048553466795, 0.020540416717529295, 0.020601856231689454, 0.02089574432373047, 0.020503551483154296, 0.020599807739257812, 0.020755456924438476, 0.020509695053100584, 0.020668415069580077, 0.020478975296020507, 0.020530176162719727, 0.02085171127319336, 0.022261760711669923, 0.021413888931274414, 0.02106982421875, 0.021121023178100586, 0.020837440490722656, 0.02063046455383301, 0.02067251205444336, 0.07491276550292969, 0.020529151916503906, 0.020552703857421875, 0.02084966468811035, 0.020824064254760744, 0.02062131118774414, 0.02108415985107422, 0.021398527145385742, 0.02079641532897949, 0.02062745666503906, 0.02059878349304199, 0.021196800231933592, 0.020633600234985353, 0.020704288482666016, 0.021265375137329103, 0.02125209617614746, 0.02085478401184082, 0.020590591430664062, 0.02062131118774414, 0.02105036735534668, 0.020569087982177735, 0.020583423614501953, 0.020937728881835937, 0.020891647338867187, 0.020775936126708985, 0.02057318305969238, 0.020725759506225586, 0.020574272155761717, 0.020619199752807616, 0.020628543853759767, 0.02058028793334961, 0.020948991775512696, 0.020566015243530272, 0.020997119903564454, 0.020587520599365236, 0.02064896011352539, 0.0206080322265625, 0.020601823806762697, 0.020732927322387695, 0.02103910446166992, 0.020678688049316406, 0.02053424072265625, 0.020520959854125977, 0.020592639923095703, 0.02064793586730957, 0.022018047332763673, 0.02227302360534668, 0.021588991165161133, 0.02166374397277832, 0.020633600234985353, 0.02060492706298828, 0.020571136474609376, 0.020560895919799805, 0.020553728103637696, 0.020620288848876952, 0.020539392471313478, 0.02048204803466797, 0.020478975296020507, 0.020536319732666015, 0.020562944412231447, 0.020546560287475587, 0.020534271240234374, 0.02050764846801758, 0.07447142028808594, 0.020553728103637696, 0.020554752349853517, 0.020596736907958983, 0.020541439056396483, 0.02058137512207031, 0.020570112228393556, 0.020577280044555665, 0.020556800842285155, 0.020595712661743162, 0.020559871673583984, 0.020493343353271486, 0.020548576354980468, 0.02068377685546875, 0.020345855712890625, 0.020436992645263673, 0.02061516761779785, 0.020519935607910156, 0.020559871673583984, 0.0204769287109375, 0.020556800842285155, 0.020539392471313478, 0.0208721923828125, 0.02109132766723633, 0.020596736907958983, 0.020518911361694335, 0.020555776596069338, 0.020591615676879883, 0.020634624481201173, 0.021158912658691405, 0.020550655364990233, 0.02083328056335449, 0.020619264602661135, 0.020574207305908202, 0.02051584053039551, 0.02067865562438965, 0.02049228858947754, 0.020544544219970703, 0.020487136840820312, 0.020590591430664062, 0.0206059513092041, 0.02053222465515137, 0.020585472106933594, 0.020562944412231447, 0.020495359420776366, 0.020488191604614257, 0.020595712661743162, 0.020534271240234374, 0.020638751983642577, 0.02068681526184082, 0.020571136474609376, 0.020542463302612304, 0.020577280044555665, 0.020545536041259766, 0.020658176422119142, 0.020560895919799805, 0.02059878349304199, 0.02082918357849121, 0.020584447860717774, 0.02065203285217285, 0.020741119384765624, 0.020510719299316405, 0.020554752349853517, 0.07528959655761719, 0.020634624481201173, 0.02071347236633301, 0.020583423614501953, 0.02048409652709961, 0.020546560287475587, 0.020542463302612304, 0.020498464584350586, 0.020568031311035156, 0.020496383666992187, 0.020564992904663085, 0.020497440338134765, 0.020544479370117188, 0.020550655364990233, 0.020541439056396483, 0.020519935607910156, 0.020454399108886717, 0.020319232940673827, 0.020497407913208008, 0.020312063217163084, 0.020316160202026368, 0.02050048065185547, 0.020551679611206054, 0.020568063735961914, 0.020388864517211915, 0.020374528884887694, 0.02064691162109375, 0.020544511795043945, 0.020530176162719727, 0.02048204803466797, 0.02053222465515137, 0.020519935607910156, 0.020365312576293947, 0.020348928451538087, 0.020312063217163084, 0.020257791519165038, 0.020295679092407228, 0.02046976089477539, 0.020517887115478514, 0.020527103424072265, 0.020571136474609376, 0.020427776336669923, 0.02031820869445801, 0.020382720947265624, 0.02045849609375, 0.020477951049804686, 0.020531200408935548, 0.020575231552124023, 0.02048307228088379, 0.020536319732666015, 0.020486143112182616, 0.020479999542236327, 0.020476959228515626, 0.02046771240234375, 0.020469728469848632, 0.02046156883239746, 0.020518943786621092, 0.0205383358001709, 0.020504575729370117, 0.02061414337158203, 0.020493312835693358, 0.02042367935180664, 0.020468736648559572, 0.07475609588623047, 0.020543487548828124, 0.020519935607910156, 0.020510719299316405, 0.020563968658447264, 0.02059878349304199, 0.02046463966369629, 0.02051584053039551, 0.02050764846801758, 0.02055990409851074, 0.02055369567871094, 0.020576255798339844, 0.020520959854125977, 0.020556800842285155, 0.020510719299316405, 0.020493312835693358, 0.020632575988769532, 0.02030182456970215, 0.02045849609375, 0.020544511795043945, 0.02050048065185547, 0.020558847427368163, 0.020546560287475587, 0.020553728103637696, 0.020557823181152343, 0.02041753578186035, 0.02035916709899902, 0.020444160461425782, 0.020580352783203124, 0.020535295486450195, 0.020553728103637696, 0.020706304550170897, 0.020562944412231447, 0.020577280044555665, 0.020402175903320312, 0.02033459281921387, 0.020555776596069338, 0.020366336822509764, 0.02033459281921387, 0.020519935607910156, 0.020619264602661135, 0.020478975296020507, 0.020486143112182616, 0.020596736907958983, 0.020510719299316405, 0.020576255798339844, 0.020539392471313478, 0.020502527236938475, 0.02052716827392578, 0.02056595230102539, 0.020568063735961914, 0.020594688415527345, 0.020583423614501953, 0.020546560287475587, 0.020617216110229493, 0.020567039489746093, 0.02046976089477539, 0.020561983108520508, 0.020559808731079102, 0.020592639923095703, 0.020546560287475587, 0.0206059513092041, 0.020554752349853517]",tokens/s,46.66535996681432,,,main,False,False,,,,,,,,,,,,,,,,,,,,,,, float32-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-70m,,cuda,0,42,,,,,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,217063f5c507ed7cc255df7e1f64c4333a0b4dfe,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,,MB,1149.927424,1057.488896,0.0,417.333248,390.916096,s,29,0.18655382394790646,0.006432890480962293,0.0006243825863804746,0.006244639873504639,0.006554713630676269,0.006909856128692626,0.008880492954254148,"[0.009556032180786133, 0.006559552192687989, 0.006544223785400391, 0.006490272045135498, 0.006211552143096924, 0.006197440147399902, 0.0061496000289916995, 0.0061785922050476075, 0.00655350399017334, 0.006248767852783203, 0.0065418238639831545, 0.006255072116851806, 0.00621504020690918, 0.006191008090972901, 0.006184000015258789, 0.006208799839019776, 0.0061847357749938965, 0.006197120189666748, 0.006208479881286621, 0.006428575992584228, 0.006230559825897217, 0.006180031776428223, 0.006181024074554444, 0.006244639873504639, 0.007143392086029053, 0.006257279872894287, 0.006292543888092041, 0.006468128204345703, 0.006252031803131103]",tokens/s,39795.4855220394,kWh,7.582184651526757e-08,4.154681081706465e-08,2.448575628806205e-07,3.6222622021295275e-07,tokens/kWh,706740665.6798551,MB,1149.927424,1059.586048,0.0,419.4304,397.46048,s,29,9.972502014160156,0.3438793797986261,0.00926896114094249,0.3454414367675781,0.35294813842773437,0.35499730224609377,0.3673164794921875,"[0.37167498779296876, 0.35610888671875, 0.3528526916503906, 0.3533299255371094, 0.3343729248046875, 0.33478692626953127, 0.3343879089355469, 0.3350140380859375, 0.35239630126953125, 0.3492184753417969, 0.3517505187988281, 0.33436032104492186, 0.3345603332519531, 0.33441622924804687, 0.3412231750488281, 0.3344357299804688, 0.33475634765625, 0.3344097900390625, 0.3482939147949219, 0.3450007934570313, 0.3463901977539062, 0.3346954040527344, 0.3341729431152344, 0.34753506469726564, 0.3525305480957031, 0.3499491271972656, 0.3484896545410156, 0.3454414367675781, 0.3459474182128906]",tokens/s,183.20377347688733,kWh,4.016638727346158e-06,2.2009346095502627e-06,6.68014797711941e-06,1.2897721314015831e-05,tokens/kWh,4884583.754460448,,s,1826,10.14188135290146,0.005554151890964655,0.0008424822024976228,0.005349376201629639,0.005629951953887939,0.005769472002983093,0.011932415962219238,"[0.006718463897705078, 0.006207488059997559, 0.006231135845184326, 0.005981088161468506, 0.0059770879745483394, 0.005895167827606201, 0.005740543842315674, 0.00573747205734253, 0.005708799839019775, 0.005724160194396972, 0.005744639873504639, 0.0056555519104003905, 0.005545983791351319, 0.005553152084350586, 0.005636096000671387, 0.005652480125427246, 0.005795839786529541, 0.005743616104125977, 0.0057712640762329105, 0.005761023998260498, 0.005738495826721191, 0.005763072013854981, 0.005884928226470947, 0.006144000053405762, 0.006030335903167725, 0.006045695781707764, 0.006055967807769775, 0.006073311805725097, 0.005933055877685547, 0.006012928009033203, 0.0059699201583862304, 0.006023168087005615, 0.006005760192871094, 0.005864448070526123, 0.005854207992553711, 0.005921792030334472, 0.005932032108306885, 0.0059023361206054685, 0.005922848224639892, 0.006025184154510498, 0.005909503936767578, 0.006093823909759521, 0.00582144021987915, 0.00582144021987915, 0.005840896129608154, 0.005746687889099121, 0.005723135948181152, 0.005628928184509278, 0.005777408123016357, 0.005811200141906738, 0.005654528141021729, 0.005939263820648194, 0.006015935897827149, 0.005873663902282715, 0.005950463771820068, 0.005631999969482422, 0.005650432109832764, 0.0057149438858032225, 0.005836800098419189, 0.005587967872619629, 0.005624832153320313, 0.005666816234588623, 0.01276211166381836, 0.005615615844726562, 0.0056258559226989744, 0.005624832153320313, 0.005635072231292724, 0.005829631805419922, 0.005846015930175781, 0.005801983833312988, 0.005794816017150879, 0.00572108793258667, 0.005693439960479736, 0.005690368175506591, 0.005577727794647217, 0.00572108793258667, 0.005619743824005127, 0.005535711765289306, 0.005612544059753418, 0.005407743930816651, 0.005426176071166992, 0.005636096000671387, 0.005543935775756836, 0.005369855880737305, 0.005305344104766845, 0.005325823783874512, 0.005351424217224121, 0.005347328186035156, 0.005795839786529541, 0.005920767784118652, 0.005858304023742676, 0.005955584049224853, 0.00601087999343872, 0.006094848155975342, 0.006065152168273926, 0.005858304023742676, 0.005668863773345947, 0.005595168113708496, 0.005618656158447265, 0.005586944103240967, 0.0056442880630493165, 0.005634079933166504, 0.005648352146148681, 0.0055920639038085935, 0.005615615844726562, 0.0056351680755615235, 0.005569439888000488, 0.005650496006011963, 0.005598144054412842, 0.005618688106536865, 0.005634047985076904, 0.005581823825836182, 0.005619711875915527, 0.005588992118835449, 0.005593088150024414, 0.005607423782348633, 0.005574656009674072, 0.005650464057922363, 0.005593056201934814, 0.005605375766754151, 0.005678112030029297, 0.0055961918830871584, 0.0056063361167907715, 0.005591040134429932, 0.005559296131134033, 0.01267404842376709, 0.005583871841430664, 0.005601280212402344, 0.005595136165618897, 0.005587967872619629, 0.005571584224700928, 0.005588992118835449, 0.005573631763458252, 0.005605375766754151, 0.005633024215698243, 0.005618688106536865, 0.005568511962890625, 0.005568511962890625, 0.005599232196807862, 0.0055552000999450684, 0.005554207801818848, 0.005577695846557617, 0.005604351997375488, 0.00563097620010376, 0.005560319900512695, 0.005581823825836182, 0.005600255966186523, 0.0055797758102417, 0.00562278413772583, 0.005583871841430664, 0.005569536209106446, 0.00562278413772583, 0.005577727794647217, 0.0056217598915100095, 0.00562278413772583, 0.005617663860321045, 0.005608448028564453, 0.005618688106536865, 0.005610559940338135, 0.0056350078582763675, 0.005575679779052735, 0.005683199882507324, 0.005574656009674072, 0.005602303981781006, 0.005599232196807862, 0.0055808000564575196, 0.0055848960876464845, 0.005603328227996827, 0.005578752040863037, 0.005559296131134033, 0.005579808235168457, 0.0055797438621521, 0.005606400012969971, 0.005605375766754151, 0.005608448028564453, 0.005557248115539551, 0.005576704025268555, 0.005583871841430664, 0.005577792167663574, 0.005609407901763916, 0.005602303981781006, 0.0055797758102417, 0.005560319900512695, 0.005596159934997558, 0.005610527992248535, 0.0056063680648803715, 0.005576704025268555, 0.005575744152069092, 0.012468159675598145, 0.005629951953887939, 0.005560319900512695, 0.005594111919403076, 0.005590015888214111, 0.005567488193511963, 0.005619711875915527, 0.0055920639038085935, 0.005643263816833496, 0.005590015888214111, 0.005587967872619629, 0.005612544059753418, 0.0055920639038085935, 0.0055797758102417, 0.005598207950592041, 0.005575679779052735, 0.00561356782913208, 0.00563097620010376, 0.005619711875915527, 0.005586944103240967, 0.005567488193511963, 0.005612544059753418, 0.005593088150024414, 0.0056145920753479005, 0.005610496044158936, 0.005575679779052735, 0.005627967834472656, 0.005615551948547363, 0.005611519813537597, 0.005649407863616943, 0.005600255966186523, 0.005618688106536865, 0.005563392162322998, 0.005610496044158936, 0.005587967872619629, 0.005876736164093017, 0.005619711875915527, 0.005603328227996827, 0.005606400012969971, 0.005566463947296142, 0.00561356782913208, 0.005585919857025146, 0.0055582718849182125, 0.005624832153320313, 0.0055797758102417, 0.0056145920753479005, 0.005610496044158936, 0.00556441593170166, 0.00561356782913208, 0.005592127799987793, 0.0056431999206542965, 0.005601280212402344, 0.005575679779052735, 0.0056145920753479005, 0.005605375766754151, 0.005597184181213379, 0.005590015888214111, 0.005554175853729248, 0.005579808235168457, 0.0056135358810424805, 0.005596159934997558, 0.005553152084350586, 0.005568511962890625, 0.011943936347961426, 0.005287936210632324, 0.005318655967712403, 0.005296127796173096, 0.005306367874145508, 0.005361663818359375, 0.005299200057983398, 0.005318655967712403, 0.005276671886444092, 0.00531763219833374, 0.0052633600234985355, 0.005239808082580567, 0.0052633600234985355, 0.005264383792877197, 0.005253119945526123, 0.005261312007904053, 0.0052971520423889164, 0.005307392120361328, 0.005306367874145508, 0.005291007995605469, 0.005272575855255127, 0.005241856098175048, 0.005269504070281982, 0.005262335777282715, 0.005259263992309571, 0.005279744148254394, 0.005352447986602784, 0.005342207908630371, 0.0053043198585510255, 0.0053606400489807126, 0.005332992076873779, 0.005307392120361328, 0.005339136123657226, 0.005328896045684814, 0.005303296089172363, 0.005299200057983398, 0.005341184139251709, 0.005292031764984131, 0.005343232154846191, 0.005347328186035156, 0.005248000144958496, 0.005280767917633057, 0.005311520099639893, 0.0053155522346496585, 0.0052971520423889164, 0.005282815933227539, 0.005319680213928223, 0.005315584182739258, 0.005321728229522705, 0.005314559936523438, 0.005346303939819336, 0.005295104026794434, 0.005281792163848877, 0.005328896045684814, 0.005306367874145508, 0.005291007995605469, 0.005332992076873779, 0.005303296089172363, 0.005294079780578613, 0.005306367874145508, 0.005326848030090332, 0.005299200057983398, 0.005262335777282715, 0.011916288375854492, 0.005311488151550293, 0.0052971520423889164, 0.0052971520423889164, 0.005313536167144775, 0.005283840179443359, 0.005311488151550293, 0.005313536167144775, 0.005320703983306885, 0.005319680213928223, 0.005296127796173096, 0.005323775768280029, 0.0053043198585510255, 0.0053422718048095705, 0.005295040130615235, 0.005279744148254394, 0.005347328186035156, 0.005313536167144775, 0.005326848030090332, 0.005320703983306885, 0.005370880126953125, 0.005285888195037842, 0.005313536167144775, 0.0053350400924682614, 0.005306367874145508, 0.005322751998901368, 0.005327871799468994, 0.005290048122406006, 0.005342144012451172, 0.005278719902038574, 0.005310463905334473, 0.005284863948822022, 0.005295104026794434, 0.005312511920928955, 0.005294079780578613, 0.005280767917633057, 0.005294079780578613, 0.005337088108062744, 0.005313536167144775, 0.005293087959289551, 0.005336031913757324, 0.005291007995605469, 0.005280767917633057, 0.0053095040321350095, 0.005327807903289795, 0.005288959980010987, 0.00531763219833374, 0.0053350400924682614, 0.005293119907379151, 0.005305280208587646, 0.00531763219833374, 0.005307392120361328, 0.0052715520858764645, 0.005324863910675049, 0.00532371187210083, 0.005259263992309571, 0.005305344104766845, 0.00530841588973999, 0.005275648117065429, 0.0053043198585510255, 0.0052971520423889164, 0.005305344104766845, 0.00531766414642334, 0.011898847579956054, 0.005294079780578613, 0.005287936210632324, 0.005338111877441406, 0.005298175811767578, 0.0052899842262268066, 0.005302271842956543, 0.005312511920928955, 0.005283840179443359, 0.0053125758171081545, 0.005312448024749756, 0.005277696132659912, 0.005283840179443359, 0.005269504070281982, 0.0052899842262268066, 0.005305344104766845, 0.005278719902038574, 0.005306367874145508, 0.005310463905334473, 0.0053012800216674805, 0.005300191879272461, 0.0053012480735778805, 0.005298175811767578, 0.0053043198585510255, 0.005321728229522705, 0.005302271842956543, 0.005257215976715088, 0.00530944013595581, 0.005288959980010987, 0.005295104026794434, 0.005282815933227539, 0.005323775768280029, 0.005303296089172363, 0.005295104026794434, 0.005338111877441406, 0.005303296089172363, 0.005273600101470947, 0.005302271842956543, 0.005313536167144775, 0.005276671886444092, 0.005314559936523438, 0.005314559936523438, 0.005277696132659912, 0.005319680213928223, 0.005315584182739258, 0.005329919815063477, 0.005323840141296387, 0.005305280208587646, 0.005310463905334473, 0.005320703983306885, 0.0053012480735778805, 0.005336063861846924, 0.005331967830657959, 0.005280767917633057, 0.005291007995605469, 0.005315584182739258, 0.005292031764984131, 0.0052705278396606445, 0.005313536167144775, 0.005294079780578613, 0.005264383792877197, 0.0053012480735778805, 0.0053012480735778805, 0.011874367713928222, 0.005317567825317383, 0.005286911964416504, 0.005322751998901368, 0.005352447986602784, 0.0053002238273620605, 0.005320703983306885, 0.005291007995605469, 0.005313536167144775, 0.005355519771575928, 0.005281792163848877, 0.005318655967712403, 0.0053002238273620605, 0.005306367874145508, 0.00530944013595581, 0.005313536167144775, 0.005323775768280029, 0.0052408318519592285, 0.005287936210632324, 0.005237760066986084, 0.005220352172851562, 0.005296127796173096, 0.005254144191741943, 0.005307392120361328, 0.00530944013595581, 0.0053002238273620605, 0.005298175811767578, 0.005286911964416504, 0.00535145616531372, 0.005286880016326904, 0.00530841588973999, 0.005276671886444092, 0.005262335777282715, 0.005256192207336426, 0.005259263992309571, 0.005258240222930908, 0.005224448204040527, 0.005244927883148193, 0.005282815933227539, 0.005281792163848877, 0.005319680213928223, 0.0053043198585510255, 0.005307392120361328, 0.0053350400924682614, 0.0052971520423889164, 0.005339136123657226, 0.005299200057983398, 0.005295104026794434, 0.005302271842956543, 0.005349376201629639, 0.005283840179443359, 0.005287936210632324, 0.005328896045684814, 0.005283840179443359, 0.005283840179443359, 0.005310463905334473, 0.005326848030090332, 0.005279744148254394, 0.005277696132659912, 0.005452799797058106, 0.005529600143432617, 0.005601280212402344, 0.0055808000564575196, 0.012562432289123536, 0.005612544059753418, 0.0055797758102417, 0.005602303981781006, 0.005606400012969971, 0.005577727794647217, 0.00556441593170166, 0.005567488193511963, 0.005590015888214111, 0.0055920639038085935, 0.005576704025268555, 0.005616640090942383, 0.005570559978485107, 0.005563392162322998, 0.00563097620010376, 0.005571584224700928, 0.005607423782348633, 0.0056145920753479005, 0.005550079822540284, 0.005532671928405761, 0.005519360065460205, 0.005588992118835449, 0.0055797758102417, 0.005575679779052735, 0.005637119770050049, 0.005577727794647217, 0.005576704025268555, 0.005594111919403076, 0.005583871841430664, 0.005634047985076904, 0.0055920639038085935, 0.005593088150024414, 0.005601280212402344, 0.0055848960876464845, 0.0056228480339050296, 0.005561279773712158, 0.005595168113708496, 0.005597152233123779, 0.005581823825836182, 0.005591040134429932, 0.005667840003967285, 0.005569536209106446, 0.005591040134429932, 0.005597184181213379, 0.005636096000671387, 0.005566463947296142, 0.0055797758102417, 0.0055848960876464845, 0.00556441593170166, 0.005570559978485107, 0.0055848960876464845, 0.005602303981781006, 0.005557248115539551, 0.005574656009674072, 0.005595136165618897, 0.005560319900512695, 0.005628928184509278, 0.005561344146728516, 0.005562431812286377, 0.005573567867279053, 0.005594111919403076, 0.005578752040863037, 0.005597184181213379, 0.01205964756011963, 0.00530944013595581, 0.005325823783874512, 0.005296127796173096, 0.0052899842262268066, 0.005302271842956543, 0.005307392120361328, 0.005307392120361328, 0.00530944013595581, 0.005314559936523438, 0.0053350400924682614, 0.005322751998901368, 0.005311552047729492, 0.005280704021453858, 0.005336063861846924, 0.00570470380783081, 0.005371903896331787, 0.0054568958282470706, 0.0057794561386108395, 0.005954559803009033, 0.005667840003967285, 0.005804031848907471, 0.00572214412689209, 0.005619679927825928, 0.005636096000671387, 0.005616640090942383, 0.005643328189849853, 0.005604288101196289, 0.005566463947296142, 0.0056258559226989744, 0.005587967872619629, 0.005610496044158936, 0.005598207950592041, 0.005574656009674072, 0.005578752040863037, 0.0056145920753479005, 0.00555622386932373, 0.005596159934997558, 0.005574656009674072, 0.005607423782348633, 0.005593088150024414, 0.005566463947296142, 0.0056145920753479005, 0.005549056053161621, 0.005591040134429932, 0.005594111919403076, 0.00557260799407959, 0.005600255966186523, 0.005586944103240967, 0.005599232196807862, 0.00556441593170166, 0.005573631763458252, 0.005593088150024414, 0.005568511962890625, 0.005577727794647217, 0.005597184181213379, 0.00555622386932373, 0.00557260799407959, 0.005596159934997558, 0.005570559978485107, 0.0055582718849182125, 0.005602303981781006, 0.0055920639038085935, 0.012612607955932617, 0.005586944103240967, 0.005586944103240967, 0.005604351997375488, 0.0055920639038085935, 0.005557248115539551, 0.005569536209106446, 0.005628928184509278, 0.00557260799407959, 0.005590015888214111, 0.005569536209106446, 0.00556441593170166, 0.005587967872619629, 0.005595136165618897, 0.005569536209106446, 0.005595136165618897, 0.005565440177917481, 0.005594111919403076, 0.005563392162322998, 0.0055582718849182125, 0.005628928184509278, 0.005550079822540284, 0.005566463947296142, 0.005597184181213379, 0.005560319900512695, 0.005574656009674072, 0.0055706238746643065, 0.005534687995910645, 0.005593056201934814, 0.005577727794647217, 0.005586944103240967, 0.005533696174621582, 0.0054917120933532714, 0.005479423999786377, 0.005464064121246338, 0.005512191772460938, 0.005591040134429932, 0.005570559978485107, 0.005583871841430664, 0.005585919857025146, 0.005588992118835449, 0.005566559791564942, 0.005578656196594239, 0.005608448028564453, 0.005566463947296142, 0.005544960021972656, 0.005609471797943116, 0.005571584224700928, 0.005604351997375488, 0.005588992118835449, 0.005593088150024414, 0.0055920639038085935, 0.005659647941589355, 0.0055920639038085935, 0.005566495895385742, 0.005584864139556884, 0.005591040134429932, 0.005593088150024414, 0.005583871841430664, 0.0055920639038085935, 0.005581823825836182, 0.0055920639038085935, 0.0055808000564575196, 0.011988991737365723, 0.005331967830657959, 0.005288959980010987, 0.0052971520423889164, 0.005351424217224121, 0.00530841588973999, 0.005287936210632324, 0.005322751998901368, 0.005334080219268799, 0.005302207946777344, 0.005328896045684814, 0.005328896045684814, 0.005268479824066162, 0.005299200057983398, 0.005286975860595703, 0.005311488151550293, 0.005281727790832519, 0.005329919815063477, 0.005321760177612305, 0.005285855770111084, 0.005313536167144775, 0.005296127796173096, 0.005321728229522705, 0.005283840179443359, 0.005281792163848877, 0.00532480001449585, 0.005295104026794434, 0.005283872127532959, 0.00533296012878418, 0.0052930560111999515, 0.005299200057983398, 0.005274623870849609, 0.005325823783874512, 0.005252096176147461, 0.005232639789581299, 0.005264383792877197, 0.0052674560546875, 0.005234687805175781, 0.0053012480735778805, 0.005328959941864013, 0.005301184177398681, 0.005313536167144775, 0.0053350720405578615, 0.0052930240631103515, 0.0053095040321350095, 0.005286848068237305, 0.005315584182739258, 0.005320767879486084, 0.005309375762939453, 0.005306367874145508, 0.005331967830657959, 0.0053043198585510255, 0.005292031764984131, 0.005284863948822022, 0.005298175811767578, 0.005272575855255127, 0.005325823783874512, 0.005284863948822022, 0.005284863948822022, 0.005341184139251709, 0.005288959980010987, 0.005288959980010987, 0.00531763219833374, 0.011937791824340821, 0.005320703983306885, 0.00531660795211792, 0.00531763219833374, 0.005320703983306885, 0.005320703983306885, 0.005318655967712403, 0.0053012480735778805, 0.005303296089172363, 0.005310495853424072, 0.005317599773406983, 0.005291007995605469, 0.005307392120361328, 0.005322751998901368, 0.005296127796173096, 0.005294079780578613, 0.005330976009368897, 0.005309408187866211, 0.005283840179443359, 0.005325823783874512, 0.005295199871063232, 0.005296031951904297, 0.005349376201629639, 0.005314559936523438, 0.0052715520858764645, 0.005323775768280029, 0.005320703983306885, 0.0052899842262268066, 0.00530944013595581, 0.005324831962585449, 0.005285855770111084, 0.005325823783874512, 0.005312511920928955, 0.005326848030090332, 0.005302271842956543, 0.00530944013595581, 0.005292031764984131, 0.005310463905334473, 0.005275648117065429, 0.005299200057983398, 0.005296127796173096, 0.005298175811767578, 0.005292031764984131, 0.005336063861846924, 0.005278719902038574, 0.005272575855255127, 0.005292096138000488, 0.005264319896697998, 0.0052408318519592285, 0.005253119945526123, 0.005283840179443359, 0.005269504070281982, 0.005264383792877197, 0.005337088108062744, 0.005291071891784668, 0.005299136161804199, 0.005303296089172363, 0.005337088108062744, 0.005321728229522705, 0.0053637118339538575, 0.00531763219833374, 0.005292031764984131, 0.0053043198585510255, 0.011889663696289063, 0.005299200057983398, 0.0052930560111999515, 0.0053350400924682614, 0.005303296089172363, 0.005299200057983398, 0.005313536167144775, 0.005296127796173096, 0.0052899842262268066, 0.005332992076873779, 0.005318655967712403, 0.005287936210632324, 0.005278719902038574, 0.005338111877441406, 0.005312511920928955, 0.005306367874145508, 0.005340159893035889, 0.005261312007904053, 0.005259263992309571, 0.005232639789581299, 0.005261312007904053, 0.005248000144958496, 0.005245952129364013, 0.005266464233398437, 0.005290976047515869, 0.0052899842262268066, 0.0053043198585510255, 0.00530841588973999, 0.005327871799468994, 0.005295104026794434, 0.005314559936523438, 0.00530944013595581, 0.005314591884613037, 0.005310431957244873, 0.005319680213928223, 0.005325823783874512, 0.005299200057983398, 0.005339136123657226, 0.005310463905334473, 0.0053043198585510255, 0.005343232154846191, 0.00530841588973999, 0.005303296089172363, 0.005307424068450928, 0.005319647789001465, 0.005280767917633057, 0.005296127796173096, 0.005323775768280029, 0.005296127796173096, 0.0053350400924682614, 0.005298175811767578, 0.005331967830657959, 0.005303296089172363, 0.005295104026794434, 0.005319712162017823, 0.005299168109893799, 0.005281792163848877, 0.005312511920928955, 0.005294079780578613, 0.005275648117065429, 0.005296192169189453, 0.005351359844207764, 0.0052971520423889164, 0.011904000282287597, 0.005277696132659912, 0.005291007995605469, 0.005320703983306885, 0.005284863948822022, 0.005284863948822022, 0.005310463905334473, 0.005321728229522705, 0.005272575855255127, 0.005325823783874512, 0.00532480001449585, 0.005295104026794434, 0.0053309440612792965, 0.005313536167144775, 0.005318655967712403, 0.005313536167144775, 0.005299200057983398, 0.005323775768280029, 0.005323775768280029, 0.005307424068450928, 0.005336031913757324, 0.005321728229522705, 0.005285888195037842, 0.0052899842262268066, 0.00532480001449585, 0.005283840179443359, 0.005291007995605469, 0.0052971520423889164, 0.005303296089172363, 0.0052726402282714845, 0.005334976196289062, 0.005322751998901368, 0.0052971520423889164, 0.005315584182739258, 0.005323775768280029, 0.005399551868438721, 0.005299200057983398, 0.005272575855255127, 0.005298175811767578, 0.005569536209106446, 0.005554175853729248, 0.005559296131134033, 0.0055972480773925784, 0.005552063941955567, 0.005629951953887939, 0.00555622386932373, 0.005567488193511963, 0.005615615844726562, 0.0055623679161071774, 0.005565440177917481, 0.005561344146728516, 0.00557369613647461, 0.0055725440979003905, 0.0055623679161071774, 0.005602303981781006, 0.0055511040687561035, 0.005576704025268555, 0.0055808000564575196, 0.005567488193511963, 0.005582848072052002, 0.005594111919403076, 0.005578752040863037, 0.005554175853729248, 0.011973631858825684, 0.005305344104766845, 0.005278719902038574, 0.005341184139251709, 0.005314559936523438, 0.0052971520423889164, 0.005319680213928223, 0.0053012480735778805, 0.005292031764984131, 0.005341184139251709, 0.005323775768280029, 0.005285888195037842, 0.0052798080444335935, 0.005326784133911132, 0.0052715520858764645, 0.005283840179443359, 0.0052971520423889164, 0.005296127796173096, 0.005278719902038574, 0.005272575855255127, 0.0053012480735778805, 0.00530841588973999, 0.005305344104766845, 0.005299200057983398, 0.00530944013595581, 0.005293119907379151, 0.0052622718811035155, 0.005346303939819336, 0.0053043198585510255, 0.0052899842262268066, 0.005325823783874512, 0.005303359985351562, 0.005273536205291748, 0.005273600101470947, 0.005331967830657959, 0.005286911964416504, 0.0052675199508667, 0.005338047981262207, 0.005282815933227539, 0.005286911964416504, 0.005298175811767578, 0.005325823783874512, 0.005273600101470947, 0.005298175811767578, 0.00530841588973999, 0.0052899842262268066, 0.00530841588973999, 0.005302271842956543, 0.005314559936523438, 0.005320703983306885, 0.005295104026794434, 0.005311488151550293, 0.005305344104766845, 0.005341184139251709, 0.005347328186035156, 0.005319680213928223, 0.005298175811767578, 0.005280767917633057, 0.005327871799468994, 0.0052971520423889164, 0.005285888195037842, 0.005285888195037842, 0.005296127796173096, 0.011891712188720703, 0.005285888195037842, 0.0053043198585510255, 0.0052971520423889164, 0.005315584182739258, 0.00530944013595581, 0.005303296089172363, 0.005318655967712403, 0.005315584182739258, 0.0053043198585510255, 0.0053012480735778805, 0.00531763219833374, 0.005313536167144775, 0.005288959980010987, 0.005312511920928955, 0.005302271842956543, 0.005282815933227539, 0.0052971520423889164, 0.005328896045684814, 0.005288959980010987, 0.005277696132659912, 0.0053350400924682614, 0.005291007995605469, 0.005292031764984131, 0.005326848030090332, 0.005302271842956543, 0.0052971520423889164, 0.005298175811767578, 0.005326848030090332, 0.005279744148254394, 0.005315584182739258, 0.0053309440612792965, 0.005296127796173096, 0.005315584182739258, 0.00530944013595581, 0.00531763219833374, 0.0053422718048095705, 0.005303232192993164, 0.005327871799468994, 0.005328896045684814, 0.005328896045684814, 0.0052971520423889164, 0.00531763219833374, 0.005325823783874512, 0.0052930560111999515, 0.005319680213928223, 0.00531660795211792, 0.005311488151550293, 0.005334047794342041, 0.005297120094299316, 0.005293119907379151, 0.005305280208587646, 0.005315584182739258, 0.0052674560546875, 0.0052971520423889164, 0.005307392120361328, 0.0053002238273620605, 0.005314559936523438, 0.0052899842262268066, 0.005314559936523438, 0.005305344104766845, 0.005323775768280029, 0.0053002238273620605, 0.011902976036071777, 0.0053340158462524415, 0.005319680213928223, 0.005295104026794434, 0.005303328037261963, 0.005331935882568359, 0.005315584182739258, 0.00531660795211792, 0.005347328186035156, 0.005314559936523438, 0.0053002238273620605, 0.005345280170440674, 0.0052971520423889164, 0.0053002238273620605, 0.005386240005493164, 0.005343232154846191, 0.005368832111358642, 0.005314591884613037, 0.00534115219116211, 0.0053043198585510255, 0.00531660795211792, 0.005294079780578613, 0.0052941122055053715, 0.005305312156677246, 0.005287936210632324, 0.005284863948822022, 0.00531660795211792, 0.005285888195037842, 0.005296127796173096, 0.005303296089172363, 0.005331967830657959, 0.005280767917633057, 0.0053340158462524415, 0.005303296089172363, 0.0052930560111999515, 0.00532480001449585, 0.005286975860595703, 0.005278656005859375, 0.005295104026794434, 0.005323775768280029, 0.005279744148254394, 0.005298175811767578, 0.005313536167144775, 0.005286911964416504, 0.005315584182739258, 0.005274623870849609, 0.005319680213928223, 0.005299200057983398, 0.005265471935272217, 0.005259200096130371, 0.005223423957824707, 0.005265408039093018, 0.005255167961120606, 0.00521727991104126, 0.005248000144958496, 0.005291007995605469, 0.005338111877441406, 0.005307392120361328, 0.005283872127532959, 0.005303264141082764, 0.005314559936523438, 0.005276671886444092, 0.005275648117065429, 0.011950079917907714, 0.005311488151550293, 0.005331967830657959, 0.0052899842262268066, 0.005288959980010987, 0.005302271842956543, 0.005268479824066162, 0.005282815933227539, 0.005283840179443359, 0.0052971520423889164, 0.005315584182739258, 0.005299200057983398, 0.005305344104766845, 0.005319680213928223, 0.005305344104766845, 0.005299200057983398, 0.0057190399169921875, 0.005601280212402344, 0.005575679779052735, 0.005617663860321045, 0.005573631763458252, 0.005578752040863037, 0.005596159934997558, 0.0055706238746643065, 0.0055715198516845706, 0.005608448028564453, 0.005581823825836182, 0.005600255966186523, 0.005583871841430664, 0.005613632202148437, 0.005562304019927978, 0.005585919857025146, 0.005624832153320313, 0.005573631763458252, 0.005590015888214111, 0.00562278413772583, 0.005601280212402344, 0.005586944103240967, 0.005590015888214111, 0.005600255966186523, 0.005583871841430664, 0.005597184181213379, 0.005589056015014649, 0.005558207988739013, 0.005577727794647217, 0.005593088150024414, 0.005601280212402344, 0.00561356782913208, 0.005609471797943116, 0.005608448028564453, 0.00557369613647461, 0.005648320198059082, 0.005568511962890625, 0.005549056053161621, 0.005596159934997558, 0.005573631763458252, 0.005571648120880127, 0.005548992156982422, 0.0055808000564575196, 0.005596159934997558, 0.005565440177917481, 0.0057497601509094234, 0.005672959804534912, 0.012422143936157226, 0.005566463947296142, 0.005587967872619629, 0.00556441593170166, 0.005569536209106446, 0.0055848960876464845, 0.005549056053161621, 0.0055920639038085935, 0.005554175853729248, 0.00555622386932373, 0.0055706238746643065, 0.005551040172576904, 0.0055552000999450684, 0.0055920639038085935, 0.005603328227996827, 0.005538815975189209, 0.0055808000564575196, 0.005567488193511963, 0.005550079822540284, 0.005559296131134033, 0.005567488193511963, 0.005285888195037842, 0.005303296089172363, 0.00532480001449585, 0.005298175811767578, 0.0053002238273620605, 0.005313536167144775, 0.005306367874145508, 0.005455872058868408, 0.005596159934997558, 0.0055797758102417, 0.005582848072052002, 0.005588992118835449, 0.0056217598915100095, 0.005570559978485107, 0.005555263996124268, 0.005626815795898438, 0.005571584224700928, 0.005594111919403076, 0.00556441593170166, 0.005574656009674072, 0.005594111919403076, 0.005582848072052002, 0.005617663860321045, 0.005588992118835449, 0.005306367874145508, 0.005407743930816651, 0.00532480001449585, 0.005361663818359375, 0.005320703983306885, 0.005307392120361328, 0.005287936210632324, 0.00531660795211792, 0.005348351955413819, 0.005287936210632324, 0.005306367874145508, 0.00530944013595581, 0.00530841588973999, 0.0053002238273620605, 0.005285888195037842, 0.005306367874145508, 0.005282815933227539, 0.005326848030090332, 0.011965503692626953, 0.005288896083831787, 0.005310463905334473, 0.005295104026794434, 0.005283840179443359, 0.0053350400924682614, 0.005328896045684814, 0.005304384231567383, 0.00528275203704834, 0.0053350400924682614, 0.0053043198585510255, 0.005291007995605469, 0.005311488151550293, 0.005322751998901368, 0.005291007995605469, 0.005306367874145508, 0.005321728229522705, 0.00532480001449585, 0.005321728229522705, 0.005313536167144775, 0.005277696132659912, 0.005305344104766845, 0.005276671886444092, 0.005472256183624268, 0.005659647941589355, 0.005574656009674072, 0.005605440139770507, 0.0055715198516845706, 0.005567488193511963, 0.005605375766754151, 0.005607423782348633, 0.005576704025268555, 0.00557260799407959, 0.005645311832427978, 0.005577727794647217, 0.005577727794647217, 0.005595136165618897, 0.0055552000999450684, 0.005578815937042236, 0.005623744010925293, 0.005582848072052002, 0.005585919857025146, 0.005601280212402344, 0.0055623679161071774, 0.005603328227996827, 0.005571584224700928, 0.0056145920753479005, 0.005568511962890625, 0.005575679779052735, 0.005627903938293457, 0.005595136165618897, 0.0056145920753479005, 0.005658624172210694, 0.0055848960876464845, 0.005598207950592041, 0.005646336078643799, 0.005628928184509278, 0.005598207950592041, 0.005596159934997558, 0.005612544059753418, 0.0056258559226989744, 0.00567299222946167, 0.005635039806365967, 0.011905023574829102, 0.005312511920928955, 0.005278719902038574, 0.005310463905334473, 0.005296127796173096, 0.005291007995605469, 0.005281792163848877, 0.005272575855255127, 0.0053350400924682614, 0.005273600101470947, 0.0053012480735778805, 0.005345280170440674, 0.005307392120361328, 0.0052971520423889164, 0.005322751998901368, 0.005319680213928223, 0.005344255924224854, 0.005318655967712403, 0.005342207908630371, 0.005283840179443359, 0.005281792163848877, 0.005380095958709717, 0.00531763219833374, 0.00530944013595581, 0.005307392120361328, 0.005310463905334473, 0.005269504070281982, 0.005315584182739258, 0.005312511920928955, 0.005302271842956543, 0.00530841588973999, 0.005307392120361328, 0.00531763219833374, 0.005311488151550293, 0.00530944013595581, 0.00531660795211792, 0.00531660795211792, 0.005288959980010987, 0.005287936210632324, 0.0053361282348632815, 0.0053257598876953124, 0.005291007995605469, 0.0053340158462524415, 0.005282815933227539, 0.005269504070281982, 0.005328896045684814, 0.005288959980010987, 0.005288959980010987, 0.005303296089172363, 0.0053350400924682614, 0.005294079780578613, 0.005321728229522705, 0.005314559936523438, 0.005287936210632324, 0.005302271842956543, 0.005319744110107422, 0.005316544055938721, 0.005302271842956543, 0.00530841588973999, 0.00530944013595581, 0.005296127796173096, 0.005268479824066162, 0.005272575855255127, 0.011892736434936524, 0.005307392120361328, 0.005331967830657959, 0.005295104026794434, 0.005287936210632324, 0.005597184181213379, 0.0053043198585510255, 0.005325856208801269, 0.005304287910461425, 0.00531763219833374, 0.005260287761688232, 0.005268479824066162, 0.005221375942230225, 0.005234687805175781, 0.005261312007904053, 0.005251071929931641, 0.005299200057983398, 0.005303296089172363, 0.00530944013595581, 0.00535859203338623, 0.005305344104766845, 0.005281792163848877, 0.0053043198585510255, 0.005251071929931641, 0.005239808082580567, 0.00521727991104126, 0.005255199909210205, 0.005204959869384766, 0.005269504070281982, 0.005356544017791748, 0.00530944013595581, 0.005303296089172363, 0.005279744148254394, 0.005278719902038574, 0.005260287761688232, 0.005272575855255127, 0.005296127796173096, 0.005244927883148193, 0.005303296089172363, 0.005336063861846924, 0.00530944013595581, 0.005288959980010987, 0.005318655967712403, 0.005339136123657226, 0.005312511920928955, 0.0053534722328186036, 0.005268479824066162, 0.00526643180847168, 0.005305344104766845, 0.005253119945526123, 0.005274623870849609, 0.005342207908630371, 0.005312511920928955, 0.005329919815063477, 0.0053350400924682614, 0.0052899842262268066, 0.005305344104766845, 0.005355519771575928, 0.0053012480735778805, 0.0052971520423889164, 0.0053534722328186036, 0.0053002238273620605, 0.005286911964416504, 0.01194598388671875, 0.0052971520423889164, 0.005275648117065429, 0.00530841588973999, 0.005235712051391602, 0.005227519989013672, 0.005250048160552978, 0.005250048160552978, 0.005244927883148193, 0.005253119945526123, 0.005303296089172363, 0.005319680213928223, 0.005311488151550293, 0.0053350400924682614, 0.005311488151550293, 0.0053002238273620605, 0.005355519771575928, 0.005319680213928223, 0.005295104026794434, 0.0053002238273620605, 0.005339136123657226, 0.0053012480735778805, 0.005401599884033203, 0.005858304023742676, 0.005696512222290039, 0.005636096000671387, 0.0061562881469726565, 0.00611846399307251, 0.005662655830383301, 0.005594111919403076, 0.005597184181213379, 0.005582848072052002, 0.005624832153320313, 0.005591040134429932, 0.005603328227996827, 0.0055848960876464845, 0.005588992118835449, 0.005633024215698243, 0.005608448028564453, 0.005599232196807862, 0.005587007999420166, 0.0055459198951721194, 0.005535744190216065, 0.005660672187805176, 0.005563392162322998, 0.005597184181213379, 0.00561356782913208, 0.005602303981781006, 0.005583871841430664, 0.005598207950592041, 0.005598207950592041, 0.005585919857025146, 0.005602303981781006, 0.005561344146728516, 0.0055808000564575196, 0.005591040134429932, 0.005587967872619629, 0.005576704025268555, 0.0055552000999450684, 0.005603328227996827, 0.005569536209106446, 0.005595136165618897, 0.005616640090942383, 0.013248512268066406, 0.0056145920753479005, 0.0056217598915100095, 0.005565440177917481, 0.005629951953887939, 0.005826560020446778, 0.005923840045928955, 0.0056145920753479005, 0.0057794561386108395, 0.005601280212402344, 0.005612544059753418, 0.005627903938293457, 0.005567488193511963, 0.0055797758102417, 0.005610496044158936, 0.005574656009674072, 0.005588992118835449, 0.005764095783233642, 0.005599232196807862, 0.005582911968231201, 0.005565375804901123, 0.005581823825836182, 0.0055920639038085935, 0.005603328227996827, 0.005700607776641845, 0.005560319900512695, 0.005587967872619629, 0.005582848072052002, 0.005563392162322998, 0.005607456207275391, 0.005568480014801026, 0.005565440177917481, 0.005583871841430664, 0.005587967872619629, 0.005599232196807862, 0.0055797758102417, 0.0055920639038085935, 0.005569536209106446, 0.005574656009674072, 0.005626880168914795, 0.005557248115539551, 0.005550079822540284, 0.005561344146728516, 0.005575679779052735, 0.0055797758102417, 0.005550079822540284, 0.00563097620010376, 0.005506048202514649, 0.005473279953002929, 0.005506048202514649, 0.005479455947875977, 0.005454815864562988, 0.005476352214813233, 0.005481472015380859, 0.005493760108947754, 0.005483520030975342, 0.0054906878471374515, 0.005483520030975342, 0.005540863990783691, 0.005602303981781006, 0.005764095783233642, 0.005581823825836182, 0.005598207950592041, 0.012050432205200195, 0.005320703983306885, 0.005327871799468994, 0.005310463905334473, 0.005311488151550293, 0.005477375984191894, 0.0053136000633239745, 0.005304255962371826, 0.005339136123657226, 0.0053975038528442385, 0.005338111877441406, 0.005326848030090332, 0.005414912223815918, 0.005331967830657959, 0.0055552000999450684, 0.005946368217468262, 0.005643263816833496, 0.005636096000671387, 0.005635072231292724, 0.005635072231292724, 0.005618688106536865, 0.005612544059753418, 0.005659647941589355, 0.005570559978485107, 0.005653503894805908, 0.005640223979949951, 0.005593056201934814, 0.005597184181213379, 0.005610496044158936, 0.0055623679161071774, 0.005550079822540284, 0.005569568157196045, 0.0055797438621521, 0.00556441593170166, 0.005594111919403076, 0.005825535774230957, 0.00563097620010376, 0.005676032066345215, 0.005600255966186523, 0.005602303981781006, 0.00562278413772583, 0.005583871841430664, 0.0056217598915100095, 0.005587967872619629, 0.005605375766754151, 0.005576704025268555, 0.005565440177917481, 0.00561356782913208, 0.005573631763458252, 0.005595136165618897, 0.005582848072052002, 0.005611519813537597, 0.00563097620010376, 0.005561344146728516, 0.005594111919403076, 0.005575679779052735, 0.005568511962890625, 0.00561356782913208, 0.005548031806945801, 0.005571584224700928, 0.005575679779052735, 0.005467135906219483, 0.005471231937408447, 0.01206272029876709, 0.005315584182739258, 0.005288959980010987, 0.005325823783874512, 0.00531763219833374, 0.005287936210632324, 0.005352511882781983, 0.005310400009155274, 0.005281792163848877, 0.005313536167144775, 0.005545983791351319, 0.0055797758102417, 0.005605375766754151, 0.005583871841430664, 0.005587967872619629, 0.005577727794647217, 0.0056217598915100095, 0.005593088150024414, 0.00557369613647461, 0.005607359886169434, 0.005575679779052735, 0.005573631763458252, 0.00562278413772583, 0.00566374397277832, 0.005568511962890625, 0.00555622386932373, 0.005482495784759522, 0.005484543800354004, 0.005462016105651855, 0.005499904155731201, 0.005468160152435303, 0.005453824043273926, 0.005497856140136719, 0.0055582718849182125, 0.005570559978485107, 0.005595136165618897, 0.005595136165618897, 0.005570559978485107, 0.005591040134429932, 0.005590015888214111, 0.005620736122131348, 0.005545983791351319, 0.005588992118835449, 0.005599232196807862, 0.005570559978485107, 0.005593088150024414, 0.005582848072052002, 0.005561344146728516, 0.00551423978805542, 0.0054906878471374515, 0.0055101442337036136, 0.005472256183624268, 0.005576767921447754, 0.00563097620010376, 0.005550015926361084, 0.005597184181213379, 0.005582848072052002, 0.0055552000999450684, 0.0056258559226989744, 0.005593088150024414, 0.005600255966186523, 0.005600255966186523, 0.005530623912811279, 0.012511232376098632, 0.005624832153320313, 0.0055808000564575196, 0.005608448028564453, 0.005595136165618897, 0.005469183921813964, 0.005497856140136719, 0.005538815975189209, 0.005585919857025146, 0.005581823825836182, 0.0055808000564575196, 0.005618688106536865, 0.005581823825836182, 0.0056145920753479005, 0.005588992118835449, 0.005459968090057373, 0.00547430419921875, 0.0054917120933532714, 0.005479423999786377, 0.005581823825836182, 0.005595136165618897, 0.005595136165618897, 0.005599232196807862, 0.0056145920753479005, 0.005586944103240967, 0.005578752040863037, 0.005569536209106446, 0.005593088150024414, 0.005578752040863037, 0.005573631763458252, 0.0055797758102417, 0.0055920639038085935, 0.005588992118835449, 0.005604351997375488, 0.0055797758102417, 0.0055582718849182125, 0.005560319900512695, 0.005598207950592041, 0.005496831893920898, 0.005467135906219483, 0.005274623870849609, 0.005295104026794434, 0.005314559936523438, 0.005325823783874512, 0.005287936210632324, 0.00530841588973999, 0.0052930560111999515, 0.0052971520423889164, 0.005506048202514649, 0.005451776027679443, 0.005349376201629639, 0.005351424217224121, 0.005310463905334473, 0.005231616020202637, 0.005344255924224854, 0.005307392120361328, 0.005286911964416504, 0.005355519771575928, 0.0052705278396606445, 0.005303296089172363, 0.005325823783874512, 0.005303296089172363, 0.0052787837982177736, 0.012004287719726562, 0.0053637118339538575, 0.005303296089172363, 0.005304384231567383, 0.0053144960403442385, 0.005305344104766845, 0.005272575855255127, 0.005310463905334473, 0.005287936210632324, 0.005380095958709717, 0.005361663818359375, 0.005279744148254394, 0.005284863948822022, 0.0053606400489807126, 0.0052930560111999515, 0.005349376201629639, 0.005315584182739258, 0.005319680213928223, 0.005283840179443359, 0.0053002238273620605, 0.005313536167144775, 0.005285888195037842, 0.005307392120361328, 0.00530841588973999, 0.005307392120361328, 0.005326848030090332, 0.005303296089172363, 0.0053095040321350095, 0.0052848000526428225, 0.005706751823425293, 0.005704736232757569, 0.005619679927825928, 0.005792768001556397, 0.006269951820373535, 0.005735424041748047, 0.005627903938293457, 0.005603328227996827, 0.005620831966400146, 0.005599135875701904, 0.0055808000564575196, 0.005634047985076904, 0.005567488193511963, 0.005617663860321045, 0.005592127799987793, 0.005610432147979736, 0.0055623679161071774, 0.005583871841430664, 0.005586944103240967, 0.005590015888214111, 0.005554175853729248, 0.005598207950592041, 0.005561344146728516, 0.00562278413772583, 0.005603328227996827, 0.005637119770050049, 0.00557260799407959, 0.005600255966186523, 0.005593088150024414, 0.00557260799407959, 0.005583871841430664, 0.005591040134429932, 0.005590015888214111, 0.00556441593170166]",tokens/s,180.04549022628876,,,main,False,False,,,,,,,,,,,,,,,,,,,,,,, float32-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,google/gemma-2b,google/gemma-2b,cuda,0,42,,,True,,,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.223-212.873.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.0,,,,1.21.4,,,,0.12.0,,,gemma,MB,1343.438848,12778.471424,0.0,12138.315776,10311.21664,s,10,0.8156489486694337,0.08156489486694335,0.00026383762270427026,0.08143335723876953,0.08200370941162109,0.08204460601806642,0.08207732330322266,"[0.08208550262451172, 0.08133808135986328, 0.08127609252929688, 0.08170243072509766, 0.08199462127685547, 0.0815745620727539, 0.08143218994140625, 0.08142899322509765, 0.08143452453613281, 0.08138195037841797]",tokens/s,3138.605161173962,kWh,1.0009059733757487e-06,5.484465569808549e-07,7.017536216660991e-06,8.566888747017595e-06,tokens/kWh,29882493.815403134,MB,1379.606528,12784.76288,0.0,12144.607232,10311.2192,s,10,14.554965942382813,1.4554965942382814,0.0012251774912942325,1.4552077026367187,1.457224951171875,1.457596728515625,1.457894150390625,"[1.457968505859375, 1.457142333984375, 1.454336181640625, 1.4546436767578126, 1.4541785888671874, 1.4558375244140624, 1.4560927734375, 1.4543509521484375, 1.455671142578125, 1.4547442626953124]",tokens/s,43.284196094577865,kWh,1.7176849155652193e-05,9.412911301357747e-06,7.350488931233878e-05,0.00010009464976934871,tokens/kWh,629404.2703098807,,s,630,14.553220096588127,0.023100349359663704,9.49616218882025e-05,0.02307072067260742,0.023177318000793458,0.023253095054626464,0.02350734275817871,"[0.02366361618041992, 0.02324684715270996, 0.023104511260986327, 0.02308095932006836, 0.023079935073852538, 0.02309836769104004, 0.023165952682495116, 0.023331840515136718, 0.023386112213134767, 0.023153663635253906, 0.02309222412109375, 0.02312294387817383, 0.023035903930664063, 0.023061504364013673, 0.023085056304931642, 0.023061504364013673, 0.02310553550720215, 0.023182336807250976, 0.023076864242553712, 0.02308710479736328, 0.023171072006225587, 0.023119871139526366, 0.02306355285644531, 0.02308710479736328, 0.023142400741577147, 0.02306764793395996, 0.023047168731689452, 0.02309939193725586, 0.02307379150390625, 0.023093248367309572, 0.02304102325439453, 0.0230645751953125, 0.02307174491882324, 0.023214080810546874, 0.023152639389038086, 0.02312499237060547, 0.023079935073852538, 0.02326323127746582, 0.023178239822387696, 0.02312396812438965, 0.02306355285644531, 0.023153663635253906, 0.023158784866333007, 0.02313113594055176, 0.023136255264282226, 0.02309119987487793, 0.023193599700927735, 0.023152639389038086, 0.02314854431152344, 0.02306768035888672, 0.023196640014648436, 0.02309939193725586, 0.023112703323364257, 0.023188480377197264, 0.023174144744873046, 0.02309734344482422, 0.02308198356628418, 0.023130111694335938, 0.02308095932006836, 0.02326937675476074, 0.02315673637390137, 0.023185407638549805, 0.023206911087036132, 0.0235100154876709, 0.02325196838378906, 0.02312499237060547, 0.02331340789794922, 0.0230830078125, 0.02309734344482422, 0.023130111694335938, 0.02312704086303711, 0.023136255264282226, 0.023096319198608398, 0.02306559944152832, 0.02305740737915039, 0.02312499237060547, 0.02309119987487793, 0.023144447326660156, 0.02308095932006836, 0.02308403205871582, 0.02315673637390137, 0.023219200134277345, 0.02307583999633789, 0.02311680030822754, 0.023052288055419923, 0.023202816009521485, 0.023197696685791015, 0.023069696426391603, 0.02305433654785156, 0.023069696426391603, 0.02305433654785156, 0.02306559944152832, 0.02304614448547363, 0.02309222412109375, 0.02313523292541504, 0.02309427261352539, 0.023500799179077148, 0.023146495819091797, 0.02307174491882324, 0.02309836769104004, 0.023059455871582032, 0.02308198356628418, 0.02305536079406738, 0.02314035224914551, 0.023068672180175782, 0.02307379150390625, 0.02307583999633789, 0.02310246467590332, 0.02307583999633789, 0.02307072067260742, 0.023068672180175782, 0.02309017562866211, 0.02307379150390625, 0.02306764793395996, 0.02307891273498535, 0.023069696426391603, 0.02308095932006836, 0.023068672180175782, 0.023047168731689452, 0.0231014404296875, 0.023153663635253906, 0.023119871139526366, 0.023638015747070314, 0.023194623947143556, 0.023158784866333007, 0.02311884880065918, 0.023478271484375, 0.023254016876220703, 0.023177215576171875, 0.023085056304931642, 0.023059455871582032, 0.02305740737915039, 0.023069696426391603, 0.023130111694335938, 0.02305638313293457, 0.02304921531677246, 0.02306662368774414, 0.02305843162536621, 0.023144447326660156, 0.023059455871582032, 0.02304614448547363, 0.02306662368774414, 0.02305023956298828, 0.023036928176879884, 0.023038976669311522, 0.023059455871582032, 0.02304921531677246, 0.02304614448547363, 0.02305023956298828, 0.02307788848876953, 0.0230645751953125, 0.02306355285644531, 0.023027711868286133, 0.02305740737915039, 0.023085056304931642, 0.023031808853149413, 0.02304204750061035, 0.023059455871582032, 0.02306559944152832, 0.023130111694335938, 0.02306355285644531, 0.023069696426391603, 0.02307379150390625, 0.02307481575012207, 0.023061504364013673, 0.023037952423095705, 0.023076864242553712, 0.023093248367309572, 0.02306252861022949, 0.02304204750061035, 0.023059455871582032, 0.02308710479736328, 0.023085056304931642, 0.02305536079406738, 0.02305740737915039, 0.023068672180175782, 0.0230645751953125, 0.02304819107055664, 0.02305536079406738, 0.02308710479736328, 0.023093248367309572, 0.0231014404296875, 0.02304819107055664, 0.023242752075195314, 0.02310041618347168, 0.023078943252563478, 0.023092191696166994, 0.023088127136230468, 0.02311065673828125, 0.02352025604248047, 0.023244800567626952, 0.023164928436279295, 0.02309836769104004, 0.02305740737915039, 0.023414783477783203, 0.023193599700927735, 0.023085056304931642, 0.02306662368774414, 0.02304819107055664, 0.02303385543823242, 0.02305843162536621, 0.023040000915527343, 0.023052288055419923, 0.023051263809204102, 0.023076864242553712, 0.023038976669311522, 0.023051263809204102, 0.02304102325439453, 0.02305740737915039, 0.02307072067260742, 0.023059455871582032, 0.023052288055419923, 0.023040000915527343, 0.02306559944152832, 0.0230645751953125, 0.02304614448547363, 0.02304921531677246, 0.02307174491882324, 0.02306252861022949, 0.02311884880065918, 0.02305740737915039, 0.02306764793395996, 0.023093248367309572, 0.02305843162536621, 0.023159807205200195, 0.023060480117797853, 0.02308403205871582, 0.02309119987487793, 0.02304102325439453, 0.023044095993041993, 0.02305740737915039, 0.023040000915527343, 0.02304204750061035, 0.02308403205871582, 0.02307788848876953, 0.02305536079406738, 0.02310041618347168, 0.02305843162536621, 0.0230830078125, 0.02308403205871582, 0.023068672180175782, 0.02305638313293457, 0.023061504364013673, 0.02308915138244629, 0.02307379150390625, 0.02306252861022949, 0.02306252861022949, 0.02306559944152832, 0.02306662368774414, 0.02311065673828125, 0.02308608055114746, 0.023133184432983397, 0.023413759231567383, 0.02323865509033203, 0.02313113594055176, 0.02308403205871582, 0.02306662368774414, 0.023060480117797853, 0.02305023956298828, 0.02307075119018555, 0.023035871505737306, 0.02305232048034668, 0.023048160552978515, 0.02306764793395996, 0.02305638313293457, 0.023165952682495116, 0.023059455871582032, 0.02311884880065918, 0.02307174491882324, 0.023051263809204102, 0.02302566337585449, 0.023044095993041993, 0.023069696426391603, 0.02304102325439453, 0.02304819107055664, 0.023045120239257814, 0.02305740737915039, 0.023061504364013673, 0.02304204750061035, 0.02304921531677246, 0.02307174491882324, 0.02305232048034668, 0.023036895751953126, 0.023040000915527343, 0.02306662368774414, 0.02308403205871582, 0.02305843162536621, 0.023052288055419923, 0.02310041618347168, 0.023059455871582032, 0.02305023956298828, 0.023045120239257814, 0.02304614448547363, 0.02307788848876953, 0.023069696426391603, 0.023059455871582032, 0.02305433654785156, 0.02308710479736328, 0.023136255264282226, 0.02312704086303711, 0.023061504364013673, 0.02305638313293457, 0.02307583999633789, 0.02306252861022949, 0.02308095932006836, 0.023068672180175782, 0.02307891273498535, 0.023061504364013673, 0.02307379150390625, 0.02310553550720215, 0.023188480377197264, 0.0231147518157959, 0.023104511260986327, 0.02307788848876953, 0.023093248367309572, 0.023373823165893554, 0.023240703582763672, 0.023134208679199218, 0.02312396812438965, 0.02306764793395996, 0.02305843162536621, 0.02309017562866211, 0.02307891273498535, 0.0230645751953125, 0.02306355285644531, 0.02306559944152832, 0.02306559944152832, 0.02306764793395996, 0.023060480117797853, 0.02305740737915039, 0.02307276725769043, 0.023044095993041993, 0.02304102325439453, 0.023028736114501954, 0.023339008331298827, 0.023213056564331053, 0.02311680030822754, 0.02323865509033203, 0.023037952423095705, 0.02307174491882324, 0.02308710479736328, 0.023040000915527343, 0.02305023956298828, 0.02305023956298828, 0.02308608055114746, 0.02307072067260742, 0.02305536079406738, 0.023038976669311522, 0.02307481575012207, 0.02306355285644531, 0.02305843162536621, 0.02327654457092285, 0.023076864242553712, 0.02309836769104004, 0.023053312301635744, 0.02305843162536621, 0.023068672180175782, 0.023164928436279295, 0.023096319198608398, 0.02308915138244629, 0.02304614448547363, 0.023079935073852538, 0.023051263809204102, 0.023044095993041993, 0.02311680030822754, 0.02374553680419922, 0.02322329521179199, 0.02311577606201172, 0.02308403205871582, 0.023052288055419923, 0.02305740737915039, 0.023076864242553712, 0.02307379150390625, 0.023069696426391603, 0.023044095993041993, 0.0231014404296875, 0.0231147518157959, 0.02310041618347168, 0.023389184951782226, 0.02327654457092285, 0.023133184432983397, 0.02309017562866211, 0.02309939193725586, 0.023111679077148437, 0.02310860824584961, 0.0230645751953125, 0.023076864242553712, 0.023143423080444335, 0.02411110305786133, 0.023138303756713868, 0.02309119987487793, 0.0230645751953125, 0.023436288833618164, 0.02306764793395996, 0.02305638313293457, 0.02304102325439453, 0.02306252861022949, 0.023040000915527343, 0.02305536079406738, 0.023052288055419923, 0.023047168731689452, 0.02307174491882324, 0.02306662368774414, 0.02304819107055664, 0.02304102325439453, 0.02306252861022949, 0.02305536079406738, 0.023043071746826172, 0.0230328311920166, 0.023053312301635744, 0.023052288055419923, 0.02305740737915039, 0.02305536079406738, 0.02305433654785156, 0.0230830078125, 0.023060480117797853, 0.023061504364013673, 0.023043071746826172, 0.02305638313293457, 0.02307174491882324, 0.02308710479736328, 0.02305638313293457, 0.0230830078125, 0.023059455871582032, 0.02305638313293457, 0.023069696426391603, 0.02307583999633789, 0.02306559944152832, 0.023068672180175782, 0.02304614448547363, 0.02307379150390625, 0.02309222412109375, 0.02309222412109375, 0.02306662368774414, 0.02305740737915039, 0.023121919631958008, 0.023186431884765626, 0.023423999786376954, 0.023085056304931642, 0.02310758399963379, 0.023117824554443358, 0.023427072525024413, 0.02328678321838379, 0.02314035224914551, 0.02309734344482422, 0.023053312301635744, 0.02307072067260742, 0.023079935073852538, 0.02306355285644531, 0.0230645751953125, 0.02305433654785156, 0.02308608055114746, 0.023069696426391603, 0.02305433654785156, 0.02304819107055664, 0.02306559944152832, 0.02304614448547363, 0.023035903930664063, 0.023047168731689452, 0.023069696426391603, 0.023037952423095705, 0.02304614448547363, 0.02305638313293457, 0.02306355285644531, 0.02305539131164551, 0.023090143203735352, 0.02303385543823242, 0.02304614448547363, 0.02306764793395996, 0.02305740737915039, 0.023045120239257814, 0.02305433654785156, 0.02305740737915039, 0.023051263809204102, 0.023037952423095705, 0.02305023956298828, 0.02306559944152832, 0.023088127136230468, 0.023463935852050782, 0.023126016616821288, 0.02305536079406738, 0.02308198356628418, 0.023068672180175782, 0.02305638313293457, 0.023053312301635744, 0.023068672180175782, 0.0230830078125, 0.023047168731689452, 0.023040000915527343, 0.02307072067260742, 0.02308198356628418, 0.023085056304931642, 0.023059455871582032, 0.02309427261352539, 0.02307072067260742, 0.023060480117797853, 0.02306252861022949, 0.02306252861022949, 0.02309119987487793, 0.02307583999633789, 0.023052288055419923, 0.02309119987487793, 0.02311680030822754, 0.023104511260986327, 0.023476224899291992, 0.023257087707519532, 0.023145471572875977, 0.02306252861022949, 0.02305843162536621, 0.02306559944152832, 0.023158784866333007, 0.02306764793395996, 0.023053312301635744, 0.023047168731689452, 0.02306355285644531, 0.02306559944152832, 0.02304819107055664, 0.023052288055419923, 0.0230645751953125, 0.02305433654785156, 0.023161855697631836, 0.023040000915527343, 0.0230645751953125, 0.02304614448547363, 0.02305740737915039, 0.02304921531677246, 0.02306252861022949, 0.023111679077148437, 0.023037952423095705, 0.023034879684448242, 0.023021568298339845, 0.02307174491882324, 0.02312396812438965, 0.023162879943847657, 0.023220224380493162, 0.0230645751953125, 0.02309222412109375, 0.023060480117797853, 0.02308710479736328, 0.02323148727416992, 0.023189504623413085, 0.02310246467590332, 0.02309119987487793, 0.02309529685974121, 0.023076864242553712, 0.023088127136230468, 0.02306252861022949, 0.02305843162536621, 0.023051263809204102, 0.023078943252563478, 0.023062496185302733, 0.023051263809204102, 0.02305843162536621, 0.02308608055114746, 0.02307174491882324, 0.02306662368774414, 0.02309939193725586, 0.023074848175048828, 0.023066591262817383, 0.0230645751953125, 0.02308198356628418, 0.02330419158935547, 0.023093248367309572, 0.02329804801940918, 0.023194623947143556, 0.023215103149414062, 0.02312294387817383, 0.023574527740478517, 0.02329497528076172, 0.023217151641845703, 0.023151615142822265, 0.02313523292541504, 0.023093248367309572, 0.02305536079406738, 0.02307072067260742, 0.02304819107055664, 0.02305740737915039, 0.023052288055419923, 0.02311577606201172, 0.02305843162536621, 0.02305740737915039, 0.02304102325439453, 0.023209983825683594, 0.02311884880065918, 0.02310246467590332, 0.02309427261352539, 0.02308095932006836, 0.02306355285644531, 0.02307174491882324, 0.023035903930664063, 0.023047168731689452, 0.02305638313293457, 0.023061504364013673, 0.02305638313293457, 0.023036928176879884, 0.023120895385742187, 0.023068672180175782, 0.02304819107055664, 0.02303385543823242, 0.023047168731689452, 0.02307174491882324, 0.02307583999633789, 0.02304204750061035, 0.02305433654785156, 0.02306764793395996, 0.023059455871582032, 0.023174144744873046, 0.02305433654785156, 0.023085056304931642, 0.023093248367309572, 0.02305740737915039, 0.023052288055419923, 0.023059455871582032, 0.023071775436401366, 0.023059423446655274, 0.023061504364013673, 0.02307788848876953, 0.02307583999633789, 0.023061504364013673, 0.02304921531677246, 0.023053312301635744, 0.02308915138244629, 0.02306764793395996, 0.02307891273498535, 0.02305740737915039, 0.02307379150390625, 0.02309734344482422, 0.0231014404296875, 0.02309222412109375, 0.0230830078125]",tokens/s,43.289388590206066,,,,,,MB,936.738816,12753.3056,0.0,12123.635712,12121.851904,s,1,6.92231787109375,6.92231787109375,0.0,6.92231787109375,6.92231787109375,6.92231787109375,6.92231787109375,[6.92231787109375],,kWh,1.8815690666659497e-06,9.811789664445764e-07,5.1641707979957485e-06,8.026918831106275e-06, float32-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,s,s,cuda,0,42,,,True,True,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/s/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66947346-6367fb272e7841bf3d3b39bb;ff678302-bfd1-4828-99da-81b9bf86e423) Repository Not Found for url: https://huggingface.co/s/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: s is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-40b,tiiuae/falcon-40b,cuda,0,42,,,True,True,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ self.load_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 250, in load_model_with_no_weights self.pretrained_model = self.automodel_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 559, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3710, in from_pretrained model = cls(config, *model_args, **model_kwargs) File ""/root/.cache/huggingface/modules/transformers_modules/tiiuae/falcon-40b/4a70170c215b36a3cce4b4253f6d0612bb7d4146/modeling_falcon.py"", line 843, in __init__ self.transformer = FalconModel(config) File ""/root/.cache/huggingface/modules/transformers_modules/tiiuae/falcon-40b/4a70170c215b36a3cce4b4253f6d0612bb7d4146/modeling_falcon.py"", line 650, in __init__ self.h = nn.ModuleList([FalconDecoderLayer(config) for _ in range(config.num_hidden_layers)]) File ""/root/.cache/huggingface/modules/transformers_modules/tiiuae/falcon-40b/4a70170c215b36a3cce4b4253f6d0612bb7d4146/modeling_falcon.py"", line 650, in self.h = nn.ModuleList([FalconDecoderLayer(config) for _ in range(config.num_hidden_layers)]) File ""/root/.cache/huggingface/modules/transformers_modules/tiiuae/falcon-40b/4a70170c215b36a3cce4b4253f6d0612bb7d4146/modeling_falcon.py"", line 420, in __init__ self.mlp = FalconMLP(config) File ""/root/.cache/huggingface/modules/transformers_modules/tiiuae/falcon-40b/4a70170c215b36a3cce4b4253f6d0612bb7d4146/modeling_falcon.py"", line 405, in __init__ self.dense_4h_to_h = FalconLinear(4 * hidden_size, hidden_size, bias=config.bias) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/linear.py"", line 98, in __init__ self.weight = Parameter(torch.empty((out_features, in_features), **factory_kwargs)) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_device.py"", line 78, in __torch_function__ return func(*args, **kwargs) torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 1024.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,google/gemma-7b,google/gemma-7b,cuda,0,42,,,True,,,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",gemma,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-125m,,cuda,0,42,,,,,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,d35829e539df8480b726c647eeabf91e41eae047,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,,MB,1063.75168,1407.71328,0.0,767.557632,582.168576,s,21,0.1947245435714722,0.009272597312927247,0.00021781516510231797,0.009289728164672852,0.009383296012878418,0.009395232200622558,0.009951929283142091,"[0.010091103553771973, 0.009395232200622558, 0.009306591987609863, 0.009295680046081544, 0.009269760131835938, 0.009262399673461914, 0.00908140754699707, 0.00902518367767334, 0.009122688293457032, 0.009079360008239746, 0.009098719596862793, 0.009125344276428223, 0.009377408027648925, 0.009383296012878418, 0.009289728164672852, 0.00908460807800293, 0.009369824409484863, 0.009115679740905762, 0.009338720321655274, 0.009321439743041992, 0.00929036808013916]",tokens/s,27608.23007412406,kWh,1.0924001234581738e-07,5.985812403584038e-08,5.493814139112079e-07,7.184795502928656e-07,tokens/kWh,356307983.84679097,MB,1063.75168,1407.71328,0.0,767.557632,601.35168,s,21,9.825072296142576,0.46786058553059895,0.010540369583952693,0.4737200622558594,0.4762123718261719,0.47770535278320314,0.48149368286132815,"[0.4824407653808594, 0.47770535278320314, 0.4737200622558594, 0.47511422729492186, 0.4717841796875, 0.46407546997070315, 0.45037109375, 0.4500010681152344, 0.4645487670898438, 0.45069720458984375, 0.4492754821777344, 0.4541901550292969, 0.4758885498046875, 0.4745844421386719, 0.47551278686523435, 0.4616695556640625, 0.47356802368164064, 0.4762123718261719, 0.47462060546875, 0.474175537109375, 0.47491659545898435]",tokens/s,134.65549770248745,kWh,5.414431724275077e-06,2.9656232714894354e-06,1.0707395268271899e-05,1.9087450264036412e-05,tokens/kWh,3300597.9912729017,,s,1322,10.00183500909806,0.007565684575717136,0.0011966433587927779,0.007502848148345947,0.007592857503890991,0.007640012574195862,0.01670176706314087,"[0.008300543785095215, 0.00819916820526123, 0.0077844481468200685, 0.00773529577255249, 0.007729152202606201, 0.008226816177368163, 0.007456768035888672, 0.007576576232910156, 0.007879680156707763, 0.007640063762664795, 0.00757862377166748, 0.0076267518997192385, 0.007599103927612305, 0.007607295989990235, 0.00759500789642334, 0.007608320236206055, 0.007559232234954834, 0.007600063800811767, 0.007585792064666748, 0.007584767818450928, 0.0075632638931274416, 0.007607295989990235, 0.007623680114746094, 0.007610367774963379, 0.007542784214019775, 0.007569407939910889, 0.0075980801582336424, 0.007565343856811523, 0.007591904163360596, 0.007634943962097168, 0.0076267518997192385, 0.007617536067962646, 0.007566336154937744, 0.007716864109039307, 0.00764518404006958, 0.007629824161529541, 0.007544832229614258, 0.007621632099151611, 0.007622655868530274, 0.007685120105743408, 0.00758681583404541, 0.00764415979385376, 0.007619647979736328, 0.007620543956756592, 0.007577600002288819, 0.007600128173828125, 0.007590911865234375, 0.007613440036773681, 0.007550975799560547, 0.007611392021179199, 0.007664639949798584, 0.007614463806152344, 0.007576608180999756, 0.0076195521354675295, 0.007622655868530274, 0.007616511821746826, 0.007544832229614258, 0.007592959880828858, 0.0075673599243164065, 0.007565311908721924, 0.007600128173828125, 0.007658495903015137, 0.01760256004333496, 0.007525375843048096, 0.007614463806152344, 0.007632895946502686, 0.007675903797149658, 0.00758681583404541, 0.007621632099151611, 0.007591936111450195, 0.007475200176239013, 0.007392255783081054, 0.007476223945617676, 0.007523327827453613, 0.007924736022949219, 0.007790592193603516, 0.007881728172302246, 0.007913472175598145, 0.007945216178894043, 0.008432640075683593, 0.00780083179473877, 0.0076277761459350585, 0.007532544136047363, 0.007527423858642578, 0.007514111995697022, 0.007481344223022461, 0.00753868818283081, 0.007469056129455566, 0.007498752117156982, 0.007523327827453613, 0.007520287990570068, 0.007478240013122558, 0.007525375843048096, 0.007525375843048096, 0.00759500789642334, 0.007483391761779785, 0.007510015964508057, 0.00754585599899292, 0.0075673599243164065, 0.007524352073669433, 0.007559167861938477, 0.007490560054779053, 0.007453695774078369, 0.007485439777374267, 0.007493631839752197, 0.007497727870941162, 0.007511040210723877, 0.0074926080703735356, 0.007549952030181885, 0.007502848148345947, 0.0074629120826721195, 0.007512063980102539, 0.007485439777374267, 0.007507967948913574, 0.007540736198425293, 0.007507967948913574, 0.00753868818283081, 0.007531519889831543, 0.007449600219726562, 0.007519231796264648, 0.0075038719177246095, 0.007532544136047363, 0.0075304961204528805, 0.007532544136047363, 0.007536640167236328, 0.017312768936157227, 0.007496767997741699, 0.007485375881195068, 0.0075008001327514645, 0.007497727870941162, 0.00748035192489624, 0.007506912231445313, 0.007489535808563232, 0.00749567985534668, 0.007518208026885987, 0.007514111995697022, 0.007489535808563232, 0.007525375843048096, 0.007479296207427978, 0.007506944179534912, 0.007494656085968018, 0.007501823902130127, 0.007501823902130127, 0.007536640167236328, 0.007536672115325928, 0.007531487941741943, 0.007481344223022461, 0.007478271961212158, 0.007426047801971435, 0.0074291200637817386, 0.0074158720970153805, 0.0074321279525756834, 0.007415808200836181, 0.007448575973510742, 0.007518208026885987, 0.007488512039184571, 0.007528448104858398, 0.00754585599899292, 0.00753766393661499, 0.007523327827453613, 0.0075038719177246095, 0.007525375843048096, 0.0074700798988342285, 0.007504896163940429, 0.007522367954254151, 0.007546815872192383, 0.007568384170532226, 0.007584832191467285, 0.0076481919288635255, 0.00765337610244751, 0.0075335679054260255, 0.0074741759300231934, 0.007517183780670166, 0.007571455955505371, 0.007550047874450684, 0.007541664123535156, 0.007550975799560547, 0.007549952030181885, 0.007513088226318359, 0.007585824012756348, 0.0074966721534729, 0.0074926080703735356, 0.0075304961204528805, 0.00753868818283081, 0.007502848148345947, 0.007515135765075683, 0.007484416007995606, 0.007517183780670166, 0.017341440200805663, 0.007561215877532959, 0.007498752117156982, 0.00753766393661499, 0.0074967041015625, 0.007528448104858398, 0.007516160011291504, 0.007502848148345947, 0.007524352073669433, 0.007513088226318359, 0.007516160011291504, 0.007526432037353516, 0.007489503860473633, 0.007494656085968018, 0.007543807983398438, 0.0074997758865356446, 0.007536704063415527, 0.007594944000244141, 0.007516160011291504, 0.007519231796264648, 0.007479296207427978, 0.0074997758865356446, 0.007498752117156982, 0.007511040210723877, 0.007523327827453613, 0.007519231796264648, 0.00753766393661499, 0.007569407939910889, 0.0074967041015625, 0.00759500789642334, 0.007494656085968018, 0.007527423858642578, 0.007549952030181885, 0.007513088226318359, 0.007486464023590088, 0.007432191848754883, 0.0074106879234313965, 0.007392255783081054, 0.007428095817565918, 0.00748748779296875, 0.007585792064666748, 0.007819263935089112, 0.008184831619262695, 0.00760319995880127, 0.007625728130340576, 0.007622655868530274, 0.007514111995697022, 0.0075673599243164065, 0.0075038719177246095, 0.007516160011291504, 0.007580671787261963, 0.007516191959381103, 0.007508959770202637, 0.00753769588470459, 0.00750486421585083, 0.0074189119338989255, 0.007421919822692871, 0.007453695774078369, 0.007550975799560547, 0.007532544136047363, 0.0075335679054260255, 0.0075345921516418455, 0.007562240123748779, 0.017300479888916014, 0.007506944179534912, 0.007513088226318359, 0.007473152160644531, 0.007457791805267334, 0.007461887836456299, 0.007457791805267334, 0.007481344223022461, 0.007488512039184571, 0.007452672004699707, 0.0074997758865356446, 0.007481344223022461, 0.007487552165985107, 0.007470016002655029, 0.007514111995697022, 0.007508992195129394, 0.007529471874237061, 0.007497727870941162, 0.0074997758865356446, 0.0075008001327514645, 0.007513088226318359, 0.007456768035888672, 0.007505951881408691, 0.007483424186706543, 0.0075140480995178225, 0.007486464023590088, 0.007490560054779053, 0.007475200176239013, 0.0074926080703735356, 0.0074629120826721195, 0.007475200176239013, 0.00753868818283081, 0.007491583824157715, 0.007505919933319092, 0.0074967041015625, 0.007460864067077637, 0.007481344223022461, 0.007478271961212158, 0.0074414081573486324, 0.007490560054779053, 0.007478335857391358, 0.007490560054779053, 0.007482304096221924, 0.007498752117156982, 0.00743936014175415, 0.007513088226318359, 0.0075008001327514645, 0.007493631839752197, 0.007516160011291504, 0.007504896163940429, 0.007504896163940429, 0.0074711041450500485, 0.007452672004699707, 0.007486464023590088, 0.0074997758865356446, 0.007505919933319092, 0.007507967948913574, 0.007484416007995606, 0.007400447845458984, 0.007374847888946533, 0.007362559795379638, 0.007392255783081054, 0.00744755220413208, 0.01724313545227051, 0.007507967948913574, 0.007516160011291504, 0.0074997758865356446, 0.007481344223022461, 0.007479328155517578, 0.007471072196960449, 0.007493696212768555, 0.0074915199279785155, 0.007475200176239013, 0.007506944179534912, 0.007525375843048096, 0.007513088226318359, 0.007518271923065186, 0.007470016002655029, 0.00749567985534668, 0.007469056129455566, 0.007507967948913574, 0.007506944179534912, 0.007521279811859131, 0.007549952030181885, 0.007521279811859131, 0.007139328002929687, 0.007129087924957276, 0.007202816009521484, 0.007168000221252442, 0.007143424034118652, 0.007139328002929687, 0.00714137601852417, 0.007139328002929687, 0.007139328002929687, 0.00710041618347168, 0.007114751815795899, 0.007120895862579346, 0.007129087924957276, 0.007146495819091797, 0.0071393918991088864, 0.007093183994293213, 0.007176191806793213, 0.007165952205657959, 0.007162879943847656, 0.007107583999633789, 0.007151616096496582, 0.007155712127685547, 0.0072202239036560055, 0.007147520065307617, 0.007180287837982178, 0.007337984085083008, 0.00760422420501709, 0.007541759967803955, 0.007507967948913574, 0.007472127914428711, 0.007431168079376221, 0.0075335679054260255, 0.007488512039184571, 0.0074967041015625, 0.007506944179534912, 0.00748748779296875, 0.007483391761779785, 0.007493631839752197, 0.007494656085968018, 0.0075335679054260255, 0.007507967948913574, 0.016677888870239257, 0.00710041618347168, 0.007175168037414551, 0.007139328002929687, 0.007146495819091797, 0.007122943878173828, 0.0071198720932006835, 0.007134208202362061, 0.007150591850280762, 0.0072130560874938965, 0.0071485438346862796, 0.0071823358535766605, 0.007144447803497315, 0.007158783912658692, 0.007132160186767578, 0.007127039909362793, 0.007178239822387696, 0.007195648193359375, 0.007150591850280762, 0.007205887794494629, 0.007107583999633789, 0.007161856174468994, 0.007132160186767578, 0.007071743965148926, 0.007009280204772949, 0.007147520065307617, 0.007161856174468994, 0.007160831928253173, 0.007116799831390381, 0.007130112171173096, 0.007153664112091064, 0.007169023990631103, 0.007161856174468994, 0.007159808158874512, 0.00714137601852417, 0.0071495680809020995, 0.007158783912658692, 0.007131135940551757, 0.007124991893768311, 0.007132160186767578, 0.007172160148620606, 0.007142335891723633, 0.007291903972625732, 0.007153664112091064, 0.007157760143280029, 0.007135231971740722, 0.007145472049713135, 0.007114751815795899, 0.007110655784606934, 0.007139328002929687, 0.007134208202362061, 0.007142399787902832, 0.007079936027526855, 0.007166975975036621, 0.007134208202362061, 0.007143424034118652, 0.007124991893768311, 0.007116799831390381, 0.007147520065307617, 0.00714035177230835, 0.007123968124389648, 0.00713318395614624, 0.007090176105499267, 0.01657753562927246, 0.007130112171173096, 0.007132160186767578, 0.007131135940551757, 0.007108607769012451, 0.007105535984039306, 0.007131135940551757, 0.007135231971740722, 0.007121920108795166, 0.00707583999633789, 0.007104512214660645, 0.007157760143280029, 0.007178239822387696, 0.007098368167877197, 0.007108607769012451, 0.007142399787902832, 0.0071792640686035155, 0.0071792640686035155, 0.007146495819091797, 0.007123968124389648, 0.007135231971740722, 0.007136256217956543, 0.007111711978912354, 0.007097311973571777, 0.0071198720932006835, 0.007143424034118652, 0.007124991893768311, 0.007129087924957276, 0.007124991893768311, 0.0071485438346862796, 0.007171072006225586, 0.007122943878173828, 0.007152703762054444, 0.007118783950805664, 0.007180287837982178, 0.00719155216217041, 0.007145535945892334, 0.007102399826049805, 0.007150591850280762, 0.007138304233551026, 0.007186431884765625, 0.007111680030822754, 0.007088128089904785, 0.007176191806793213, 0.007131135940551757, 0.0071526398658752445, 0.007136256217956543, 0.007116799831390381, 0.007138368129730225, 0.007160799980163574, 0.007143392086029053, 0.0071055998802185055, 0.00715667200088501, 0.007157760143280029, 0.007112703800201416, 0.007120895862579346, 0.007138336181640625, 0.007147488117218018, 0.007195648193359375, 0.007138304233551026, 0.007112703800201416, 0.007169023990631103, 0.00714137601852417, 0.016661504745483398, 0.007150591850280762, 0.00709939193725586, 0.007097343921661377, 0.007114751815795899, 0.007134208202362061, 0.007132160186767578, 0.007114751815795899, 0.007105535984039306, 0.007155712127685547, 0.007135231971740722, 0.007144447803497315, 0.007096320152282715, 0.007117824077606201, 0.0071393918991088864, 0.007132095813751221, 0.0071157760620117185, 0.007158783912658692, 0.007078911781311035, 0.0071495680809020995, 0.007124991893768311, 0.007871488094329833, 0.007566336154937744, 0.007483391761779785, 0.007526400089263916, 0.0074926080703735356, 0.0074741759300231934, 0.007485439777374267, 0.007480319976806641, 0.0074700798988342285, 0.007488512039184571, 0.0074414081573486324, 0.0074414081573486324, 0.007470111846923828, 0.007455711841583252, 0.007469056129455566, 0.007543807983398438, 0.007484416007995606, 0.007472127914428711, 0.0075038719177246095, 0.00749567985534668, 0.0074997758865356446, 0.007544832229614258, 0.007549983978271484, 0.007407584190368652, 0.0073820161819458, 0.007391232013702393, 0.007387135982513428, 0.007386112213134765, 0.0075090560913085936, 0.007497663974761963, 0.0075335679054260255, 0.007479296207427978, 0.007508992195129394, 0.00748748779296875, 0.0074967041015625, 0.0074670081138610836, 0.007463935852050781, 0.007461887836456299, 0.007488512039184571, 0.007514111995697022, 0.0074629120826721195, 0.007465983867645264, 0.016703487396240235, 0.007176191806793213, 0.007142399787902832, 0.007135231971740722, 0.007150623798370361, 0.0071075520515441895, 0.007110655784606934, 0.007126016139984131, 0.007120895862579346, 0.007128064155578613, 0.007108607769012451, 0.007144447803497315, 0.007127039909362793, 0.007168000221252442, 0.007142399787902832, 0.007108607769012451, 0.007145472049713135, 0.00714137601852417, 0.007151616096496582, 0.007142399787902832, 0.00709939193725586, 0.007138304233551026, 0.00714035177230835, 0.007131135940551757, 0.007102464199066162, 0.007134208202362061, 0.007195648193359375, 0.007131135940551757, 0.007107583999633789, 0.007121920108795166, 0.007147520065307617, 0.007151616096496582, 0.007144447803497315, 0.007132160186767578, 0.007112703800201416, 0.007147552013397217, 0.007179232120513916, 0.007143424034118652, 0.007432191848754883, 0.0074332160949707035, 0.007153664112091064, 0.0071485438346862796, 0.007121920108795166, 0.007185408115386963, 0.00713318395614624, 0.007139328002929687, 0.0071823358535766605, 0.007122943878173828, 0.007130112171173096, 0.007159808158874512, 0.007159808158874512, 0.007137279987335205, 0.007113728046417236, 0.007128064155578613, 0.007135231971740722, 0.007144447803497315, 0.007130112171173096, 0.007137279987335205, 0.007120895862579346, 0.007145472049713135, 0.007143424034118652, 0.0071188478469848635, 0.007132160186767578, 0.016663551330566406, 0.007114751815795899, 0.007136256217956543, 0.007134208202362061, 0.007131135940551757, 0.0072130560874938965, 0.007105535984039306, 0.0071526398658752445, 0.007170048236846924, 0.007156735897064209, 0.0070891518592834475, 0.007151616096496582, 0.00709939193725586, 0.007123968124389648, 0.007122975826263428, 0.007086048126220703, 0.007117824077606201, 0.007036928176879883, 0.007058432102203369, 0.007001088142395019, 0.007051263809204102, 0.007136256217956543, 0.007130112171173096, 0.007111680030822754, 0.00710041618347168, 0.007130112171173096, 0.00714137601852417, 0.007202847957611084, 0.007120863914489746, 0.0071823358535766605, 0.007143424034118652, 0.007101439952850342, 0.007104512214660645, 0.007108607769012451, 0.007137279987335205, 0.007142399787902832, 0.007128064155578613, 0.007101439952850342, 0.0071495680809020995, 0.007170048236846924, 0.007149631977081299, 0.007115712165832519, 0.0071157760620117185, 0.0072540159225463864, 0.007175168037414551, 0.0071526398658752445, 0.007153664112091064, 0.007132160186767578, 0.007135231971740722, 0.007163904190063477, 0.007171072006225586, 0.0071270718574523925, 0.007160799980163574, 0.0071495680809020995, 0.007150591850280762, 0.007113728046417236, 0.007074816226959229, 0.007067647933959961, 0.007101439952850342, 0.007060480117797851, 0.0070594558715820314, 0.007124991893768311, 0.0070594558715820314, 0.016663551330566406, 0.007132160186767578, 0.007127039909362793, 0.007186431884765625, 0.007117824077606201, 0.007159808158874512, 0.007142399787902832, 0.007168000221252442, 0.007129087924957276, 0.007098368167877197, 0.007201791763305664, 0.007175168037414551, 0.007159808158874512, 0.0071495680809020995, 0.0071485438346862796, 0.007154687881469727, 0.007230463981628418, 0.00709939193725586, 0.007130112171173096, 0.007164927959442138, 0.007163904190063477, 0.007164927959442138, 0.00714137601852417, 0.007074816226959229, 0.007060480117797851, 0.007096320152282715, 0.007045119762420654, 0.007102464199066162, 0.007157760143280029, 0.0071485438346862796, 0.007154687881469727, 0.00713321590423584, 0.007133152008056641, 0.007165952205657959, 0.007164927959442138, 0.00714137601852417, 0.007124991893768311, 0.007142399787902832, 0.007178239822387696, 0.007177216053009033, 0.00713318395614624, 0.0071485438346862796, 0.007169023990631103, 0.007127039909362793, 0.0072499198913574215, 0.007218175888061523, 0.007163904190063477, 0.007251967906951904, 0.007173120021820068, 0.007177216053009033, 0.007123968124389648, 0.007128064155578613, 0.007174143791198731, 0.007158783912658692, 0.007161856174468994, 0.007432191848754883, 0.007244800090789795, 0.0072202239036560055, 0.008088576316833495, 0.0077916159629821775, 0.007622655868530274, 0.007580671787261963, 0.007570464134216309, 0.017376224517822267, 0.00753868818283081, 0.008096768379211425, 0.007606272220611572, 0.007543807983398438, 0.007547904014587403, 0.0074967041015625, 0.007507967948913574, 0.007464960098266602, 0.007493631839752197, 0.007511040210723877, 0.007618559837341309, 0.007508992195129394, 0.007550975799560547, 0.007543807983398438, 0.0074926080703735356, 0.007505919933319092, 0.007543807983398438, 0.0075632638931274416, 0.0075304961204528805, 0.007579648017883301, 0.007518208026885987, 0.007542784214019775, 0.007504896163940429, 0.007515135765075683, 0.007490560054779053, 0.007522304058074952, 0.007510015964508057, 0.007514111995697022, 0.007532544136047363, 0.007535615921020508, 0.0075345921516418455, 0.007473152160644531, 0.007503967761993408, 0.00756931209564209, 0.007511040210723877, 0.0075345921516418455, 0.007532544136047363, 0.007502912044525146, 0.007499711990356445, 0.007453695774078369, 0.0074967041015625, 0.007517183780670166, 0.007506944179534912, 0.007560192108154297, 0.007520256042480469, 0.007526400089263916, 0.007612415790557861, 0.007514111995697022, 0.007532544136047363, 0.007507967948913574, 0.0076277761459350585, 0.007540736198425293, 0.007561215877532959, 0.00783564805984497, 0.007583744049072265, 0.007547904014587403, 0.007521279811859131, 0.007572480201721191, 0.00753766393661499, 0.007592959880828858, 0.00752953577041626, 0.007631807804107666, 0.017369087219238282, 0.0074967041015625, 0.007536640167236328, 0.007553023815155029, 0.007504896163940429, 0.0075345921516418455, 0.007501823902130127, 0.007525375843048096, 0.007482367992401123, 0.007560192108154297, 0.0075008001327514645, 0.007524352073669433, 0.007518208026885987, 0.007517183780670166, 0.007508992195129394, 0.007570432186126709, 0.007475200176239013, 0.007521279811859131, 0.007583744049072265, 0.007504896163940429, 0.007521279811859131, 0.007526400089263916, 0.007582719802856445, 0.0075121278762817385, 0.0075181441307067874, 0.007501823902130127, 0.0075939841270446775, 0.007519231796264648, 0.0075304961204528805, 0.007536640167236328, 0.007527423858642578, 0.007511104106903076, 0.007532512187957764, 0.007502816200256348, 0.007497727870941162, 0.007501823902130127, 0.007588863849639893, 0.007504896163940429, 0.007501823902130127, 0.007512063980102539, 0.0075304961204528805, 0.007585792064666748, 0.007548927783966064, 0.0074700798988342285, 0.007528448104858398, 0.007539711952209473, 0.007553023815155029, 0.007517183780670166, 0.007515135765075683, 0.007625728130340576, 0.007512063980102539, 0.007507967948913574, 0.007505919933319092, 0.007522304058074952, 0.007517183780670166, 0.00753766393661499, 0.007554048061370849, 0.00748748779296875, 0.007493631839752197, 0.007515135765075683, 0.007502848148345947, 0.007516160011291504, 0.007609344005584716, 0.017302528381347656, 0.007511040210723877, 0.007507967948913574, 0.007511040210723877, 0.007557119846343994, 0.007519231796264648, 0.00753766393661499, 0.0075335679054260255, 0.007506944179534912, 0.007478271961212158, 0.007510015964508057, 0.007532544136047363, 0.007544832229614258, 0.007546879768371582, 0.007512063980102539, 0.008287263870239257, 0.007617504119873047, 0.007558144092559814, 0.00749567985534668, 0.007516160011291504, 0.007522304058074952, 0.007557119846343994, 0.007528448104858398, 0.007507967948913574, 0.007522304058074952, 0.007504896163940429, 0.007460864067077637, 0.00748748779296875, 0.0074967041015625, 0.007549952030181885, 0.007535615921020508, 0.007518208026885987, 0.007520256042480469, 0.007518208026885987, 0.007491583824157715, 0.007528448104858398, 0.007535615921020508, 0.007527423858642578, 0.007555136203765869, 0.007522240161895752, 0.007547904014587403, 0.007524352073669433, 0.00753868818283081, 0.007573503971099854, 0.00753766393661499, 0.007497727870941162, 0.007568384170532226, 0.0075304961204528805, 0.007534656047821045, 0.0075447678565979, 0.007524352073669433, 0.007555071830749512, 0.007523327827453613, 0.007489535808563232, 0.007528448104858398, 0.00753868818283081, 0.007532544136047363, 0.007515135765075683, 0.007608320236206055, 0.007576576232910156, 0.007506944179534912, 0.007558144092559814, 0.00748748779296875, 0.016695295333862305, 0.0071198720932006835, 0.007171135902404785, 0.007152575969696045, 0.007150591850280762, 0.0071485438346862796, 0.00714137601852417, 0.0071495680809020995, 0.0071495680809020995, 0.007137279987335205, 0.007142399787902832, 0.00714137601852417, 0.0071495680809020995, 0.007153664112091064, 0.007126016139984131, 0.007130144119262695, 0.007153632164001465, 0.007162879943847656, 0.007159808158874512, 0.007143424034118652, 0.007113728046417236, 0.007156735897064209, 0.007214079856872559, 0.007160831928253173, 0.007113728046417236, 0.007153664112091064, 0.007153664112091064, 0.007145472049713135, 0.007231488227844239, 0.00714035177230835, 0.007175168037414551, 0.007180287837982178, 0.007164927959442138, 0.007153664112091064, 0.007129087924957276, 0.007181312084197998, 0.007170048236846924, 0.007155712127685547, 0.007184383869171143, 0.007320576190948487, 0.008632320404052735, 0.007762944221496582, 0.0076267518997192385, 0.007546879768371582, 0.007570432186126709, 0.007541759967803955, 0.00754585599899292, 0.007508992195129394, 0.007527423858642578, 0.0075162239074707034, 0.007516096115112305, 0.0075673599243164065, 0.007576576232910156, 0.007526400089263916, 0.007585792064666748, 0.007515135765075683, 0.007548927783966064, 0.007515135765075683, 0.007543807983398438, 0.007514111995697022, 0.007555071830749512, 0.007550975799560547, 0.007623712062835694, 0.017332191467285155, 0.007562240123748779, 0.007532544136047363, 0.007535615921020508, 0.007526400089263916, 0.007614463806152344, 0.007512063980102539, 0.007476223945617676, 0.007513088226318359, 0.0074700798988342285, 0.007512063980102539, 0.007579648017883301, 0.007535615921020508, 0.007501823902130127, 0.0075345921516418455, 0.007488512039184571, 0.007520256042480469, 0.007494656085968018, 0.007619584083557129, 0.007488512039184571, 0.007501823902130127, 0.0075642881393432615, 0.007523327827453613, 0.007508992195129394, 0.007498752117156982, 0.007481344223022461, 0.007508992195129394, 0.00821452808380127, 0.007549952030181885, 0.0075642881393432615, 0.00753868818283081, 0.007476223945617676, 0.007525375843048096, 0.0074997758865356446, 0.007504896163940429, 0.007493631839752197, 0.007532544136047363, 0.007541759967803955, 0.00759500789642334, 0.007529471874237061, 0.00753766393661499, 0.007530528068542481, 0.007550943851470948, 0.007535615921020508, 0.0075632638931274416, 0.008127488136291505, 0.007316480159759522, 0.007829504013061523, 0.007287807941436767, 0.007670783996582031, 0.007845888137817383, 0.0075970559120178225, 0.007161856174468994, 0.007177216053009033, 0.007154751777648926, 0.007186367988586426, 0.007206912040710449, 0.007501823902130127, 0.007738368034362793, 0.00719155216217041, 0.007176224231719971, 0.007155680179595948, 0.007435264110565186, 0.016763904571533202, 0.008265727996826172, 0.007964704036712646, 0.007598048210144043, 0.007525375843048096, 0.007504960060119629, 0.007465919971466064, 0.007514111995697022, 0.0074783039093017575, 0.007524320125579834, 0.007561215877532959, 0.007557119846343994, 0.007526400089263916, 0.007542784214019775, 0.007531519889831543, 0.007504896163940429, 0.007506944179534912, 0.007488512039184571, 0.007423999786376953, 0.007448575973510742, 0.007512063980102539, 0.007514111995697022, 0.007507967948913574, 0.007521279811859131, 0.007483391761779785, 0.007517183780670166, 0.007502848148345947, 0.007519231796264648, 0.007727104187011719, 0.007549952030181885, 0.007512063980102539, 0.0075008001327514645, 0.0075345921516418455, 0.007518208026885987, 0.007532544136047363, 0.007504960060119629, 0.007509952068328858, 0.007553023815155029, 0.007539711952209473, 0.007531519889831543, 0.007514111995697022, 0.00748748779296875, 0.007612415790557861, 0.007541759967803955, 0.0075345921516418455, 0.0075304961204528805, 0.007568384170532226, 0.007606272220611572, 0.00764518404006958, 0.007696383953094482, 0.007664639949798584, 0.007591936111450195, 0.00759603214263916, 0.0075939841270446775, 0.007558144092559814, 0.007568384170532226, 0.0075632638931274416, 0.007526400089263916, 0.007480319976806641, 0.007526400089263916, 0.007529471874237061, 0.007539711952209473, 0.00757862377166748, 0.017319936752319336, 0.00748748779296875, 0.0075345921516418455, 0.007494656085968018, 0.007557119846343994, 0.007577600002288819, 0.007568384170532226, 0.007544832229614258, 0.0075038719177246095, 0.007513088226318359, 0.007540736198425293, 0.007639039993286132, 0.007516160011291504, 0.007502848148345947, 0.007528448104858398, 0.0075008001327514645, 0.007590911865234375, 0.007526400089263916, 0.007490560054779053, 0.007549952030181885, 0.007520256042480469, 0.007517183780670166, 0.007523327827453613, 0.00753766393661499, 0.007518208026885987, 0.0075335679054260255, 0.0075335679054260255, 0.007505919933319092, 0.007546879768371582, 0.007524352073669433, 0.007516160011291504, 0.007504896163940429, 0.007527423858642578, 0.007489535808563232, 0.007607295989990235, 0.007498752117156982, 0.007510015964508057, 0.007469056129455566, 0.007519231796264648, 0.007511040210723877, 0.007481344223022461, 0.007505919933319092, 0.007543807983398438, 0.0074670081138610836, 0.00753766393661499, 0.007527423858642578, 0.00753868818283081, 0.007536640167236328, 0.007514111995697022, 0.007555071830749512, 0.007491583824157715, 0.007590911865234375, 0.00753766393661499, 0.007566336154937744, 0.007532544136047363, 0.007548927783966064, 0.007558144092559814, 0.0075131521224975585, 0.007548863887786865, 0.007485439777374267, 0.007505919933319092, 0.0074997758865356446, 0.007512063980102539, 0.017298431396484376, 0.007486464023590088, 0.007489535808563232, 0.00753766393661499, 0.007518208026885987, 0.0075673599243164065, 0.007532544136047363, 0.007523327827453613, 0.00758784008026123, 0.0075335679054260255, 0.007565311908721924, 0.007539711952209473, 0.0075304961204528805, 0.007513088226318359, 0.007549952030181885, 0.007507967948913574, 0.007522304058074952, 0.007515135765075683, 0.007499839782714844, 0.007501760005950928, 0.007517183780670166, 0.007476223945617676, 0.007516160011291504, 0.007482367992401123, 0.007514111995697022, 0.007522304058074952, 0.007458816051483155, 0.007577600002288819, 0.0075335679054260255, 0.0075345921516418455, 0.007517183780670166, 0.007502848148345947, 0.007590911865234375, 0.0075038719177246095, 0.00749567985534668, 0.0075980801582336424, 0.007536640167236328, 0.007508992195129394, 0.00753766393661499, 0.0075038719177246095, 0.007517248153686523, 0.007477183818817138, 0.007513088226318359, 0.0074926080703735356, 0.007522304058074952, 0.007521279811859131, 0.00753868818283081, 0.007480319976806641, 0.007583744049072265, 0.007512063980102539, 0.007516160011291504, 0.007463935852050781, 0.007510015964508057, 0.007519231796264648, 0.0074997758865356446, 0.007521279811859131, 0.007529471874237061, 0.007529471874237061, 0.007514111995697022, 0.007515135765075683, 0.007511040210723877, 0.0074967041015625, 0.007493631839752197, 0.017253376007080077, 0.007525375843048096, 0.007505919933319092, 0.007521279811859131, 0.007527455806732177, 0.007559135913848877, 0.007511104106903076, 0.00751200008392334, 0.007479296207427978, 0.00749567985534668, 0.007540736198425293, 0.007539711952209473, 0.007522304058074952, 0.007523327827453613, 0.0075304961204528805, 0.00748748779296875, 0.0074926080703735356, 0.0074926080703735356, 0.008326144218444824, 0.007541759967803955, 0.007514111995697022, 0.007552000045776367, 0.007528448104858398, 0.007497727870941162, 0.007515135765075683, 0.007517183780670166, 0.007484416007995606, 0.0075008001327514645, 0.007543807983398438, 0.0075632638931274416, 0.00760319995880127, 0.007526400089263916, 0.007491583824157715, 0.007489535808563232, 0.007531519889831543, 0.007483391761779785, 0.007510015964508057, 0.007491583824157715, 0.007531519889831543, 0.0075335679054260255, 0.007517183780670166, 0.007519231796264648, 0.007602176189422607, 0.007508992195129394, 0.00753766393661499, 0.007513088226318359, 0.007529471874237061, 0.007527423858642578, 0.007508992195129394, 0.007521279811859131, 0.007483391761779785, 0.007502848148345947, 0.007531519889831543, 0.007521279811859131, 0.007516160011291504, 0.007528448104858398, 0.007515135765075683, 0.00753868818283081, 0.0074711041450500485, 0.007527423858642578, 0.0075008001327514645, 0.007515135765075683, 0.0075038719177246095]",tokens/s,132.17574563042263,,,main,False,False,,,,,,,,,,,,,,,,,,,,,,, float32-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,databricks/dbrx-base,databricks/dbrx-base,cuda,0,42,,,True,,,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",dbrx,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-350m,,cuda,0,42,,,,,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,d35829e539df8480b726c647eeabf91e41eae047,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,,MB,1068.552192,2181.562368,0.0,1541.40672,1436.182528,s,12,0.23721007919311526,0.01976750659942627,0.0003854873300060162,0.01976827144622803,0.019917565536499024,0.020365829277038575,0.020800436363220216,"[0.020909088134765625, 0.019883552551269532, 0.019494592666625978, 0.019426591873168947, 0.019793407440185547, 0.019498559951782228, 0.019796928405761718, 0.019776607513427736, 0.019759935379028322, 0.01992134475708008, 0.019532703399658204, 0.019416767120361327]",tokens/s,12950.545821870632,kWh,2.347058828466784e-07,1.2860518705089082e-07,1.4881708333928426e-06,1.851481903290412e-06,tokens/kWh,138267622.03024644,MB,1068.552192,2181.562368,0.0,1541.40672,1491.80672,s,12,10.230781188964846,0.8525650990804036,0.017568516374491507,0.8614872131347656,0.8671965515136719,0.8696436676025391,0.8719482159423828,"[0.8725243530273438, 0.8655316772460937, 0.8224754638671875, 0.8391669921875, 0.8672867431640625, 0.8500016479492187, 0.8655090942382813, 0.8663848266601563, 0.8383928833007812, 0.8660814819335938, 0.819960693359375, 0.85746533203125]",tokens/s,73.89465047062477,kWh,1.0192779495957353e-05,5.5852522588750045e-06,2.2920205174775116e-05,3.869823692960747e-05,tokens/kWh,1627981.1432907837,,s,755,10.432233480453483,0.013817527788680123,0.0024169315888991644,0.013705151557922363,0.013815398597717285,0.014001356601715089,0.03295234191894531,"[0.014635007858276367, 0.01440665626525879, 0.014334976196289062, 0.014276608467102051, 0.014304256439208985, 0.014358528137207031, 0.01444863986968994, 0.014229503631591797, 0.01452236843109131, 0.014353407859802245, 0.013615103721618652, 0.013674495697021484, 0.01354854393005371, 0.013545472145080567, 0.013566975593566894, 0.013574144363403321, 0.01375334358215332, 0.01375641632080078, 0.013553664207458497, 0.01377894401550293, 0.013591551780700683, 0.013586432456970214, 0.013664256095886231, 0.013649920463562011, 0.01358233642578125, 0.013565952301025391, 0.013656064033508301, 0.013771776199340821, 0.013707263946533203, 0.01359772777557373, 0.013598688125610351, 0.013508607864379883, 0.013561856269836426, 0.013500415802001953, 0.013371423721313476, 0.013505536079406738, 0.014265312194824218, 0.01426636791229248, 0.014387200355529785, 0.013872127532958984, 0.013793279647827148, 0.013792256355285644, 0.013848575592041015, 0.013888511657714844, 0.013796352386474609, 0.013750271797180176, 0.013707263946533203, 0.013728768348693847, 0.013757439613342285, 0.01377280044555664, 0.013757439613342285, 0.013723648071289063, 0.013873151779174805, 0.013759488105773926, 0.013858816146850587, 0.013767680168151856, 0.013727744102478028, 0.013685759544372558, 0.01375436782836914, 0.013707263946533203, 0.013768704414367675, 0.013660160064697266, 0.034148353576660156, 0.013682687759399414, 0.013701120376586913, 0.013740032196044923, 0.013660160064697266, 0.01367244815826416, 0.013726719856262207, 0.013728768348693847, 0.013699071884155273, 0.013745152473449707, 0.013718591690063477, 0.013705151557922363, 0.013791232109069825, 0.01378713607788086, 0.013721599578857421, 0.013664287567138672, 0.013733856201171875, 0.013667327880859375, 0.013818880081176758, 0.013834239959716797, 0.013699071884155273, 0.013709312438964843, 0.013684736251831055, 0.013657088279724122, 0.013656064033508301, 0.013720576286315917, 0.013734911918640137, 0.01375641632080078, 0.013733887672424316, 0.013669376373291015, 0.013719552040100098, 0.013724672317504882, 0.013677568435668945, 0.013751296043395997, 0.013685759544372558, 0.013744128227233888, 0.01376972770690918, 0.013891584396362304, 0.013709312438964843, 0.013785087585449218, 0.01376460838317871, 0.01366528034210205, 0.013736960411071777, 0.013900799751281738, 0.013766655921936035, 0.013699071884155273, 0.013758463859558106, 0.013807616233825683, 0.013767680168151856, 0.013729791641235351, 0.013709312438964843, 0.013783040046691895, 0.013705216407775878, 0.01369600009918213, 0.013760512351989745, 0.013697024345397948, 0.013708288192749024, 0.013700096130371094, 0.013728768348693847, 0.01367142391204834, 0.013806591987609864, 0.01376153564453125, 0.013702143669128418, 0.03293491363525391, 0.013025279998779296, 0.012980223655700684, 0.013008895874023438, 0.013010944366455078, 0.013012991905212403, 0.012998656272888183, 0.013006848335266113, 0.013082624435424805, 0.012959744453430176, 0.012993535995483398, 0.013029376029968261, 0.013010944366455078, 0.012982272148132324, 0.012992544174194336, 0.012998623847961426, 0.013039615631103516, 0.013006848335266113, 0.012986399650573731, 0.01301807975769043, 0.01298742389678955, 0.013021151542663574, 0.013036543846130372, 0.013065216064453124, 0.013087743759155274, 0.01305292797088623, 0.01306009578704834, 0.013142016410827637, 0.013073408126831054, 0.01304371166229248, 0.013066240310668945, 0.013042688369750977, 0.013063167572021485, 0.01304576015472412, 0.013256704330444336, 0.013081600189208984, 0.013100031852722169, 0.013031423568725586, 0.013065216064453124, 0.013082624435424805, 0.013011967658996582, 0.012999679565429688, 0.013087743759155274, 0.01306112003326416, 0.013044735908508302, 0.013047807693481446, 0.013117440223693847, 0.013085696220397949, 0.01297715187072754, 0.013377535820007324, 0.013042688369750977, 0.01304576015472412, 0.013017087936401368, 0.01307852840423584, 0.013040639877319337, 0.013034496307373047, 0.013106176376342774, 0.01306214427947998, 0.013057024002075195, 0.013115391731262208, 0.013039615631103516, 0.013015040397644043, 0.01306214427947998, 0.03284479904174804, 0.01304576015472412, 0.012942367553710938, 0.013052895545959472, 0.013007871627807617, 0.013007871627807617, 0.013001728057861327, 0.012992511749267579, 0.012993535995483398, 0.013021183967590331, 0.013011967658996582, 0.01305395221710205, 0.013029376029968261, 0.013028351783752442, 0.013050880432128906, 0.01297920036315918, 0.012981247901916505, 0.013112319946289062, 0.012958720207214355, 0.012989439964294434, 0.01298739242553711, 0.012974080085754394, 0.013042688369750977, 0.012956671714782715, 0.013080608367919922, 0.01301091194152832, 0.013006848335266113, 0.012982272148132324, 0.013035519599914551, 0.012998656272888183, 0.013003775596618652, 0.01298534393310547, 0.012989439964294434, 0.012900351524353027, 0.012972031593322754, 0.013720576286315917, 0.01375641632080078, 0.013723648071289063, 0.013729791641235351, 0.013692928314208984, 0.013750271797180176, 0.01389568042755127, 0.013729791641235351, 0.013705216407775878, 0.01368883228302002, 0.013712384223937989, 0.013725695610046386, 0.013666303634643554, 0.013727744102478028, 0.013670399665832519, 0.013724672317504882, 0.013640704154968262, 0.013733887672424316, 0.013651968002319336, 0.01346560001373291, 0.013570048332214356, 0.013640704154968262, 0.013722623825073242, 0.013748224258422852, 0.01368166446685791, 0.013712384223937989, 0.013691904067993164, 0.01368883228302002, 0.033974273681640625, 0.013700096130371094, 0.013645824432373046, 0.013677568435668945, 0.013705247879028321, 0.01372771167755127, 0.01371548843383789, 0.013800415992736816, 0.013667327880859375, 0.013700096130371094, 0.01368064022064209, 0.01376153564453125, 0.013699071884155273, 0.013766719818115234, 0.013713343620300293, 0.01368883228302002, 0.013694975852966309, 0.013759488105773926, 0.013750271797180176, 0.013767680168151856, 0.013714431762695312, 0.013702143669128418, 0.014334976196289062, 0.013881343841552735, 0.013873151779174805, 0.01377996826171875, 0.013757439613342285, 0.013735936164855958, 0.013682687759399414, 0.013734911918640137, 0.013742079734802246, 0.013775872230529786, 0.013722623825073242, 0.013827072143554688, 0.01377894401550293, 0.013720576286315917, 0.013685759544372558, 0.01375436782836914, 0.013750271797180176, 0.013732864379882812, 0.013744128227233888, 0.013740032196044923, 0.013719552040100098, 0.013709312438964843, 0.013685759544372558, 0.013748224258422852, 0.01369600009918213, 0.013751296043395997, 0.013785087585449218, 0.013837311744689941, 0.013718527793884277, 0.013749247550964355, 0.013747200012207032, 0.013725695610046386, 0.013751296043395997, 0.013774847984313965, 0.013774847984313965, 0.013718527793884277, 0.013762592315673828, 0.013716447830200195, 0.013695039749145509, 0.014172096252441407, 0.01406156826019287, 0.03308755111694336, 0.013034432411193848, 0.013143039703369141, 0.013138943672180176, 0.013017087936401368, 0.013025279998779296, 0.013018112182617187, 0.01306726360321045, 0.013042688369750977, 0.013028351783752442, 0.013033472061157226, 0.013003775596618652, 0.013017087936401368, 0.013087743759155274, 0.013021183967590331, 0.013061152458190919, 0.012977120399475097, 0.013025279998779296, 0.013015040397644043, 0.013034496307373047, 0.013014016151428222, 0.01297920036315918, 0.013096960067749023, 0.01386905574798584, 0.013690879821777344, 0.01374617576599121, 0.013729791641235351, 0.013751296043395997, 0.013724672317504882, 0.013736960411071777, 0.013752320289611816, 0.013692928314208984, 0.013827136039733887, 0.013697983741760254, 0.013730815887451172, 0.013747200012207032, 0.013799424171447755, 0.01374617576599121, 0.013770751953125, 0.014030847549438476, 0.013752320289611816, 0.01377791976928711, 0.013745152473449707, 0.013722623825073242, 0.013711359977722168, 0.013731840133666993, 0.013719552040100098, 0.013718527793884277, 0.013716511726379395, 0.013730784416198731, 0.013775872230529786, 0.013693951606750488, 0.013708288192749024, 0.013716480255126954, 0.01368166446685791, 0.013715456008911133, 0.013714431762695312, 0.013724672317504882, 0.013707263946533203, 0.013668352127075196, 0.013684736251831055, 0.013690879821777344, 0.013749247550964355, 0.03395891189575195, 0.013742143630981446, 0.013660096168518066, 0.013732864379882812, 0.013731840133666993, 0.013766655921936035, 0.013757439613342285, 0.013716480255126954, 0.013749247550964355, 0.013702143669128418, 0.013732864379882812, 0.01375436782836914, 0.01375641632080078, 0.013806591987609864, 0.013751296043395997, 0.01377996826171875, 0.013725695610046386, 0.013875200271606445, 0.013703167915344238, 0.013751296043395997, 0.013714431762695312, 0.013723648071289063, 0.013730815887451172, 0.013698047637939453, 0.013743103981018067, 0.013723648071289063, 0.013711423873901367, 0.013723584175109863, 0.013705280303955078, 0.013736895561218262, 0.013690879821777344, 0.013783040046691895, 0.013686783790588379, 0.013751296043395997, 0.013678591728210449, 0.013714495658874512, 0.013791168212890625, 0.013749247550964355, 0.01375334358215332, 0.013723648071289063, 0.013784064292907714, 0.013722623825073242, 0.013711359977722168, 0.013709312438964843, 0.013710335731506347, 0.013699071884155273, 0.013727744102478028, 0.013763584136962891, 0.01370419216156006, 0.01376460838317871, 0.013699071884155273, 0.013699071884155273, 0.013732864379882812, 0.013700096130371094, 0.013791232109069825, 0.01378713607788086, 0.013725695610046386, 0.013697024345397948, 0.013719552040100098, 0.013727744102478028, 0.013657088279724122, 0.013741056442260742, 0.013676575660705566, 0.033946590423583986, 0.013678591728210449, 0.01374617576599121, 0.013719552040100098, 0.01367142391204834, 0.013684736251831055, 0.013712384223937989, 0.013721664428710938, 0.013729727745056153, 0.013760512351989745, 0.013693951606750488, 0.013631487846374512, 0.013717503547668456, 0.013659135818481445, 0.013669376373291015, 0.013639743804931641, 0.013729727745056153, 0.013670399665832519, 0.013709312438964843, 0.013691904067993164, 0.01366220760345459, 0.013658111572265624, 0.013716480255126954, 0.013615103721618652, 0.013683712005615235, 0.013676544189453126, 0.014004223823547364, 0.013785087585449218, 0.013702143669128418, 0.01365401554107666, 0.013701120376586913, 0.01377791976928711, 0.01368166446685791, 0.013716544151306152, 0.013929408073425293, 0.014581760406494141, 0.014315520286560059, 0.014000127792358399, 0.013724672317504882, 0.013774847984313965, 0.013718527793884277, 0.013715456008911133, 0.013703167915344238, 0.013742079734802246, 0.013643775939941406, 0.013706239700317382, 0.013881407737731933, 0.013676480293273926, 0.013782015800476074, 0.013742079734802246, 0.013750271797180176, 0.013715456008911133, 0.013728768348693847, 0.013715456008911133, 0.01369600009918213, 0.013758463859558106, 0.013656064033508301, 0.013725695610046386, 0.013664256095886231, 0.013730815887451172, 0.013732864379882812, 0.013697024345397948, 0.013759488105773926, 0.03320115280151367, 0.012998656272888183, 0.013011967658996582, 0.01293824005126953, 0.012943360328674316, 0.012912639617919922, 0.013179903984069824, 0.012993535995483398, 0.012955712318420411, 0.0129617280960083, 0.012967935562133789, 0.012989472389221192, 0.013068256378173828, 0.012992511749267579, 0.013011967658996582, 0.013005824089050292, 0.013012991905212403, 0.013054976463317871, 0.01303756809234619, 0.013058048248291015, 0.012991488456726074, 0.012976127624511719, 0.013041664123535156, 0.012995583534240723, 0.013021183967590331, 0.013035519599914551, 0.01304371166229248, 0.013065216064453124, 0.013015040397644043, 0.013023232460021973, 0.01308672046661377, 0.013035519599914551, 0.01304371166229248, 0.013048831939697265, 0.013079551696777344, 0.01303756809234619, 0.013009920120239257, 0.013015040397644043, 0.013027327537536621, 0.013021183967590331, 0.013009984016418457, 0.01406764793395996, 0.013816831588745117, 0.013784064292907714, 0.013850624084472657, 0.013782015800476074, 0.01379532814025879, 0.013747200012207032, 0.013810688018798829, 0.013725695610046386, 0.013791232109069825, 0.01376153564453125, 0.013752320289611816, 0.013743103981018067, 0.013791232109069825, 0.01439948844909668, 0.013914112091064454, 0.013837311744689941, 0.013805567741394043, 0.013790207862854004, 0.013765631675720215, 0.01376255989074707, 0.01377280044555664, 0.03412070465087891, 0.013693951606750488, 0.013683712005615235, 0.013775872230529786, 0.013718527793884277, 0.013729791641235351, 0.013731840133666993, 0.013737983703613281, 0.013698047637939453, 0.01379532814025879, 0.013718527793884277, 0.013735936164855958, 0.013722623825073242, 0.013710335731506347, 0.01375641632080078, 0.013815808296203613, 0.013752320289611816, 0.013748224258422852, 0.013773823738098144, 0.013770751953125, 0.013790207862854004, 0.013798399925231934, 0.013780991554260253, 0.013768704414367675, 0.013742079734802246, 0.013745152473449707, 0.014112768173217773, 0.013828096389770507, 0.013755392074584961, 0.013700096130371094, 0.01368992042541504, 0.013662143707275391, 0.013768704414367675, 0.013829119682312012, 0.01387007999420166, 0.013660160064697266, 0.013766655921936035, 0.013705216407775878, 0.013743103981018067, 0.013728832244873047, 0.013722559928894044, 0.013674495697021484, 0.013656064033508301, 0.013677568435668945, 0.013744159698486328, 0.013710304260253907, 0.01374617576599121, 0.013683712005615235, 0.01384447956085205, 0.013686783790588379, 0.013698047637939453, 0.013706239700317382, 0.013814784049987794, 0.013670399665832519, 0.013706239700317382, 0.013706239700317382, 0.013667327880859375, 0.013724672317504882, 0.01374617576599121, 0.013718527793884277, 0.013706239700317382, 0.013679679870605468, 0.013714367866516114, 0.03297280120849609, 0.013124608039855956, 0.013116415977478027, 0.013063167572021485, 0.013018112182617187, 0.013026304244995117, 0.012999679565429688, 0.013031423568725586, 0.012974080085754394, 0.012957695960998536, 0.013007871627807617, 0.013000703811645508, 0.012943360328674316, 0.01297923183441162, 0.012977120399475097, 0.01298534393310547, 0.012950528144836425, 0.012957695960998536, 0.012982272148132324, 0.012985440254211425, 0.012975008010864257, 0.012981247901916505, 0.013088768005371093, 0.013016063690185547, 0.012993535995483398, 0.012996607780456543, 0.012969984054565429, 0.012986368179321289, 0.012988415718078614, 0.01298739242553711, 0.01296895980834961, 0.012988415718078614, 0.012986368179321289, 0.013003775596618652, 0.012973055839538575, 0.012990464210510254, 0.012966912269592285, 0.012981247901916505, 0.01298739242553711, 0.013022208213806152, 0.013017087936401368, 0.013010944366455078, 0.013001728057861327, 0.013088768005371093, 0.013031423568725586, 0.013032447814941407, 0.01304576015472412, 0.01301814365386963, 0.012978143692016601, 0.013002752304077148, 0.013039615631103516, 0.013050880432128906, 0.013010944366455078, 0.013005824089050292, 0.012992511749267579, 0.013066240310668945, 0.01297920036315918, 0.0130764799118042, 0.013016063690185547, 0.013017087936401368, 0.013010944366455078, 0.013033472061157226, 0.013100031852722169, 0.0328908805847168, 0.012998656272888183, 0.013111295700073243, 0.013095935821533204, 0.013014016151428222, 0.01298739242553711, 0.013003775596618652, 0.01305190372467041, 0.013030400276184082, 0.01298739242553711, 0.01306726360321045, 0.013089792251586914, 0.013025279998779296, 0.01304371166229248, 0.01306931209564209, 0.013072383880615235, 0.012989439964294434, 0.01304576015472412, 0.013801535606384278, 0.013717439651489258, 0.013890560150146485, 0.013760512351989745, 0.013819904327392578, 0.013771776199340821, 0.013773823738098144, 0.014641152381896973, 0.014643199920654297, 0.014138367652893067, 0.013828096389770507, 0.013699071884155273, 0.013752320289611816, 0.013768704414367675, 0.013755392074584961, 0.013745152473449707, 0.013743103981018067, 0.013792256355285644, 0.013694975852966309, 0.013802495956420899, 0.013775936126708984, 0.013754303932189942, 0.013785087585449218, 0.013703200340270996, 0.013962207794189453, 0.013705216407775878, 0.013813759803771973, 0.013789183616638183, 0.013820927619934082, 0.013705216407775878, 0.013757439613342285, 0.013691904067993164, 0.013853695869445801, 0.013729791641235351, 0.01374617576599121, 0.013719552040100098, 0.013748224258422852, 0.013752320289611816, 0.013809663772583008, 0.013785087585449218, 0.01406668758392334, 0.01374617576599121, 0.01389363193511963, 0.013843456268310546, 0.013836288452148437]",tokens/s,72.37184649045834,,,main,False,False,,,,,,,,,,,,,,,,,,,,,,, float32-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,-,-,cuda,0,42,,,True,True,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 106, in _inner_fn validate_repo_id(arg_value) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 160, in validate_repo_id raise HFValidationError( huggingface_hub.errors.HFValidationError: Repo id must use alphanumeric chars or '-', '_', '.', '--' and '..' are forbidden, '-' and '.' cannot start or end the name, max length is 96: '-'. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 466, in cached_file raise EnvironmentError( OSError: Incorrect path_or_model_id: '-'. Please provide either the path to a local folder or the repo_id of a model on the Hub. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,stabilityai/stablelm-2-12b,stabilityai/stablelm-2-12b,cuda,0,42,,,True,True,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ self.load_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 250, in load_model_with_no_weights self.pretrained_model = self.automodel_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3710, in from_pretrained model = cls(config, *model_args, **model_kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/stablelm/modeling_stablelm.py"", line 1188, in __init__ self.model = StableLmModel(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/stablelm/modeling_stablelm.py"", line 966, in __init__ [StableLmDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/stablelm/modeling_stablelm.py"", line 966, in [StableLmDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/stablelm/modeling_stablelm.py"", line 748, in __init__ self.mlp = StableLmMLP(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/stablelm/modeling_stablelm.py"", line 204, in __init__ self.gate_proj = nn.Linear(self.hidden_size, self.intermediate_size, bias=False) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/linear.py"", line 98, in __init__ self.weight = Parameter(torch.empty((out_features, in_features), **factory_kwargs)) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_device.py"", line 78, in __torch_function__ return func(*args, **kwargs) torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 270.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,v,v,cuda,0,42,,,True,True,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/v/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66932555-6be0430b45c14f001d5f97ae;4e6e15c4-769a-4f0e-b1ed-0c51aca67099) Repository Not Found for url: https://huggingface.co/v/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: v is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,openai-community/gpt2,openai-community/gpt2,cuda,0,42,,,True,True,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1099.210752,1418.19904,0.0,778.043392,601.166848,s,19,0.18128963279724122,0.009541559620907433,0.00031779269862037553,0.009458847999572754,0.009708159828186035,0.010222595596313475,0.010578356246948242,"[0.010667296409606933, 0.010173184394836426, 0.009462528228759766, 0.009500479698181153, 0.009489824295043945, 0.009458847999572754, 0.009479328155517578, 0.00946134376525879, 0.009448384284973145, 0.009422528266906737, 0.009321439743041992, 0.009472767829895019, 0.00940332794189453, 0.009591903686523438, 0.009431520462036132, 0.009341440200805663, 0.009337087631225587, 0.00940169620513916, 0.009424703598022461]",tokens/s,26829.99532267803,kWh,1.1473024290099243e-07,6.284304177807445e-08,5.747377844931974e-07,7.523110691722643e-07,tokens/kWh,340284771.1408339,MB,1099.538432,1418.19904,0.0,778.043392,620.356608,s,19,9.940461730957031,0.5231821963661595,0.006936860479069573,0.5211837768554688,0.5285371215820313,0.5336169982910156,0.5456140246582031,"[0.54861328125, 0.5200792236328124, 0.5259476318359375, 0.5217031860351562, 0.5217098388671875, 0.5319507446289062, 0.5205490112304687, 0.5212012939453124, 0.52119287109375, 0.521158447265625, 0.5182402954101563, 0.5211837768554688, 0.5231117553710938, 0.5208383178710938, 0.5200266723632813, 0.515190673828125, 0.5189518432617187, 0.521129150390625, 0.5276837158203125]",tokens/s,120.41694162679075,kWh,6.329118204940518e-06,3.4681341634669864e-06,1.1839526532875298e-05,2.1636778901282805e-05,tokens/kWh,2911708.8216982633,,s,1196,10.10165556907654,0.008446200308592423,0.0012225374068946526,0.008260607719421387,0.008543744087219238,0.008691968202590942,0.018165042972564698,"[0.009374719619750976, 0.009341952323913574, 0.00898252773284912, 0.008909824371337891, 0.008951807975769043, 0.008835071563720704, 0.008820735931396484, 0.008653887748718262, 0.00871724796295166, 0.008668160438537598, 0.008626175880432128, 0.008643584251403809, 0.008753151893615722, 0.008596511840820312, 0.00855446434020996, 0.008706175804138183, 0.008697728157043456, 0.008782848358154297, 0.008606719970703124, 0.008599552154541015, 0.008691712379455567, 0.008648703575134278, 0.008502304077148437, 0.008571871757507324, 0.008642560005187988, 0.00858726406097412, 0.008644607543945313, 0.008657919883728027, 0.008777728080749512, 0.008855551719665527, 0.008734720230102539, 0.008647680282592773, 0.008728575706481934, 0.008791040420532227, 0.008677375793457032, 0.00858521556854248, 0.008523776054382324, 0.008739935874938964, 0.008729503631591798, 0.008763392448425293, 0.008556544303894043, 0.0087326717376709, 0.008560640335083008, 0.008565759658813477, 0.00858521556854248, 0.008911871910095214, 0.00882380771636963, 0.008761343955993652, 0.008822848320007324, 0.008705056190490723, 0.008563615798950196, 0.008794112205505371, 0.00852684783935547, 0.008408063888549805, 0.008756223678588868, 0.008516608238220215, 0.00838144016265869, 0.008282112121582032, 0.008230912208557128, 0.008299584388732911, 0.008713184356689453, 0.008756192207336425, 0.019679231643676756, 0.008623104095458984, 0.008548352241516113, 0.008539135932922364, 0.008547327995300292, 0.008492032051086425, 0.008426624298095703, 0.008375167846679688, 0.00839577579498291, 0.008140800476074218, 0.00809062385559082, 0.00813980770111084, 0.008227807998657227, 0.008119296073913575, 0.00810700798034668, 0.008113151550292968, 0.0081397762298584, 0.008121343612670898, 0.008101920127868652, 0.008088543891906738, 0.008086527824401855, 0.00819814395904541, 0.00820531177520752, 0.008143872261047362, 0.008115391731262207, 0.008131391525268554, 0.008151040077209473, 0.008119296073913575, 0.00808448028564453, 0.008117376327514648, 0.008102848052978515, 0.008093631744384766, 0.0081080322265625, 0.008170495986938477, 0.008227840423583984, 0.008317952156066894, 0.008291328430175781, 0.008214624404907226, 0.008226719856262207, 0.008170495986938477, 0.008251551628112794, 0.008176480293273926, 0.008205344200134277, 0.008396767616271972, 0.008272895812988281, 0.008242207527160645, 0.008279007911682129, 0.00828825569152832, 0.008260607719421387, 0.008252479553222657, 0.008307647705078126, 0.008246272087097169, 0.00828825569152832, 0.008223744392395019, 0.008246272087097169, 0.008268799781799317, 0.008292351722717285, 0.00822169589996338, 0.008245247840881348, 0.008274944305419921, 0.008256511688232422, 0.008257535934448243, 0.00832102394104004, 0.01823027229309082, 0.008279040336608886, 0.008247296333312988, 0.008256511688232422, 0.00828006362915039, 0.008269824028015137, 0.008269824028015137, 0.00892416000366211, 0.008607744216918945, 0.009148447990417481, 0.008775648117065429, 0.008606719970703124, 0.008601632118225098, 0.008605664253234863, 0.008607744216918945, 0.00860262393951416, 0.008560640335083008, 0.008464384078979491, 0.008464384078979491, 0.00827187156677246, 0.008309760093688966, 0.008245311737060546, 0.008244159698486329, 0.008264703750610352, 0.008203264236450195, 0.008270848274230956, 0.008250368118286134, 0.008244223594665527, 0.008243200302124023, 0.008232959747314453, 0.008248319625854492, 0.008265727996826172, 0.008259584426879883, 0.008257535934448243, 0.008257535934448243, 0.008247296333312988, 0.008263680458068847, 0.008260607719421387, 0.00829030418395996, 0.008268799781799317, 0.008265727996826172, 0.008272895812988281, 0.008247296333312988, 0.008276991844177246, 0.008282112121582032, 0.008230912208557128, 0.008289279937744141, 0.008313856124877929, 0.008282112121582032, 0.008260607719421387, 0.00830463981628418, 0.008284159660339355, 0.008274944305419921, 0.008253439903259278, 0.008228863716125488, 0.008272895812988281, 0.008279040336608886, 0.008253439903259278, 0.008260607719421387, 0.008266752243041992, 0.008334336280822753, 0.008247296333312988, 0.008259584426879883, 0.0182794246673584, 0.00820633602142334, 0.008225791931152344, 0.008299519538879394, 0.008212639808654785, 0.008256352424621582, 0.008260607719421387, 0.008264703750610352, 0.008232959747314453, 0.008239168167114258, 0.008250304222106934, 0.008249343872070313, 0.008294400215148925, 0.008276991844177246, 0.008249343872070313, 0.008491007804870606, 0.008254464149475099, 0.00830463981628418, 0.00821350383758545, 0.008239104270935058, 0.008261631965637208, 0.008259584426879883, 0.008202239990234375, 0.008267775535583496, 0.008233983993530274, 0.008242176055908204, 0.0082227201461792, 0.008220671653747558, 0.008255488395690918, 0.008260607719421387, 0.008225791931152344, 0.00824118423461914, 0.008240096092224121, 0.008202239990234375, 0.008259584426879883, 0.008227840423583984, 0.008187040328979493, 0.008275808334350586, 0.008254464149475099, 0.008259584426879883, 0.008279040336608886, 0.008245247840881348, 0.008272895812988281, 0.008253439903259278, 0.008224767684936523, 0.008265727996826172, 0.008223744392395019, 0.008734784126281737, 0.008686528205871583, 0.008285280227661132, 0.00826255989074707, 0.008541184425354004, 0.008240127563476562, 0.008256511688232422, 0.008248384475708009, 0.008268735885620116, 0.008258560180664062, 0.008247296333312988, 0.008283136367797851, 0.008293375968933106, 0.008229887962341309, 0.008283136367797851, 0.008249343872070313, 0.01817087936401367, 0.008275967597961426, 0.008310784339904785, 0.008278016090393067, 0.008229887962341309, 0.008243200302124023, 0.008264703750610352, 0.008226816177368163, 0.00819814395904541, 0.008283136367797851, 0.008235008239746093, 0.008256511688232422, 0.008253472328186035, 0.008289248466491698, 0.008261631965637208, 0.008209407806396484, 0.008203264236450195, 0.00830463981628418, 0.008283136367797851, 0.00830463981628418, 0.008259584426879883, 0.008258560180664062, 0.008285183906555176, 0.008246272087097169, 0.008251392364501953, 0.008258560180664062, 0.008282112121582032, 0.008279040336608886, 0.008324095726013184, 0.008264703750610352, 0.008340479850769043, 0.008230912208557128, 0.008253439903259278, 0.008284159660339355, 0.008294400215148925, 0.008235008239746093, 0.008648703575134278, 0.008472576141357421, 0.008267775535583496, 0.008245247840881348, 0.00824732780456543, 0.008243167877197265, 0.008246272087097169, 0.008273920059204102, 0.008250368118286134, 0.008254464149475099, 0.008242176055908204, 0.008235008239746093, 0.008251392364501953, 0.008281087875366211, 0.008275967597961426, 0.008253439903259278, 0.008253439903259278, 0.008263680458068847, 0.008281087875366211, 0.008279040336608886, 0.008341504096984862, 0.008265727996826172, 0.008293375968933106, 0.008285183906555176, 0.00828006362915039, 0.008281120300292969, 0.00827184009552002, 0.01822822380065918, 0.008253439903259278, 0.008268799781799317, 0.008255488395690918, 0.008249343872070313, 0.008263680458068847, 0.008273920059204102, 0.008833024024963379, 0.008607744216918945, 0.008302623748779298, 0.009087967872619628, 0.008692735671997071, 0.008630271911621093, 0.008647680282592773, 0.008583168029785156, 0.00857907199859619, 0.008657919883728027, 0.00859545612335205, 0.008541184425354004, 0.008582143783569337, 0.008301568031311036, 0.008240127563476562, 0.008259584426879883, 0.008261631965637208, 0.008287232398986816, 0.008243200302124023, 0.008281120300292969, 0.00828003215789795, 0.008276991844177246, 0.008253439903259278, 0.008279040336608886, 0.008301568031311036, 0.008278016090393067, 0.00841113567352295, 0.008590335845947266, 0.008523776054382324, 0.008557567596435547, 0.008532992362976074, 0.00856169605255127, 0.008530912399291992, 0.008527872085571288, 0.008546303749084473, 0.008477696418762207, 0.008524800300598144, 0.008511487960815429, 0.008522751808166504, 0.00858726406097412, 0.008498175621032715, 0.008551424026489257, 0.008538111686706543, 0.008540160179138183, 0.008570879936218261, 0.008535103797912598, 0.008517567634582519, 0.00852070426940918, 0.008295424461364746, 0.008268799781799317, 0.008265727996826172, 0.008264703750610352, 0.008258560180664062, 0.008170495986938477, 0.008286208152770995, 0.008284159660339355, 0.01825382423400879, 0.00827187156677246, 0.008236031532287597, 0.008315903663635254, 0.008229887962341309, 0.008308735847473145, 0.008258560180664062, 0.00818380832672119, 0.008256511688232422, 0.008237055778503418, 0.008228863716125488, 0.008302592277526855, 0.008220671653747558, 0.008227840423583984, 0.008262656211853027, 0.00822169589996338, 0.008216575622558593, 0.008283136367797851, 0.008227840423583984, 0.008270848274230956, 0.008218624114990235, 0.0082227201461792, 0.00830668830871582, 0.00829030418395996, 0.008225791931152344, 0.008224767684936523, 0.008252415657043457, 0.008263680458068847, 0.008240127563476562, 0.008244223594665527, 0.008218624114990235, 0.008226816177368163, 0.008229887962341309, 0.008391679763793946, 0.008232959747314453, 0.008275967597961426, 0.008237055778503418, 0.008248319625854492, 0.008253439903259278, 0.008256511688232422, 0.008274944305419921, 0.008299551963806152, 0.008275936126708985, 0.008269824028015137, 0.00827187156677246, 0.008254464149475099, 0.008332287788391114, 0.008243200302124023, 0.008244223594665527, 0.008263680458068847, 0.00829849624633789, 0.008225791931152344, 0.008239104270935058, 0.008235008239746093, 0.008260640144348144, 0.00825443172454834, 0.008270848274230956, 0.008255488395690918, 0.008254464149475099, 0.008239104270935058, 0.008226816177368163, 0.008227840423583984, 0.00826576042175293, 0.018193376541137694, 0.00828825569152832, 0.008286208152770995, 0.008251392364501953, 0.008233983993530274, 0.008267775535583496, 0.008228863716125488, 0.008241151809692383, 0.008246272087097169, 0.008246272087097169, 0.00822169589996338, 0.008268832206726074, 0.008218591690063476, 0.00830668830871582, 0.008258560180664062, 0.008254464149475099, 0.00830668830871582, 0.008252415657043457, 0.008232959747314453, 0.008278016090393067, 0.008299519538879394, 0.00831283187866211, 0.008294400215148925, 0.008301568031311036, 0.008254464149475099, 0.008244223594665527, 0.008294400215148925, 0.008240127563476562, 0.008287232398986816, 0.008254464149475099, 0.008260607719421387, 0.008243200302124023, 0.008245247840881348, 0.008272895812988281, 0.008315903663635254, 0.008256511688232422, 0.008284159660339355, 0.008240127563476562, 0.008261631965637208, 0.008267775535583496, 0.008342528343200683, 0.008256511688232422, 0.008248319625854492, 0.008238080024719239, 0.008262656211853027, 0.008236031532287597, 0.008259584426879883, 0.008233983993530274, 0.008241151809692383, 0.008266752243041992, 0.008250368118286134, 0.008266752243041992, 0.00831283187866211, 0.008244223594665527, 0.00828825569152832, 0.008256511688232422, 0.00831283187866211, 0.008262656211853027, 0.008262656211853027, 0.008285183906555176, 0.008299519538879394, 0.008274944305419921, 0.008281087875366211, 0.01820569610595703, 0.008229887962341309, 0.00823094367980957, 0.008260576248168946, 0.008353792190551757, 0.008250368118286134, 0.008275967597961426, 0.008233983993530274, 0.008249343872070313, 0.008237055778503418, 0.008249343872070313, 0.008247296333312988, 0.008253439903259278, 0.008258560180664062, 0.008302592277526855, 0.008341504096984862, 0.008279040336608886, 0.008278016090393067, 0.008294400215148925, 0.008275967597961426, 0.008252415657043457, 0.008259584426879883, 0.008219648361206054, 0.008256511688232422, 0.008255488395690918, 0.008236031532287597, 0.008247296333312988, 0.008257535934448243, 0.008263680458068847, 0.008250368118286134, 0.008255488395690918, 0.008258560180664062, 0.008245280265808105, 0.008268832206726074, 0.00829638385772705, 0.008232959747314453, 0.008276991844177246, 0.008228863716125488, 0.008301568031311036, 0.008267775535583496, 0.008273920059204102, 0.008217599868774414, 0.008269824028015137, 0.008226816177368163, 0.008269824028015137, 0.008263680458068847, 0.008262751579284668, 0.008260512351989746, 0.008248319625854492, 0.008247296333312988, 0.008258560180664062, 0.008254464149475099, 0.008276000022888183, 0.008285152435302734, 0.008275967597961426, 0.008274944305419921, 0.008281087875366211, 0.008282112121582032, 0.008400896072387695, 0.008282112121582032, 0.008293375968933106, 0.00827187156677246, 0.008256511688232422, 0.018164735794067383, 0.008260607719421387, 0.008272895812988281, 0.008287232398986816, 0.008270848274230956, 0.008302592277526855, 0.008247360229492188, 0.008172479629516602, 0.008267775535583496, 0.008261631965637208, 0.008254464149475099, 0.008275967597961426, 0.008264703750610352, 0.008273920059204102, 0.008242176055908204, 0.008243200302124023, 0.008266752243041992, 0.008264703750610352, 0.00829036808013916, 0.008282048225402832, 0.00830668830871582, 0.008285183906555176, 0.008259584426879883, 0.008289279937744141, 0.008336383819580078, 0.008287232398986816, 0.00829644775390625, 0.008286208152770995, 0.008261631965637208, 0.008265727996826172, 0.008287232398986816, 0.008233983993530274, 0.00827187156677246, 0.008266752243041992, 0.008247296333312988, 0.008273920059204102, 0.008267775535583496, 0.008241151809692383, 0.008266752243041992, 0.008245247840881348, 0.008253439903259278, 0.008235008239746093, 0.008260607719421387, 0.008245247840881348, 0.008230912208557128, 0.008262656211853027, 0.008231936454772949, 0.008236031532287597, 0.008246272087097169, 0.008274944305419921, 0.008273920059204102, 0.008291328430175781, 0.008248319625854492, 0.008237055778503418, 0.008273920059204102, 0.008253439903259278, 0.008262656211853027, 0.008274944305419921, 0.008230912208557128, 0.00818380832672119, 0.008248319625854492, 0.008333312034606934, 0.008364031791687012, 0.018010112762451173, 0.008252415657043457, 0.008248319625854492, 0.008126463890075684, 0.008143872261047362, 0.008130720138549805, 0.008145759582519532, 0.008138751983642578, 0.008130559921264649, 0.008114175796508789, 0.008159232139587403, 0.008241151809692383, 0.008241151809692383, 0.008268799781799317, 0.008259584426879883, 0.008299519538879394, 0.008261631965637208, 0.008254464149475099, 0.008275967597961426, 0.008261631965637208, 0.008248319625854492, 0.008251392364501953, 0.00828006362915039, 0.008257535934448243, 0.008263680458068847, 0.008229887962341309, 0.008257535934448243, 0.008238080024719239, 0.008243200302124023, 0.008230912208557128, 0.008254464149475099, 0.008236031532287597, 0.008243200302124023, 0.008351743698120117, 0.008211520195007324, 0.00822265625, 0.008246272087097169, 0.008143872261047362, 0.008269824028015137, 0.008196224212646484, 0.008152095794677735, 0.008130399703979493, 0.008174592018127442, 0.008231936454772949, 0.00811520004272461, 0.008135680198669434, 0.008121343612670898, 0.008111104011535645, 0.008113151550292968, 0.008143903732299805, 0.008210399627685546, 0.008335359573364258, 0.008289279937744141, 0.008229887962341309, 0.008262656211853027, 0.008245247840881348, 0.008256511688232422, 0.008264703750610352, 0.008208383560180664, 0.008260607719421387, 0.008257535934448243, 0.008262656211853027, 0.008251392364501953, 0.01820876884460449, 0.008242176055908204, 0.008248319625854492, 0.008261664390563965, 0.00826364803314209, 0.00828006362915039, 0.008257535934448243, 0.008263680458068847, 0.008263680458068847, 0.008246272087097169, 0.008267775535583496, 0.008261631965637208, 0.008264703750610352, 0.008250368118286134, 0.008220671653747558, 0.008250368118286134, 0.008259584426879883, 0.008293375968933106, 0.00827187156677246, 0.00824732780456543, 0.008262623786926269, 0.008279040336608886, 0.008256511688232422, 0.008223744392395019, 0.008279040336608886, 0.008262656211853027, 0.008246272087097169, 0.008211456298828124, 0.008282112121582032, 0.008328191757202149, 0.008286208152770995, 0.008257535934448243, 0.008263680458068847, 0.008257535934448243, 0.008260607719421387, 0.008247296333312988, 0.008276991844177246, 0.008243200302124023, 0.008254464149475099, 0.008244223594665527, 0.008267775535583496, 0.008250368118286134, 0.00827187156677246, 0.00827187156677246, 0.008227871894836427, 0.008282079696655273, 0.008252415657043457, 0.008317952156066894, 0.00827187156677246, 0.008285183906555176, 0.008283167839050294, 0.008286175727844238, 0.008297472000122071, 0.008247296333312988, 0.008263680458068847, 0.008294400215148925, 0.008332287788391114, 0.008286208152770995, 0.008245247840881348, 0.008226816177368163, 0.008258560180664062, 0.00830361557006836, 0.008311807632446289, 0.018150400161743165, 0.008276991844177246, 0.008250368118286134, 0.008244223594665527, 0.008231936454772949, 0.008278016090393067, 0.00830463981628418, 0.008255488395690918, 0.008275967597961426, 0.008246272087097169, 0.008251392364501953, 0.008256511688232422, 0.008247296333312988, 0.008232959747314453, 0.008241151809692383, 0.008236031532287597, 0.008262656211853027, 0.008223744392395019, 0.008266752243041992, 0.008220671653747558, 0.008392704010009766, 0.008273920059204102, 0.009038847923278808, 0.008653823852539062, 0.008942591667175292, 0.008707072257995606, 0.008226816177368163, 0.008275967597961426, 0.008268799781799317, 0.008255488395690918, 0.008282112121582032, 0.008281087875366211, 0.008262656211853027, 0.00825654411315918, 0.008259552001953125, 0.008263680458068847, 0.008278016090393067, 0.008258560180664062, 0.008203264236450195, 0.008269856452941895, 0.008262623786926269, 0.008243200302124023, 0.008261631965637208, 0.008275967597961426, 0.008255488395690918, 0.00828825569152832, 0.008344575881958008, 0.008249343872070313, 0.008250368118286134, 0.008258560180664062, 0.008242176055908204, 0.008231936454772949, 0.008238080024719239, 0.008239104270935058, 0.008244223594665527, 0.008287232398986816, 0.008236031532287597, 0.00828006362915039, 0.008251392364501953, 0.008250368118286134, 0.008258560180664062, 0.008235008239746093, 0.008243200302124023, 0.018347007751464844, 0.008209407806396484, 0.008228863716125488, 0.008250399589538574, 0.008270815849304199, 0.008247360229492188, 0.008241087913513183, 0.008245247840881348, 0.008245247840881348, 0.008250368118286134, 0.008251392364501953, 0.008256511688232422, 0.008257535934448243, 0.008261631965637208, 0.008254464149475099, 0.00831283187866211, 0.008270848274230956, 0.008245247840881348, 0.008252415657043457, 0.008345600128173827, 0.008294400215148925, 0.008231936454772949, 0.008241151809692383, 0.008269824028015137, 0.008252415657043457, 0.008218624114990235, 0.008262656211853027, 0.008224767684936523, 0.008268799781799317, 0.008240127563476562, 0.008239104270935058, 0.008248319625854492, 0.008243200302124023, 0.008249343872070313, 0.008241151809692383, 0.008238080024719239, 0.008258560180664062, 0.008310784339904785, 0.00827187156677246, 0.008261631965637208, 0.008246272087097169, 0.008256511688232422, 0.008317952156066894, 0.008241151809692383, 0.008267775535583496, 0.008247296333312988, 0.008281087875366211, 0.008220671653747558, 0.008229887962341309, 0.008263680458068847, 0.008291328430175781, 0.008250368118286134, 0.008269824028015137, 0.008261631965637208, 0.008251392364501953, 0.008286208152770995, 0.00830463981628418, 0.008269856452941895, 0.008304608345031739, 0.008263680458068847, 0.008243200302124023, 0.0083374080657959, 0.008251392364501953, 0.018182144165039063, 0.008268799781799317, 0.008261631965637208, 0.008259584426879883, 0.008261631965637208, 0.008248319625854492, 0.008278016090393067, 0.008243200302124023, 0.008264703750610352, 0.008235008239746093, 0.008235008239746093, 0.008265727996826172, 0.008253439903259278, 0.008250368118286134, 0.008236031532287597, 0.008249343872070313, 0.008252415657043457, 0.008251392364501953, 0.008241151809692383, 0.008301568031311036, 0.00827187156677246, 0.00822169589996338, 0.0083056640625, 0.008263680458068847, 0.008261631965637208, 0.008260640144348144, 0.008270815849304199, 0.008263680458068847, 0.00827187156677246, 0.008251456260681153, 0.008375231742858886, 0.008242176055908204, 0.008281087875366211, 0.00827187156677246, 0.008270848274230956, 0.008238080024719239, 0.008334336280822753, 0.008263680458068847, 0.008259584426879883, 0.008246272087097169, 0.008259584426879883, 0.008241151809692383, 0.008274944305419921, 0.008270848274230956, 0.008179712295532226, 0.0082227201461792, 0.008263680458068847, 0.008264703750610352, 0.008261631965637208, 0.008284223556518555, 0.008256447792053223, 0.008124608039855957, 0.008112095832824706, 0.008235872268676758, 0.008159232139587403, 0.008217599868774414, 0.008226816177368163, 0.008240127563476562, 0.008168448448181152, 0.008169471740722656, 0.008161279678344726, 0.008228863716125488, 0.008196096420288086, 0.017939456939697264, 0.008122367858886719, 0.008211456298828124, 0.008127488136291505, 0.008119359970092773, 0.008130496025085449, 0.008157183647155761, 0.008208383560180664, 0.008144895553588867, 0.008172575950622558, 0.00817353630065918, 0.008156160354614257, 0.00820736026763916, 0.008217599868774414, 0.008183967590332032, 0.008142687797546387, 0.008202239990234375, 0.008152064323425292, 0.00819711971282959, 0.008274944305419921, 0.008174592018127442, 0.008112128257751466, 0.00818073558807373, 0.00819916820526123, 0.00823516845703125, 0.008149951934814452, 0.008150943756103516, 0.008125439643859863, 0.008324095726013184, 0.008185855865478516, 0.008195072174072266, 0.008164352416992187, 0.008177663803100586, 0.008135680198669434, 0.008136799812316894, 0.008122400283813476, 0.008171520233154296, 0.0081396484375, 0.0081397762298584, 0.008121439933776856, 0.008140704154968263, 0.008135680198669434, 0.008162303924560547, 0.008113216400146485, 0.008119232177734375, 0.008119359970092773, 0.008203200340270996, 0.008232959747314453, 0.008279040336608886, 0.00831488037109375, 0.008162303924560547, 0.008119296073913575, 0.008132608413696289, 0.008130559921264649, 0.008120320320129394, 0.008141823768615723, 0.008146944046020508, 0.008129535675048828, 0.008233983993530274, 0.008136896133422852, 0.00821331214904785, 0.008194047927856446, 0.008301631927490234, 0.01793324851989746, 0.008103936195373536, 0.008190976142883302, 0.008193023681640625, 0.008126463890075684, 0.008236031532287597, 0.008135807991027833, 0.008161151885986328, 0.008132608413696289, 0.008160256385803222, 0.008151040077209473, 0.008136704444885253, 0.0081080322265625, 0.008133631706237793, 0.00812339210510254, 0.00819711971282959, 0.008269824028015137, 0.00824630355834961, 0.008289248466491698, 0.00830463981628418, 0.008291328430175781, 0.008240127563476562, 0.008193023681640625, 0.008219648361206054, 0.008166399955749512, 0.008212479591369629, 0.008227840423583984, 0.008219648361206054, 0.008252415657043457, 0.008233983993530274, 0.008275967597961426, 0.008282112121582032, 0.00840499210357666, 0.008261631965637208, 0.008258560180664062, 0.008220671653747558, 0.00829849624633789, 0.008245247840881348, 0.008240256309509277, 0.008218496322631837, 0.008245247840881348, 0.008250368118286134, 0.008265727996826172, 0.008242176055908204, 0.008274944305419921, 0.008273920059204102, 0.008227840423583984, 0.008287263870239257, 0.008283103942871094, 0.008279040336608886, 0.008276991844177246, 0.008240192413330079, 0.00831174373626709, 0.008235008239746093, 0.008254464149475099, 0.008281087875366211, 0.008260671615600585, 0.008253376007080078, 0.008255488395690918, 0.008173567771911621, 0.008254528045654297, 0.008232895851135253, 0.008352767944335938, 0.018137088775634767, 0.008227840423583984, 0.008237055778503418, 0.008284159660339355, 0.008263680458068847, 0.008300607681274413, 0.008325056076049804, 0.008283167839050294, 0.008232928276062012, 0.008283136367797851, 0.008275967597961426, 0.008245247840881348, 0.008259584426879883, 0.008240127563476562, 0.008249343872070313, 0.008235008239746093, 0.008270879745483399, 0.008240096092224121, 0.008264703750610352, 0.008230912208557128, 0.008284159660339355, 0.008253439903259278, 0.008239104270935058, 0.008243200302124023, 0.008256511688232422, 0.008263680458068847, 0.008289279937744141, 0.008407039642333984, 0.008258560180664062, 0.008266752243041992, 0.008258560180664062, 0.00821350383758545, 0.008254464149475099, 0.008285183906555176, 0.008257535934448243, 0.008248319625854492, 0.008238080024719239, 0.008255488395690918, 0.008238080024719239, 0.008251392364501953, 0.008287232398986816, 0.008246272087097169, 0.008261631965637208, 0.008270879745483399, 0.008243167877197265, 0.008255488395690918, 0.008223744392395019, 0.00821555233001709, 0.008285183906555176, 0.008273920059204102, 0.008279040336608886, 0.008268799781799317, 0.00827187156677246, 0.008263680458068847, 0.008249343872070313, 0.008263680458068847, 0.008266752243041992, 0.008262656211853027, 0.008340479850769043, 0.008260607719421387, 0.008266783714294434, 0.008297439575195312, 0.008256511688232422, 0.018292736053466797, 0.008232959747314453, 0.008257535934448243, 0.008282112121582032, 0.00824015998840332, 0.008266719818115234, 0.008231936454772949, 0.008250368118286134, 0.008265727996826172, 0.008268799781799317, 0.008261631965637208, 0.008258560180664062, 0.008260640144348144, 0.008283103942871094, 0.008264703750610352, 0.008270848274230956, 0.008270848274230956, 0.008254464149475099, 0.008262656211853027, 0.008273920059204102, 0.008258560180664062, 0.008285216331481934, 0.008334303855895996, 0.008269824028015137, 0.008264703750610352, 0.008238080024719239, 0.008263680458068847, 0.008260671615600585, 0.008264639854431152, 0.008259584426879883, 0.008281087875366211, 0.008258560180664062, 0.008289312362670899, 0.008259552001953125, 0.008261631965637208, 0.008243200302124023, 0.008263680458068847, 0.008391679763793946, 0.00827187156677246, 0.00821452808380127, 0.008239104270935058, 0.008252415657043457, 0.008282112121582032, 0.008272895812988281, 0.00822169589996338, 0.008888319969177246, 0.009343999862670899, 0.008645631790161134, 0.00860371208190918, 0.008592320442199706, 0.00861184024810791, 0.008594431877136231, 0.008583168029785156, 0.00859340763092041, 0.008556544303894043, 0.008529919624328614, 0.008550399780273438, 0.008525856018066406, 0.008541152000427247, 0.00851865577697754, 0.00852070426940918, 0.008537088394165039, 0.008536064147949218]",tokens/s,118.39643430936485,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Salesforce/codegen-6B-nl,Salesforce/codegen-6B-nl,cuda,0,42,,,True,,,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",codegen,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-410m,,cuda,0,42,,,,,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,217063f5c507ed7cc255df7e1f64c4333a0b4dfe,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,,MB,1149.001728,2605.187072,0.0,1965.031424,1890.996224,s,10,0.20988812828063966,0.020988812828063967,0.0005355747849219581,0.020890799522399903,0.021313901138305665,0.021906086540222167,0.022379834861755373,"[0.022498271942138673, 0.02067964744567871, 0.020683040618896486, 0.02118230438232422, 0.020906368255615235, 0.020609184265136717, 0.020998016357421875, 0.020876991271972657, 0.020904607772827148, 0.02054969596862793]",tokens/s,12196.973792519819,kWh,2.475870372725344e-07,1.356627377859443e-07,1.5029781415848897e-06,1.8862279166433682e-06,tokens/kWh,135720608.17314383,MB,1149.001728,2605.187072,0.0,1965.031424,1948.84096,s,10,11.4994541015625,1.14994541015625,0.02259531597582969,1.1540154418945314,1.1748104248046876,1.1748933715820313,1.1749597290039064,"[1.132986083984375, 1.1180478515625, 1.1256864013671875, 1.174976318359375, 1.1742025146484374, 1.1569588623046876, 1.1747919921875, 1.1707353515625, 1.151072021484375, 1.1199967041015626]",tokens/s,54.78520931827521,kWh,1.357429260460475e-05,7.438304037412697e-06,2.742707278021723e-05,4.843966942223466e-05,tokens/kWh,1300586.9105102087,,s,629,11.666427900314329,0.01854758012768574,0.0025782765884541027,0.018424800872802734,0.01868211135864258,0.018801049423217774,0.03874668548583985,"[0.019333120346069335, 0.01822719955444336, 0.018052095413208007, 0.01785036849975586, 0.017912832260131836, 0.01824563217163086, 0.01822003173828125, 0.017979391098022462, 0.018020351409912108, 0.01784115219116211, 0.017582080841064454, 0.018381824493408205, 0.017802240371704102, 0.01782374382019043, 0.017821695327758787, 0.017728511810302734, 0.017769472122192383, 0.01779302406311035, 0.017754112243652344, 0.01775103950500488, 0.0178288631439209, 0.017802240371704102, 0.018108415603637695, 0.01839308738708496, 0.018181119918823242, 0.0177838077545166, 0.017814559936523436, 0.017944543838500977, 0.01792617607116699, 0.017824735641479492, 0.01819545555114746, 0.018593791961669923, 0.01862860870361328, 0.01864396858215332, 0.017701887130737306, 0.01794867134094238, 0.017785856246948242, 0.017640447616577147, 0.017756160736083985, 0.017952768325805665, 0.018299903869628906, 0.01778892707824707, 0.0178155517578125, 0.01779814338684082, 0.01785651206970215, 0.01778892707824707, 0.0176312313079834, 0.01764352035522461, 0.018555967330932618, 0.018244543075561524, 0.017858591079711914, 0.017780704498291014, 0.01780940818786621, 0.01779916763305664, 0.017780736923217775, 0.017581056594848633, 0.01767731285095215, 0.017687551498413084, 0.017617919921875, 0.01763839912414551, 0.01765376091003418, 0.017624063491821287, 0.038637569427490234, 0.01761894416809082, 0.017738752365112305, 0.017760320663452147, 0.018018239974975585, 0.017758207321166994, 0.01780531120300293, 0.017740800857543947, 0.01776742362976074, 0.017753087997436523, 0.01776131248474121, 0.0177161922454834, 0.01757798385620117, 0.017582080841064454, 0.01756979179382324, 0.01768448066711426, 0.017797119140625, 0.017819648742675782, 0.017902591705322265, 0.017725439071655275, 0.017622016906738282, 0.017764352798461915, 0.01778892707824707, 0.01756159973144531, 0.01760153579711914, 0.017518592834472657, 0.01761177635192871, 0.01767430305480957, 0.017612735748291017, 0.017827840805053712, 0.01821388816833496, 0.01779199981689453, 0.017655807495117186, 0.017730560302734375, 0.01763430404663086, 0.017555456161499023, 0.01758515167236328, 0.017778688430786133, 0.017779712677001954, 0.017760255813598632, 0.017878015518188475, 0.018019327163696287, 0.01798041534423828, 0.01784832000732422, 0.017753087997436523, 0.017728511810302734, 0.017722368240356445, 0.017688575744628905, 0.017781759262084963, 0.017738752365112305, 0.01775103950500488, 0.017705984115600586, 0.017722368240356445, 0.01775103950500488, 0.017740800857543947, 0.017727487564086913, 0.017756160736083985, 0.01780636787414551, 0.01770800018310547, 0.017709056854248048, 0.017884159088134767, 0.017832000732421874, 0.01772640037536621, 0.03880038452148438, 0.01774284744262695, 0.01780940818786621, 0.01776639938354492, 0.01764454460144043, 0.017582080841064454, 0.01765990447998047, 0.017566719055175782, 0.017531904220581054, 0.01756979179382324, 0.017554431915283202, 0.017549312591552735, 0.01776639938354492, 0.017694719314575197, 0.017758207321166994, 0.017738752365112305, 0.017740800857543947, 0.017714208602905273, 0.0177161922454834, 0.017741823196411134, 0.017597471237182617, 0.017669088363647462, 0.0177807674407959, 0.017611743927001954, 0.01764249610900879, 0.017719295501708983, 0.017574911117553712, 0.01759846305847168, 0.017722368240356445, 0.017743871688842772, 0.017709056854248048, 0.017756160736083985, 0.017721343994140625, 0.017657855987548828, 0.01784832000732422, 0.017743871688842772, 0.017755136489868165, 0.01762918472290039, 0.017661951065063478, 0.017686527252197267, 0.017732608795166017, 0.017581056594848633, 0.017687551498413084, 0.017762304306030274, 0.01760153579711914, 0.01761587142944336, 0.017802240371704102, 0.01785958480834961, 0.018259967803955078, 0.018498559951782227, 0.018420736312866212, 0.01839411163330078, 0.018342912673950194, 0.018408447265625, 0.018513919830322266, 0.018503679275512695, 0.01840127944946289, 0.018415615081787108, 0.018391040802001952, 0.01841459274291992, 0.018415615081787108, 0.01843404769897461, 0.01845248031616211, 0.04163174438476563, 0.019153919219970703, 0.01859071922302246, 0.018542591094970702, 0.01859686470031738, 0.018619392395019533, 0.018637823104858398, 0.018589696884155273, 0.018534400939941405, 0.018463743209838866, 0.018759679794311524, 0.01862348747253418, 0.018619392395019533, 0.01861427116394043, 0.01860710334777832, 0.01860505676269531, 0.01863167953491211, 0.01862348747253418, 0.018504703521728515, 0.018364416122436524, 0.018339839935302735, 0.01841868782043457, 0.018386943817138672, 0.018563072204589845, 0.01875660705566406, 0.01859584045410156, 0.018408447265625, 0.018340864181518556, 0.01860915184020996, 0.018661376953125, 0.01863987159729004, 0.0185743350982666, 0.01846784019470215, 0.018307071685791015, 0.019081247329711913, 0.019641311645507812, 0.01926246452331543, 0.01869004821777344, 0.01863065528869629, 0.018507776260375978, 0.01863680076599121, 0.01880473518371582, 0.01883033561706543, 0.018593791961669923, 0.01878118324279785, 0.018426944732666015, 0.018834367752075195, 0.01860505676269531, 0.01864806365966797, 0.018687007904052734, 0.01850262451171875, 0.018469888687133788, 0.018565120697021483, 0.018506752014160157, 0.01887539291381836, 0.018353151321411132, 0.018382848739624022, 0.01844428825378418, 0.018586624145507814, 0.018646015167236327, 0.01864499282836914, 0.01856716728210449, 0.01858252716064453, 0.03990118408203125, 0.018689023971557618, 0.018391040802001952, 0.018456575393676757, 0.018364416122436524, 0.01858252716064453, 0.01844223976135254, 0.018463743209838866, 0.01861734390258789, 0.018555904388427736, 0.01858252716064453, 0.01862041664123535, 0.018593791961669923, 0.018715648651123046, 0.018874368667602538, 0.018565120697021483, 0.01862553596496582, 0.018663423538208008, 0.018584640502929687, 0.018612159729003906, 0.0186562557220459, 0.01880985641479492, 0.0186429443359375, 0.018539520263671876, 0.018581504821777343, 0.018557952880859374, 0.01856518363952637, 0.01857836723327637, 0.01840332794189453, 0.0184597110748291, 0.018613183975219726, 0.018578432083129884, 0.01918569564819336, 0.019470304489135743, 0.01905971145629883, 0.018675712585449217, 0.01864396858215332, 0.018569215774536133, 0.01860812759399414, 0.018555904388427736, 0.01862758445739746, 0.01862246322631836, 0.018585599899291993, 0.018675712585449217, 0.01862246322631836, 0.018536447525024414, 0.018603008270263673, 0.0186746883392334, 0.018584575653076172, 0.01866035270690918, 0.01864396858215332, 0.018626560211181642, 0.01863065528869629, 0.018704383850097657, 0.018695167541503906, 0.018736127853393555, 0.01862041664123535, 0.01860095977783203, 0.018619392395019533, 0.018626560211181642, 0.018601984024047852, 0.01860710334777832, 0.018592767715454102, 0.03876761627197266, 0.01776742362976074, 0.017737728118896484, 0.017914880752563478, 0.01778892707824707, 0.0177541446685791, 0.017742816925048827, 0.017729536056518554, 0.017701887130737306, 0.017730592727661133, 0.017787872314453126, 0.017812480926513673, 0.01783193588256836, 0.017712127685546874, 0.01775103950500488, 0.017693695068359376, 0.017688575744628905, 0.017854463577270507, 0.017709056854248048, 0.017733631134033204, 0.017722368240356445, 0.01789952087402344, 0.018052095413208007, 0.018482175827026368, 0.018684928894042968, 0.018456575393676757, 0.018522111892700196, 0.018542591094970702, 0.018708480834960937, 0.018362367630004883, 0.018495487213134765, 0.018905088424682616, 0.01884364891052246, 0.01866444778442383, 0.01839107131958008, 0.018424800872802734, 0.0186060791015625, 0.018566144943237304, 0.01942835235595703, 0.02079539108276367, 0.019021823883056642, 0.018715648651123046, 0.0186296329498291, 0.018586624145507814, 0.018601984024047852, 0.018651136398315428, 0.018679807662963867, 0.018680864334106446, 0.01868796730041504, 0.018542591094970702, 0.01862246322631836, 0.01862451171875, 0.018569215774536133, 0.01861840057373047, 0.018598880767822266, 0.018554880142211915, 0.0186296329498291, 0.01868191909790039, 0.018625471115112306, 0.018704383850097657, 0.018687999725341797, 0.018564096450805666, 0.018556928634643553, 0.0400711669921875, 0.018571264266967775, 0.018587648391723634, 0.018572288513183592, 0.018428928375244142, 0.018332672119140626, 0.01866649627685547, 0.01860710334777832, 0.01863680076599121, 0.018708480834960937, 0.01861734390258789, 0.01908531188964844, 0.01901260757446289, 0.018676736831665038, 0.01863372802734375, 0.018671615600585938, 0.01861529541015625, 0.018715648651123046, 0.01864089584350586, 0.018570240020751954, 0.018610176086425782, 0.01864396858215332, 0.018682880401611326, 0.01864192008972168, 0.018651136398315428, 0.0186746883392334, 0.01865011215209961, 0.018670591354370117, 0.01861734390258789, 0.018594816207885743, 0.018662399291992187, 0.018581504821777343, 0.018634752273559572, 0.018699264526367186, 0.018626560211181642, 0.01860406494140625, 0.018647008895874024, 0.01859686470031738, 0.018637823104858398, 0.01860812759399414, 0.018795520782470702, 0.018687999725341797, 0.018673728942871094, 0.018618303298950194, 0.018572288513183592, 0.018663423538208008, 0.01857535934448242, 0.018826271057128908, 0.018756576538085937, 0.01859891128540039, 0.018700288772583007, 0.01866035270690918, 0.01861734390258789, 0.01856924819946289, 0.01868182373046875, 0.018526208877563476, 0.018509824752807616, 0.018593791961669923, 0.018587648391723634, 0.018561023712158203, 0.018670591354370117, 0.0186060791015625, 0.01861324882507324, 0.039984127044677735, 0.018713600158691408, 0.01862041664123535, 0.0183767032623291, 0.018524160385131837, 0.018594816207885743, 0.01860095977783203, 0.018676736831665038, 0.01866963195800781, 0.018686912536621095, 0.018390016555786134, 0.01855897521972656, 0.01864089584350586, 0.01864192008972168, 0.01879347229003906, 0.01859071922302246, 0.018584575653076172, 0.01865318489074707, 0.01863167953491211, 0.018603008270263673, 0.01867366409301758, 0.01859071922302246, 0.018578432083129884, 0.018663423538208008, 0.01860095977783203, 0.01854060745239258, 0.018655168533325196, 0.018570240020751954, 0.01843404769897461, 0.018512895584106445, 0.018512895584106445, 0.018358272552490236, 0.01847091293334961, 0.018565120697021483, 0.01859891128540039, 0.018535423278808593, 0.018537471771240235, 0.01882521629333496, 0.018682880401611326, 0.018626560211181642, 0.018692096710205077, 0.01851910400390625, 0.018366399765014647, 0.018589759826660158, 0.018634687423706053, 0.01860505676269531, 0.018725887298583984, 0.018481184005737304, 0.018608095169067383, 0.01856716728210449, 0.018620479583740236, 0.018594751358032225, 0.01855897521972656, 0.018555904388427736, 0.018577407836914063, 0.018571264266967775, 0.01863167953491211, 0.01864089584350586, 0.018587648391723634, 0.018375680923461913, 0.01840230369567871, 0.018354175567626953, 0.018328575134277342, 0.03990220642089844, 0.018370559692382812, 0.018415615081787108, 0.018601984024047852, 0.018585599899291993, 0.018576383590698242, 0.018538496017456055, 0.018528255462646484, 0.018397184371948243, 0.018325504302978517, 0.01838387107849121, 0.01856211280822754, 0.01857529640197754, 0.018578432083129884, 0.018413568496704103, 0.018363391876220703, 0.018530399322509765, 0.018537376403808595, 0.01856716728210449, 0.018585599899291993, 0.018489343643188477, 0.018374656677246092, 0.018332672119140626, 0.018350080490112306, 0.018700288772583007, 0.018562047958374024, 0.018525184631347655, 0.018594816207885743, 0.018594816207885743, 0.01861529541015625, 0.018557952880859374, 0.018594816207885743, 0.018573312759399413, 0.0186060791015625, 0.018559999465942383, 0.018603008270263673, 0.01859993553161621, 0.0186060791015625, 0.01860095977783203, 0.01859174346923828, 0.018563072204589845, 0.018676736831665038, 0.017737728118896484, 0.017745920181274414, 0.0178657283782959, 0.01777663993835449, 0.01773465538024902, 0.017696767807006835, 0.01761894416809082, 0.017663999557495116, 0.017772544860839845, 0.017740800857543947, 0.017760255813598632, 0.017765439987182618, 0.01773356819152832, 0.01776639938354492, 0.017708032608032227, 0.017718271255493166, 0.017731584548950196, 0.017696767807006835, 0.017729536056518554, 0.017686527252197267, 0.017754112243652344, 0.03869286346435547, 0.017753087997436523, 0.017721343994140625, 0.017739776611328126, 0.017718271255493166, 0.017733663558959962, 0.017747936248779298, 0.017736703872680663, 0.01775923156738281, 0.017720319747924804, 0.017721343994140625, 0.017748992919921876, 0.0178606071472168, 0.017722368240356445, 0.017768447875976562, 0.017736703872680663, 0.017787904739379884, 0.017738752365112305, 0.017804288864135744, 0.018058240890502928, 0.0177838077545166, 0.017785856246948242, 0.01783302307128906, 0.017777599334716798, 0.01784012794494629, 0.017761280059814453, 0.0178155517578125, 0.017747968673706056, 0.017820735931396485, 0.01770284843444824, 0.017772544860839845, 0.01770086479187012, 0.017787904739379884, 0.017833984375, 0.01824870491027832, 0.017516544342041016, 0.017665023803710937, 0.017746944427490235, 0.017778688430786133, 0.017886207580566405, 0.01787494468688965, 0.017752063751220702, 0.017787904739379884, 0.017755136489868165, 0.017726463317871095, 0.017755136489868165, 0.01779199981689453, 0.017703935623168944, 0.017765375137329103, 0.017661951065063478, 0.01773472023010254, 0.018082752227783203, 0.017748992919921876, 0.017746944427490235, 0.017741823196411134, 0.017689599990844726, 0.017728511810302734, 0.017745920181274414, 0.017687551498413084, 0.017747968673706056, 0.017731584548950196, 0.017731584548950196, 0.01779814338684082]",tokens/s,53.915389129782625,,,main,False,False,,,,,,,,,,,,,,,,,,,,,,, float16-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,.,.,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 373, in cached_file raise EnvironmentError( OSError: . does not appear to have a file named config.json. Checkout 'https://huggingface.co/./tree/None' for available files. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,a,a,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/a/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66947477-5fb528cb577bb69b158aa93e;28cdf4a1-8e4b-4f04-b3a2-4b548c956072) Repository Not Found for url: https://huggingface.co/a/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: a is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-7B,Qwen/Qwen-7B,cuda,0,42,,,True,,,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",qwen,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-66b,facebook/opt-66b,cuda,0,42,,,True,,,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",opt,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,stabilityai/stablelm-base-alpha-7b,stabilityai/stablelm-base-alpha-7b,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1286.004736,17205.559296,0.0,16559.112192,16333.262848,s,10,0.8190601959228515,0.08190601959228515,0.0008894510012711463,0.08160480117797851,0.08195909805297852,0.08326615867614746,0.08431180717468262,"[0.08457321929931641, 0.08160256195068359, 0.08166864013671875, 0.08162210845947265, 0.08160704040527343, 0.08160124969482421, 0.0815909423828125, 0.08157849884033203, 0.08163753509521485, 0.08157839965820313]",tokens/s,3125.5334012606936,kWh,9.675536496529412e-07,5.299783271899598e-07,5.8282634701952525e-06,7.325795447038154e-06,tokens/kWh,34945010.6614022,MB,1286.004736,17205.559296,0.0,16559.112192,16464.787968,s,10,23.537657714843746,2.353765771484375,0.0030467430168322355,2.353167602539062,2.3545855468749997,2.3586080810546872,2.361826108398437,"[2.362630615234375, 2.35303955078125, 2.35365673828125, 2.35329638671875, 2.351913330078125, 2.353691650390625, 2.3525146484375, 2.353295654296875, 2.351259765625, 2.352359375]",tokens/s,26.76561991139406,kWh,2.7722381541157158e-05,1.5193045944165864e-05,0.0001337570426424043,0.0001766724701277273,tokens/kWh,356592.0596143443,,s,629,24.234010593414304,0.038527838781262806,0.009732435356185313,0.03733916854858398,0.037465087890625,0.03766134338378906,0.11927461486816406,"[0.03816550445556641, 0.038217727661132815, 0.03773952102661133, 0.03817779159545898, 0.03826073455810547, 0.03797401428222656, 0.03802316665649414, 0.038280193328857424, 0.03813894271850586, 0.037709758758544924, 0.03750707244873047, 0.03788800048828125, 0.037718017578125, 0.03751321411132812, 0.03730022430419922, 0.03722854232788086, 0.03731353759765625, 0.03730022430419922, 0.03722956848144531, 0.03726028823852539, 0.03731558227539063, 0.037308414459228514, 0.0372674560546875, 0.037292030334472655, 0.03728179168701172, 0.03732691192626953, 0.037301216125488285, 0.03730531311035156, 0.037372928619384765, 0.0373043212890625, 0.03733196640014649, 0.03732070541381836, 0.03736678314208984, 0.0373125114440918, 0.037308414459228514, 0.03734425735473633, 0.037343231201171875, 0.037349407196044924, 0.0373606071472168, 0.03731558227539063, 0.037349441528320315, 0.037396415710449216, 0.037346336364746095, 0.03732271957397461, 0.03738623809814453, 0.03735756683349609, 0.03734732818603516, 0.03738726425170898, 0.03740671920776367, 0.03740979385375977, 0.037375999450683595, 0.037367809295654295, 0.037381118774414065, 0.037438465118408204, 0.03740262222290039, 0.03743027114868164, 0.037453887939453125, 0.03753670501708985, 0.03745382308959961, 0.037416961669921874, 0.037422080993652344, 0.03742924880981445, 0.11929497528076172, 0.03723574447631836, 0.037195743560791014, 0.037222400665283206, 0.03723263931274414, 0.03720294570922852, 0.037176319122314457, 0.03724697494506836, 0.037272575378417966, 0.037213214874267576, 0.03728585433959961, 0.03728076934814453, 0.0372408332824707, 0.03723161697387695, 0.03726847839355469, 0.037236736297607424, 0.03727052688598633, 0.037269504547119144, 0.03723775863647461, 0.03727974319458008, 0.0373043212890625, 0.037310462951660156, 0.037343326568603515, 0.03725814437866211, 0.037465087890625, 0.037318656921386716, 0.03728384017944336, 0.037302272796630856, 0.03733913421630859, 0.037288959503173826, 0.03732889556884766, 0.037314624786376954, 0.03753670501708985, 0.03753881454467774, 0.037369857788085936, 0.0374015998840332, 0.03733708953857422, 0.03736883163452148, 0.03743743896484375, 0.03734732818603516, 0.037348350524902346, 0.037367809295654295, 0.03736064147949219, 0.037392383575439454, 0.037407745361328126, 0.037343231201171875, 0.03743539047241211, 0.037381118774414065, 0.03749481582641601, 0.037422046661376954, 0.03740671920776367, 0.037400577545166014, 0.0374015998840332, 0.03736576080322265, 0.03738521575927734, 0.037449726104736326, 0.03744768142700195, 0.03740467071533203, 0.037440513610839846, 0.03742822265625, 0.03752243041992188, 0.03752447891235351, 0.03743231964111328, 0.11940351867675782, 0.037310462951660156, 0.037200897216796876, 0.037269504547119144, 0.037177345275878904, 0.037275646209716795, 0.03723980712890625, 0.037236736297607424, 0.03721318435668945, 0.03735244750976562, 0.037566463470458986, 0.03750400161743164, 0.03728179168701172, 0.03729817581176758, 0.03729100799560547, 0.03724185562133789, 0.037238784790039066, 0.03727667236328125, 0.03722444915771484, 0.03723980712890625, 0.037238784790039066, 0.03730124664306641, 0.03724697494506836, 0.037370880126953124, 0.037282817840576174, 0.037275646209716795, 0.037341182708740234, 0.03733196640014649, 0.0373043212890625, 0.037323776245117186, 0.03733606338500976, 0.03738828659057617, 0.03738828659057617, 0.03736576080322265, 0.03728691101074219, 0.03739852905273437, 0.037369857788085936, 0.03731148910522461, 0.03735039901733399, 0.03741593551635742, 0.03737395095825195, 0.03738214492797851, 0.037340160369873046, 0.03741593551635742, 0.037379070281982424, 0.03738623809814453, 0.03736064147949219, 0.03740364837646484, 0.03736883163452148, 0.03740367889404297, 0.03744457626342773, 0.03749990463256836, 0.03741183853149414, 0.03742105484008789, 0.03786547088623047, 0.03749990463256836, 0.037424129486083986, 0.03742822265625, 0.03741900634765625, 0.03744460678100586, 0.03742310333251953, 0.03742310333251953, 0.037384193420410154, 0.1193154525756836, 0.037152767181396484, 0.03720601654052735, 0.03720601654052735, 0.037217281341552735, 0.03718143844604492, 0.03726540756225586, 0.03722137451171875, 0.03719475173950195, 0.03724697494506836, 0.03722547149658203, 0.037217281341552735, 0.0372490234375, 0.03717539215087891, 0.03718339157104492, 0.03724800109863281, 0.03726335906982422, 0.03722751998901367, 0.037250049591064455, 0.03722547149658203, 0.03722444915771484, 0.03720723342895508, 0.03725606536865234, 0.03726124954223633, 0.037207168579101564, 0.03729804611206055, 0.03729100799560547, 0.03725209426879883, 0.03725414276123047, 0.037303329467773434, 0.0372696647644043, 0.03730720138549805, 0.03724800109863281, 0.037272575378417966, 0.037294078826904296, 0.037321727752685545, 0.03735244750976562, 0.037356544494628906, 0.037392383575439454, 0.0374128646850586, 0.03829555130004883, 0.03767295837402344, 0.037424129486083986, 0.03741798400878906, 0.03739852905273437, 0.03751424026489258, 0.03745792007446289, 0.03771187210083008, 0.037733375549316404, 0.03738521575927734, 0.037389312744140625, 0.03742310333251953, 0.037424129486083986, 0.03757056045532227, 0.03740979385375977, 0.03745177459716797, 0.037449726104736326, 0.03741183853149414, 0.037443584442138675, 0.037449726104736326, 0.03740671920776367, 0.03745792007446289, 0.03739955139160156, 0.1193175048828125, 0.037184513092041016, 0.037198848724365234, 0.03725516891479492, 0.037168128967285156, 0.03719782257080078, 0.03721318435668945, 0.03722038269042969, 0.03718960189819336, 0.03736576080322265, 0.037294078826904296, 0.03721625518798828, 0.037253120422363284, 0.03726540756225586, 0.03725721740722656, 0.03726335906982422, 0.03723161697387695, 0.03733407974243164, 0.03722643280029297, 0.03730124664306641, 0.037354496002197264, 0.037308414459228514, 0.03731763076782227, 0.0372490234375, 0.03726028823852539, 0.03731558227539063, 0.03726438522338867, 0.03728179168701172, 0.0373196792602539, 0.03742617416381836, 0.0373125114440918, 0.037310462951660156, 0.0373043212890625, 0.03729612731933594, 0.037356544494628906, 0.037348350524902346, 0.03727360153198242, 0.037410816192626956, 0.03733299255371094, 0.0373493766784668, 0.037392383575439454, 0.03734732818603516, 0.037349407196044924, 0.037323776245117186, 0.03735958480834961, 0.037378047943115236, 0.037416961669921874, 0.03732992172241211, 0.03735859298706055, 0.0374128646850586, 0.0374128646850586, 0.037408767700195314, 0.03740467071533203, 0.03741183853149414, 0.03747020721435547, 0.037449726104736326, 0.037419071197509764, 0.037399486541748045, 0.03741388702392578, 0.0374128646850586, 0.03745587158203125, 0.03745075225830078, 0.0374200325012207, 0.1192325439453125, 0.037181407928466796, 0.037223423004150394, 0.03721420669555664, 0.037171199798583986, 0.037238784790039066, 0.03721113586425781, 0.03743747329711914, 0.03728278350830078, 0.03728793716430664, 0.03718860626220703, 0.03724185562133789, 0.03724595260620117, 0.03723161697387695, 0.03729100799560547, 0.03728486251831055, 0.037266433715820314, 0.03726438522338867, 0.037324798583984374, 0.03722751998901367, 0.03728179168701172, 0.037359615325927735, 0.03725209426879883, 0.03735756683349609, 0.03728179168701172, 0.037303295135498044, 0.037454849243164064, 0.037443584442138675, 0.037395454406738284, 0.037477375030517575, 0.0373196792602539, 0.037533695220947266, 0.037295135498046875, 0.03740156936645508, 0.03730636978149414, 0.037308414459228514, 0.03734425735473633, 0.03732787322998047, 0.037318656921386716, 0.037779457092285154, 0.037648384094238284, 0.0374477424621582, 0.03735136032104492, 0.03743027114868164, 0.037501022338867186, 0.03740047836303711, 0.03737395095825195, 0.03738623809814453, 0.03745280075073242, 0.03739852905273437, 0.03739033508300781, 0.037353473663330077, 0.03743539047241211, 0.03741593551635742, 0.037410816192626956, 0.037408767700195314, 0.03743129730224609, 0.03739648056030274, 0.03736576080322265, 0.03740979385375977, 0.037427200317382815, 0.03750092697143555, 0.037408767700195314, 0.11928985595703125, 0.03720191955566406, 0.03722751998901367, 0.03722444915771484, 0.0372408332824707, 0.03732787322998047, 0.03727360153198242, 0.03721420669555664, 0.03721932983398438, 0.03725516891479492, 0.03721420669555664, 0.03725107192993164, 0.03723775863647461, 0.03728486251831055, 0.03723263931274414, 0.03728076934814453, 0.03728076934814453, 0.03754598236083984, 0.037479423522949216, 0.037282817840576174, 0.03727465438842773, 0.037279712677001954, 0.03725113677978516, 0.03727251052856445, 0.03729715347290039, 0.03727872085571289, 0.037303295135498044, 0.03731455993652344, 0.03733196640014649, 0.03728079986572266, 0.03732278442382812, 0.037330879211425784, 0.037318656921386716, 0.03735551834106445, 0.03736064147949219, 0.03729919815063477, 0.037375999450683595, 0.03736576080322265, 0.037292030334472655, 0.03737395095825195, 0.03737299346923828, 0.03738515090942383, 0.03738009643554688, 0.037354496002197264, 0.03743539047241211, 0.037340160369873046, 0.03736883163452148, 0.03736064147949219, 0.03735244750976562, 0.03739340972900391, 0.03742105484008789, 0.03736576080322265, 0.03736883163452148, 0.03739648056030274, 0.037465087890625, 0.03740979385375977, 0.03741593551635742, 0.03741183853149414, 0.03741593551635742, 0.037465087890625, 0.03744563293457031, 0.03747020721435547, 0.03749273681640625, 0.11923558044433594, 0.03723161697387695, 0.03720908737182617, 0.03722649765014648, 0.037220352172851565, 0.03726131057739258, 0.037185535430908204, 0.03723161697387695, 0.03720601654052735, 0.03720499038696289, 0.03726339340209961, 0.03725104141235352, 0.03720908737182617, 0.03723263931274414, 0.03726438522338867, 0.03725414276123047, 0.03724492645263672, 0.037303295135498044, 0.037236736297607424, 0.03726540756225586, 0.0372592658996582, 0.037305343627929685, 0.03728384017944336, 0.03730022430419922, 0.03726028823852539, 0.03732787322998047, 0.037323776245117186, 0.03728384017944336, 0.03733097457885742, 0.03732579040527344, 0.03731763076782227, 0.03731353759765625, 0.03731353759765625, 0.03746099090576172, 0.03739033508300781, 0.03735756683349609, 0.037316638946533205, 0.037353439331054686, 0.03737497711181641, 0.03733916854858398, 0.0373073616027832, 0.03736678314208984, 0.03735244750976562, 0.037338111877441404, 0.03737190246582031, 0.03737395095825195, 0.037392383575439454, 0.03742310333251953, 0.03812972640991211, 0.03766998291015625, 0.03739120101928711, 0.037318656921386716, 0.03738521575927734, 0.03737702560424805, 0.03781836700439453, 0.03761270523071289, 0.03736764907836914, 0.03741289520263672, 0.03739235305786133, 0.03735756683349609, 0.037368896484375, 0.037615615844726565, 0.03747532653808594, 0.119289794921875, 0.037147647857666014, 0.037177345275878904, 0.03714563369750976, 0.037135326385498046, 0.037184513092041016, 0.03723267364501953, 0.03721011352539062, 0.03725616073608398, 0.03723571014404297, 0.0372490234375, 0.03728793716430664, 0.03725823974609375, 0.037207038879394534, 0.03726131057739258, 0.0372490234375, 0.03727155303955078, 0.03724800109863281, 0.03726540756225586, 0.037250049591064455, 0.03729100799560547, 0.037292030334472655, 0.03730022430419922, 0.03738214492797851, 0.03729100799560547, 0.03734220886230469, 0.03725721740722656, 0.037305343627929685, 0.03728179168701172, 0.03728076934814453, 0.03731353759765625, 0.037282817840576174, 0.0373309440612793, 0.037318656921386716, 0.037308448791503905, 0.037330913543701175, 0.03733401489257813, 0.037379070281982424, 0.03733708953857422, 0.037341182708740234, 0.03733299255371094, 0.03736678314208984, 0.03733708953857422, 0.037359615325927735, 0.03734732818603516, 0.037372928619384765, 0.03734630584716797, 0.03736166381835938, 0.03739955139160156, 0.03739033508300781, 0.03738009643554688, 0.03736678314208984, 0.03738214492797851, 0.037410816192626956, 0.03740262222290039, 0.03747840118408203, 0.03743027114868164, 0.03745587158203125, 0.03741593551635742, 0.037449726104736326, 0.037422080993652344, 0.037440513610839846, 0.037438465118408204, 0.11929603576660157, 0.03719164657592773, 0.03717222213745117, 0.037195777893066405, 0.03721420669555664, 0.037174270629882815, 0.03722854232788086, 0.037250049591064455, 0.037198848724365234, 0.03721932983398438, 0.03724288177490234, 0.03726335906982422, 0.0372592658996582, 0.03722547149658203, 0.03728998565673828, 0.0372674560546875, 0.037266433715820314, 0.03724390411376953, 0.03729305648803711, 0.03722547149658203, 0.03729919815063477, 0.0372490234375, 0.03728179168701172, 0.03729305648803711, 0.037302272796630856, 0.03729100799560547, 0.037354496002197264, 0.037335041046142575, 0.0373043212890625, 0.037333023071289065, 0.03734832000732422, 0.03730944061279297, 0.03734527969360352, 0.03733708953857422, 0.037294078826904296, 0.037384193420410154, 0.037379070281982424, 0.03732582473754883, 0.037525505065917966, 0.037356544494628906, 0.03730944061279297, 0.03734527969360352, 0.037335041046142575, 0.037343231201171875, 0.03737395095825195, 0.037354496002197264, 0.03778867340087891, 0.037700607299804685, 0.037459968566894535, 0.037424129486083986, 0.03739033508300781, 0.03740364837646484, 0.037440513610839846, 0.03742310333251953, 0.03735244750976562, 0.03736483383178711, 0.03736975860595703, 0.037424129486083986, 0.037395454406738284, 0.037362686157226564, 0.03738521575927734, 0.037424190521240235, 0.03743328094482422]",tokens/s,25.955258110307728,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/gpt-neo-2.7B,EleutherAI/gpt-neo-2.7B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1224.35584,6568.804352,0.0,5922.357248,5577.217024,s,10,0.4181464958190918,0.04181464958190918,0.0006958654947977685,0.04170903968811035,0.04205739555358887,0.0429289701461792,0.04362622982025146,"[0.04380054473876953, 0.041311710357666016, 0.0417088623046875, 0.04128147125244141, 0.0417740478515625, 0.04125260925292969, 0.04169116973876953, 0.041863712310791015, 0.041753150939941405, 0.0417092170715332]",tokens/s,6122.25625611261,kWh,4.910977013707986e-07,2.690972570427783e-07,2.586877357651452e-06,3.347072316065029e-06,tokens/kWh,76484753.18900946,MB,1224.35584,6568.804352,0.0,5922.357248,5677.7216,s,10,15.071193725585939,1.5071193725585936,0.013919136889699726,1.513021240234375,1.5191654296875,1.520555859375,1.521668203125,"[1.5219462890625, 1.4936270751953125, 1.5188564453125, 1.4842158203125, 1.5126497802734375, 1.482237548828125, 1.5132919921875, 1.51275048828125, 1.5146500244140626, 1.51696826171875]",tokens/s,41.80159922770198,kWh,1.763207485333408e-05,9.662340437715863e-06,6.078642334094919e-05,8.808083863199911e-05,tokens/kWh,715252.045489864,,s,629,15.421263866424555,0.024517112665221875,0.0049896068008657805,0.023986175537109376,0.024215961074829102,0.02452869110107422,0.06469328765869141,"[0.023702527999877928, 0.02346700859069824, 0.02386534309387207, 0.023566335678100587, 0.023803903579711915, 0.0237260799407959, 0.02367283248901367, 0.023522335052490233, 0.024013792037963867, 0.02392780876159668, 0.025779199600219727, 0.02533580780029297, 0.0249487361907959, 0.024756223678588866, 0.02507161521911621, 0.024961023330688475, 0.024808448791503908, 0.02456985664367676, 0.02469375991821289, 0.024790016174316407, 0.024847360610961915, 0.024400896072387695, 0.024198144912719727, 0.02404761505126953, 0.024053760528564453, 0.024054784774780274, 0.02408038330078125, 0.024011775970458983, 0.02424630355834961, 0.024010719299316405, 0.024255487442016603, 0.024172544479370117, 0.02406707191467285, 0.023996416091918944, 0.024115232467651366, 0.02395644760131836, 0.02389504051208496, 0.02404351997375488, 0.023977983474731446, 0.023967744827270508, 0.023998464584350586, 0.023997440338134765, 0.0238919677734375, 0.024001535415649415, 0.02392268753051758, 0.02409676742553711, 0.024029184341430664, 0.024349695205688478, 0.024198144912719727, 0.023954431533813478, 0.023983104705810547, 0.023504896163940428, 0.023941120147705077, 0.023979007720947267, 0.024004608154296874, 0.024052736282348632, 0.023962623596191408, 0.023967744827270508, 0.024026111602783205, 0.02412748718261719, 0.024000511169433594, 0.023957504272460937, 0.064716796875, 0.022793216705322264, 0.022795263290405272, 0.022979583740234375, 0.022732799530029296, 0.022683647155761717, 0.022777856826782225, 0.02305536079406738, 0.022742015838623047, 0.022817792892456053, 0.02281062316894531, 0.02281881523132324, 0.022849536895751952, 0.022841344833374022, 0.022746112823486327, 0.02307072067260742, 0.022808576583862306, 0.02281881523132324, 0.02309734344482422, 0.024195072174072265, 0.02413260841369629, 0.024039424896240235, 0.02392268753051758, 0.02414080047607422, 0.023953407287597657, 0.023987199783325194, 0.024164352416992187, 0.02412544059753418, 0.023916543960571288, 0.024071168899536134, 0.02408243179321289, 0.024163328170776367, 0.02410188865661621, 0.024246271133422852, 0.02395136070251465, 0.023928831100463867, 0.02390323257446289, 0.023824384689331055, 0.02392268753051758, 0.024177663803100585, 0.023941120147705077, 0.02390937614440918, 0.024196096420288086, 0.024517631530761717, 0.02409881591796875, 0.02429030418395996, 0.02448076820373535, 0.023998464584350586, 0.023993343353271485, 0.02391859245300293, 0.024026111602783205, 0.023973888397216796, 0.024217599868774413, 0.024048639297485352, 0.023992319107055664, 0.024046592712402344, 0.02388582420349121, 0.023954431533813478, 0.02413363265991211, 0.024011775970458983, 0.024441856384277344, 0.024069120407104492, 0.023926816940307617, 0.0662558364868164, 0.02402617645263672, 0.02421343994140625, 0.024191999435424806, 0.023953407287597657, 0.023941120147705077, 0.02408755111694336, 0.024247360229492188, 0.02403014373779297, 0.024065023422241212, 0.024004608154296874, 0.023974912643432617, 0.024122367858886717, 0.02412031936645508, 0.024237056732177735, 0.024367103576660155, 0.02409164810180664, 0.024049663543701173, 0.02415001678466797, 0.0241080322265625, 0.024044544219970702, 0.024590335845947265, 0.025598976135253908, 0.02472243118286133, 0.024053760528564453, 0.02408550453186035, 0.02411724853515625, 0.023975936889648438, 0.02392780876159668, 0.024123392105102538, 0.023989248275756835, 0.02412031936645508, 0.02408755111694336, 0.023942144393920898, 0.024004608154296874, 0.024016895294189454, 0.023983104705810547, 0.02394316864013672, 0.024215551376342775, 0.024052736282348632, 0.024034303665161134, 0.023977983474731446, 0.024175647735595704, 0.02399228858947754, 0.023941120147705077, 0.024011775970458983, 0.023984128952026368, 0.02404761505126953, 0.024028160095214843, 0.023963647842407225, 0.02408243179321289, 0.02408448028564453, 0.02409574317932129, 0.024041471481323243, 0.024024063110351563, 0.02411929512023926, 0.024164352416992187, 0.024006656646728516, 0.02407935905456543, 0.024006656646728516, 0.024033279418945314, 0.024009727478027345, 0.023846912384033202, 0.06453657531738281, 0.024250368118286132, 0.024689664840698244, 0.024176639556884767, 0.0240762882232666, 0.02391756820678711, 0.023970815658569337, 0.02408550453186035, 0.023933952331542968, 0.023987199783325194, 0.023974912643432617, 0.022748159408569335, 0.02271232032775879, 0.0228351993560791, 0.02292531204223633, 0.02286489677429199, 0.022804479598999023, 0.02291814422607422, 0.022806528091430665, 0.022854656219482423, 0.022798336029052735, 0.022879232406616212, 0.02294169616699219, 0.022965248107910157, 0.023023616790771483, 0.022773759841918945, 0.022775808334350587, 0.02290995216369629, 0.02287001609802246, 0.02280147171020508, 0.022824895858764647, 0.023021568298339845, 0.022845439910888672, 0.02453606414794922, 0.024366079330444337, 0.02414899253845215, 0.024164352416992187, 0.02282700729370117, 0.02293350410461426, 0.02290995216369629, 0.022833152770996092, 0.023589887619018556, 0.023979007720947267, 0.023924736022949217, 0.02392268753051758, 0.024048639297485352, 0.02389811134338379, 0.023986175537109376, 0.024030208587646484, 0.023988224029541014, 0.02407526397705078, 0.024137727737426756, 0.024154111862182616, 0.024102912902832032, 0.024057855606079103, 0.024160255432128908, 0.023957504272460937, 0.024088640213012696, 0.024088512420654296, 0.024049663543701173, 0.023931903839111326, 0.024070144653320313, 0.02409369659423828, 0.06637366485595703, 0.02399945640563965, 0.023895072937011718, 0.02400048065185547, 0.023804927825927736, 0.02392166328430176, 0.02453811264038086, 0.024240127563476564, 0.02390425682067871, 0.023999488830566407, 0.023937023162841797, 0.024030208587646484, 0.023940095901489256, 0.023974912643432617, 0.0239554557800293, 0.024170495986938476, 0.023941120147705077, 0.024060928344726562, 0.023968767166137696, 0.02409881591796875, 0.02406809616088867, 0.02404761505126953, 0.02392268753051758, 0.02415001678466797, 0.02395136070251465, 0.023985151290893555, 0.02391449546813965, 0.023982080459594726, 0.023857152938842774, 0.024016895294189454, 0.023963647842407225, 0.024015871047973633, 0.023992319107055664, 0.02412441635131836, 0.023940095901489256, 0.02390937614440918, 0.023932928085327147, 0.02391756820678711, 0.02388787269592285, 0.02406707191467285, 0.023996416091918944, 0.024073215484619142, 0.024025087356567384, 0.02391142463684082, 0.02392166328430176, 0.0239052791595459, 0.02390630340576172, 0.02413465690612793, 0.024228864669799805, 0.023875583648681642, 0.023976959228515626, 0.024035327911376952, 0.023974912643432617, 0.024275968551635742, 0.023946239471435548, 0.024088607788085938, 0.02397590446472168, 0.02405891227722168, 0.02394108772277832, 0.023968767166137696, 0.02389401626586914, 0.02387763214111328, 0.023993343353271485, 0.0646328353881836, 0.02274003219604492, 0.022840255737304686, 0.022915071487426757, 0.022675455093383787, 0.02269593620300293, 0.022765567779541016, 0.02290278434753418, 0.02272870445251465, 0.02271027183532715, 0.02285055923461914, 0.022790143966674805, 0.02269491195678711, 0.02274406433105469, 0.02273587226867676, 0.02285772705078125, 0.022716415405273437, 0.022757375717163086, 0.02280243110656738, 0.022813695907592774, 0.022749183654785156, 0.02286796760559082, 0.022738943099975584, 0.022784000396728517, 0.023946239471435548, 0.024412160873413087, 0.024033279418945314, 0.02398316764831543, 0.023886783599853516, 0.024020992279052734, 0.023994367599487306, 0.02410393524169922, 0.024201215744018553, 0.02389708709716797, 0.023579647064208984, 0.023825408935546875, 0.02407526397705078, 0.02394419288635254, 0.024005632400512695, 0.023945215225219727, 0.02394316864013672, 0.02384588813781738, 0.02390937614440918, 0.023990272521972656, 0.023900159835815428, 0.02390118408203125, 0.023953407287597657, 0.023946271896362305, 0.023935968399047852, 0.02393087959289551, 0.023932960510253905, 0.02397795104980469, 0.023931903839111326, 0.024135679244995118, 0.023968767166137696, 0.024130559921264647, 0.023814144134521483, 0.023952415466308594, 0.024034271240234373, 0.024012800216674804, 0.02392166328430176, 0.024035327911376952, 0.0241080322265625, 0.06620877075195312, 0.02390425682067871, 0.023989248275756835, 0.02406707191467285, 0.023900159835815428, 0.023883775711059572, 0.02395955276489258, 0.024028160095214843, 0.02409881591796875, 0.02395136070251465, 0.024015871047973633, 0.023974912643432617, 0.023941120147705077, 0.023990272521972656, 0.023957504272460937, 0.02407935905456543, 0.024101951599121093, 0.023955392837524413, 0.024070144653320313, 0.023953407287597657, 0.023975936889648438, 0.023948287963867186, 0.024016895294189454, 0.024061952590942383, 0.02389401626586914, 0.023995391845703123, 0.02390323257446289, 0.0239237117767334, 0.023880704879760743, 0.023946239471435548, 0.023965696334838867, 0.02408345603942871, 0.02408243179321289, 0.023875583648681642, 0.02395955276489258, 0.023977983474731446, 0.02395136070251465, 0.02387353515625, 0.0239913272857666, 0.0239400634765625, 0.023867391586303712, 0.024819711685180663, 0.02526924705505371, 0.024327167510986326, 0.02408755111694336, 0.023847936630249023, 0.02387660789489746, 0.024040447235107423, 0.02392985534667969, 0.02392780876159668, 0.023999488830566407, 0.0239554557800293, 0.023963647842407225, 0.023948287963867186, 0.023963712692260743, 0.024064960479736327, 0.023990272521972656, 0.023940095901489256, 0.024056831359863282, 0.0239554557800293, 0.023979040145874025, 0.02396054458618164, 0.023993343353271485, 0.06662348937988281, 0.023967744827270508, 0.023908351898193358, 0.024180736541748047, 0.02411929512023926, 0.02389401626586914, 0.023971839904785155, 0.024238079071044923, 0.02413465690612793, 0.02394726371765137, 0.02389504051208496, 0.024073215484619142, 0.023953407287597657, 0.023954431533813478, 0.02391142463684082, 0.024244224548339844, 0.023635967254638672, 0.023966720581054687, 0.024005632400512695, 0.02392064094543457, 0.023957504272460937, 0.024060928344726562, 0.024276992797851563, 0.02409062385559082, 0.024030208587646484, 0.024052736282348632, 0.023953407287597657, 0.023953407287597657, 0.024046592712402344, 0.023967744827270508, 0.023928831100463867, 0.023970815658569337, 0.024035327911376952, 0.024019968032836913, 0.024034303665161134, 0.024056831359863282, 0.023982080459594726, 0.024048639297485352, 0.023958528518676758, 0.023963680267333986, 0.02398409652709961, 0.02387353515625, 0.024016895294189454, 0.02393497657775879, 0.02391244888305664, 0.02387660789489746, 0.023961599349975587, 0.023979007720947267, 0.024007680892944337, 0.023847936630249023, 0.02404351997375488, 0.024081407546997072, 0.024025087356567384, 0.023989248275756835, 0.02392985534667969, 0.0241582088470459, 0.02391551971435547, 0.023980031967163085, 0.024030208587646484, 0.02408038330078125, 0.023938047409057618, 0.02404150390625, 0.02408239936828613, 0.06629273223876953, 0.023961599349975587, 0.02391244888305664, 0.024033279418945314, 0.023908351898193358, 0.023960575103759766, 0.023942144393920898, 0.02411008071899414, 0.024114175796508788, 0.024154111862182616, 0.02411008071899414, 0.024345600128173828, 0.02409267234802246, 0.024060928344726562, 0.023980031967163085, 0.02405887985229492, 0.02410700798034668, 0.024012800216674804, 0.024027135848999022, 0.024000511169433594, 0.02394316864013672, 0.02409676742553711, 0.024017919540405275, 0.02410495948791504, 0.02389504051208496, 0.023977983474731446, 0.02389708709716797, 0.023954431533813478, 0.02391756820678711, 0.023976959228515626, 0.02390118408203125, 0.023998464584350586, 0.024046592712402344, 0.024009727478027345, 0.023974912643432617, 0.023964672088623046, 0.023991296768188477, 0.023875583648681642, 0.024039424896240235, 0.023950336456298828, 0.023946239471435548, 0.023931903839111326, 0.023973888397216796, 0.02384899139404297, 0.023880672454833985, 0.024056831359863282, 0.02391551971435547, 0.023967744827270508, 0.023965696334838867, 0.023977983474731446, 0.024016895294189454, 0.023931903839111326, 0.025040895462036132, 0.02511564826965332, 0.024242176055908202, 0.024224767684936522, 0.024010751724243166, 0.0240762882232666, 0.02409574317932129, 0.024003583908081053, 0.02388172721862793, 0.023799808502197265, 0.023843839645385743, 0.06624665832519532, 0.023966720581054687, 0.023916543960571288, 0.024000511169433594, 0.02409267234802246, 0.023859199523925782, 0.023932960510253905, 0.024041439056396486, 0.024039424896240235, 0.02388787269592285, 0.023962623596191408, 0.024070144653320313, 0.0239554557800293, 0.02409369659423828, 0.023989248275756835, 0.024129535675048826, 0.024054784774780274, 0.024180736541748047, 0.024052736282348632, 0.02410188865661621, 0.023964672088623046, 0.023999488830566407, 0.024024063110351563, 0.024048639297485352, 0.023953407287597657, 0.023979007720947267, 0.02411724853515625, 0.02415001678466797, 0.02412748718261719, 0.024147968292236328, 0.023969791412353517, 0.025000959396362304, 0.024792064666748048, 0.024194047927856444, 0.02411315155029297, 0.024164352416992187, 0.024329216003417968, 0.0239554557800293, 0.024147968292236328, 0.02424115180969238, 0.024040447235107423, 0.02394419288635254, 0.023937023162841797, 0.023932928085327147, 0.02394419288635254, 0.02387049674987793, 0.023878623962402343, 0.023984128952026368, 0.02388787269592285, 0.02390220832824707, 0.023937023162841797, 0.023908351898193358, 0.0245166072845459, 0.024081407546997072, 0.023981056213378905, 0.024105983734130858, 0.024060928344726562, 0.024065023422241212, 0.024303615570068358, 0.02432204818725586, 0.023970815658569337, 0.024014848709106446, 0.02412646484375]",tokens/s,40.78783719987242,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,1,1,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/1/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669325f6-7bd54d296d8d351e22abeb2b;884dcd0c-c096-4d98-bd29-9192f38daf56) Repository Not Found for url: https://huggingface.co/1/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: 1 is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-7b,tiiuae/falcon-7b,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1278.64832,15196.48768,0.0,14550.040576,13944.957952,s,10,0.8168599700927734,0.08168599700927734,0.0017125434954674181,0.08090859222412108,0.08301425933837892,0.0847530174255371,0.08614402389526367,"[0.08649177551269531, 0.08102448272705078, 0.081532958984375, 0.08075113677978515, 0.08064419555664062, 0.08055107116699219, 0.08077008056640625, 0.0807927017211914, 0.08262786865234376, 0.08167369842529297]",tokens/s,3133.9520771341663,kWh,9.578823829637206e-07,5.248717686505011e-07,5.411986856467704e-06,6.894741008081926e-06,tokens/kWh,37129748.557620965,MB,1278.976,15196.48768,0.0,14550.040576,13945.037824,s,10,29.099600341796872,2.9099600341796874,0.007177875359054941,2.9103421630859376,2.9147032958984376,2.9201170288085936,2.924448015136719,"[2.92553076171875, 2.91293896484375, 2.90659423828125, 2.89784130859375, 2.900981689453125, 2.913500244140625, 2.913074462890625, 2.910731201171875, 2.908454345703125, 2.909953125]",tokens/s,21.649781873296273,kWh,3.440782479594686e-05,1.8857076072979383e-05,0.00012732428599913426,0.00018058918686806048,tokens/kWh,348858.0966147667,,s,629,29.782428653717,0.04734885318555968,0.009705406038681034,0.04612710571289062,0.04654249038696289,0.046845132446289066,0.12721811279296877,"[0.04724531173706055, 0.04674355316162109, 0.04698316955566406, 0.047265792846679686, 0.04716134262084961, 0.04681830215454102, 0.04755251312255859, 0.04667596817016602, 0.047184928894042966, 0.04760367965698242, 0.047064064025878906, 0.04638924789428711, 0.04663097763061524, 0.04603897476196289, 0.046258174896240234, 0.04610559844970703, 0.046148609161376954, 0.04613324737548828, 0.04612300872802735, 0.046173183441162106, 0.046808063507080076, 0.04640153503417969, 0.0463636474609375, 0.046004222869873046, 0.046124031066894534, 0.046413822174072264, 0.04611481475830078, 0.04610047912597656, 0.04619878387451172, 0.046088191986083986, 0.04631552124023437, 0.04662681579589844, 0.046306304931640625, 0.046071807861328126, 0.04640358352661133, 0.04619673538208008, 0.04623155212402344, 0.046061569213867185, 0.04623360061645508, 0.046031871795654294, 0.04622444915771484, 0.04634515380859375, 0.046496768951416016, 0.04604108810424805, 0.04610662460327149, 0.04598988723754883, 0.04610559844970703, 0.046086143493652344, 0.04684799957275391, 0.047483905792236325, 0.04658380889892578, 0.04656742477416992, 0.046712833404541014, 0.046077953338623044, 0.046235649108886716, 0.04609638214111328, 0.04611276626586914, 0.04597555160522461, 0.046074878692626955, 0.04595916748046875, 0.04613119888305664, 0.046042110443115236, 0.12811776733398436, 0.04605952072143555, 0.04600115203857422, 0.04608204650878906, 0.0464087028503418, 0.04651212692260742, 0.0463636474609375, 0.046663681030273435, 0.04601036834716797, 0.046137344360351565, 0.04633804702758789, 0.04612710571289062, 0.04618137741088867, 0.04613017654418945, 0.04610662460327149, 0.04616499328613281, 0.04618137741088867, 0.04732723236083984, 0.04656639862060547, 0.046723072052001956, 0.04606054306030274, 0.04642406463623047, 0.0460871696472168, 0.04612812805175781, 0.046025726318359376, 0.046159870147705076, 0.04615679931640625, 0.04614144134521484, 0.046077953338623044, 0.046176254272460936, 0.04607078552246094, 0.04613119888305664, 0.04616499328613281, 0.046139392852783206, 0.0460882568359375, 0.04608607864379883, 0.04655001449584961, 0.04639641571044922, 0.04612198257446289, 0.046192638397216795, 0.046088191986083986, 0.04611481475830078, 0.0464640007019043, 0.04693196868896484, 0.04614656066894531, 0.04637286376953125, 0.0463267822265625, 0.04611379241943359, 0.04614348983764648, 0.046134273529052736, 0.046066688537597655, 0.046129150390625, 0.046033920288085936, 0.046412799835205076, 0.04630835342407227, 0.046104576110839846, 0.0459950065612793, 0.046139392852783206, 0.046080001831054686, 0.046091262817382815, 0.046063617706298826, 0.046069793701171875, 0.04605436706542969, 0.12855807495117189, 0.04613017654418945, 0.04614553451538086, 0.046115840911865234, 0.04598476791381836, 0.046150657653808595, 0.04603596878051758, 0.04613020706176758, 0.046008289337158205, 0.046139392852783206, 0.046064640045166014, 0.046167041778564455, 0.045859840393066405, 0.0459315185546875, 0.04600115203857422, 0.046088191986083986, 0.046102527618408204, 0.046164031982421874, 0.04594899368286133, 0.04687449645996094, 0.04614553451538086, 0.046265342712402346, 0.04603801727294922, 0.04614553451538086, 0.04600729751586914, 0.046104576110839846, 0.04601036834716797, 0.04622441482543945, 0.046024673461914065, 0.046072830200195314, 0.04602982330322265, 0.046050304412841796, 0.04598169708251953, 0.046182430267333985, 0.04615779113769531, 0.04609740829467773, 0.04604108810424805, 0.04616908645629883, 0.04602982330322265, 0.04614348983764648, 0.04624076843261719, 0.046134273529052736, 0.04596121597290039, 0.04598886489868164, 0.04605132675170898, 0.046124031066894534, 0.04602880096435547, 0.046102527618408204, 0.04608204650878906, 0.04615167999267578, 0.04605235290527344, 0.046077953338623044, 0.04600627136230469, 0.04659609603881836, 0.04694937515258789, 0.046682113647460936, 0.04607078552246094, 0.04613328170776367, 0.045957088470458984, 0.04610969543457031, 0.046020606994628906, 0.04612300872802735, 0.04600729751586914, 0.12716134643554688, 0.04579328155517578, 0.04561920166015625, 0.04581683349609375, 0.04567244720458984, 0.04595609664916992, 0.045908992767333984, 0.045808639526367184, 0.04564889526367188, 0.04589567947387695, 0.04581990432739258, 0.046002174377441404, 0.0459315185546875, 0.04613324737548828, 0.0460134391784668, 0.04604313659667969, 0.046031871795654294, 0.04607897567749023, 0.045985790252685545, 0.046080001831054686, 0.04597555160522461, 0.045770751953125, 0.0458526725769043, 0.046034942626953124, 0.04603289413452148, 0.04609228897094726, 0.04603596878051758, 0.04611481475830078, 0.0459950065612793, 0.04586822509765625, 0.045786945343017575, 0.046077953338623044, 0.04607385635375977, 0.04610969543457031, 0.04605542373657227, 0.046069759368896485, 0.0460840950012207, 0.04621414566040039, 0.04606054306030274, 0.04615270233154297, 0.04593356704711914, 0.046043167114257814, 0.046147552490234375, 0.046104576110839846, 0.04616396713256836, 0.04612505722045898, 0.04605952072143555, 0.0460840950012207, 0.0459048957824707, 0.04605440139770508, 0.045695999145507815, 0.04575033569335937, 0.04584646224975586, 0.04615475082397461, 0.04641996765136719, 0.046148609161376954, 0.04630220794677734, 0.04613017654418945, 0.04597555160522461, 0.04587417602539062, 0.04580966567993164, 0.04609740829467773, 0.045960193634033204, 0.12724018859863281, 0.04617523193359375, 0.04598169708251953, 0.046050304412841796, 0.046004222869873046, 0.04606771087646484, 0.04600729751586914, 0.046840831756591796, 0.046383102416992186, 0.04615884780883789, 0.04669440078735351, 0.046377983093261715, 0.046058494567871096, 0.04604528045654297, 0.04600412750244141, 0.04619673538208008, 0.0461415023803711, 0.046054336547851564, 0.0458076171875, 0.0458526725769043, 0.04584960174560547, 0.04589056015014648, 0.046167041778564455, 0.045965312957763675, 0.04626432037353516, 0.04598896026611328, 0.0458064956665039, 0.04604108810424805, 0.045841407775878903, 0.045843616485595706, 0.04580438232421875, 0.045888671875, 0.0458238410949707, 0.0459417610168457, 0.04587519836425781, 0.04611686325073242, 0.046020606994628906, 0.04628582382202148, 0.046341121673583986, 0.04613539123535156, 0.04677519989013672, 0.04654188919067383, 0.04608710479736328, 0.04647116851806641, 0.0460134391784668, 0.04598988723754883, 0.04575539016723633, 0.04590591812133789, 0.045795326232910154, 0.04592435073852539, 0.04572985458374024, 0.04600620651245117, 0.045720577239990234, 0.04581587219238281, 0.046367679595947266, 0.04591939163208008, 0.04584534454345703, 0.04603903961181641, 0.045674495697021485, 0.04581478500366211, 0.0458721923828125, 0.04580550384521485, 0.04575743865966797, 0.12702105712890624, 0.04650393676757812, 0.045930496215820314, 0.046371841430664064, 0.04579328155517578, 0.046322689056396485, 0.04591206359863281, 0.04608512115478516, 0.04635443115234375, 0.04627763366699219, 0.046440448760986325, 0.046430206298828124, 0.04608822250366211, 0.04622025680541992, 0.04643430328369141, 0.04610559844970703, 0.04599603271484375, 0.04643430328369141, 0.046266368865966793, 0.04613631820678711, 0.046031871795654294, 0.046734336853027345, 0.04647628784179687, 0.04646912002563477, 0.04619468688964844, 0.046102527618408204, 0.046042110443115236, 0.0460840950012207, 0.04624486541748047, 0.046120960235595705, 0.04608102416992187, 0.046139392852783206, 0.046304256439208984, 0.046502910614013675, 0.04657254409790039, 0.046514175415039063, 0.046148609161376954, 0.046238719940185545, 0.046150657653808595, 0.046440448760986325, 0.046170112609863284, 0.047333377838134766, 0.04649574279785156, 0.04621823883056641, 0.04622335815429687, 0.04635238265991211, 0.046048255920410154, 0.04623360061645508, 0.046167041778564455, 0.04617113494873047, 0.04607590484619141, 0.04604927825927734, 0.046024768829345704, 0.046093246459960935, 0.04625612640380859, 0.04633599853515625, 0.04598067092895508, 0.046601215362548826, 0.046064640045166014, 0.046053375244140625, 0.046148609161376954, 0.04622335815429687, 0.04604006576538086, 0.12749721527099608, 0.04660736083984375, 0.046058494567871096, 0.04618239974975586, 0.04592127990722656, 0.04613529586791992, 0.04612505722045898, 0.04609228897094726, 0.046050304412841796, 0.04617113494873047, 0.046072830200195314, 0.047718399047851565, 0.04685619354248047, 0.04643430328369141, 0.04601446533203125, 0.04626227188110352, 0.046099456787109375, 0.04617932891845703, 0.04600320053100586, 0.046104576110839846, 0.046074878692626955, 0.04616908645629883, 0.04643840026855469, 0.046415870666503906, 0.046235649108886716, 0.046120960235595705, 0.04616089630126953, 0.04636774444580078, 0.04645273590087891, 0.04626739120483398, 0.046140415191650394, 0.0462213134765625, 0.046020606994628906, 0.046548992156982424, 0.04602675247192383, 0.04628070449829102, 0.04603084945678711, 0.04619878387451172, 0.04623462295532227, 0.04618751907348633, 0.046170112609863284, 0.04618547058105469, 0.046066688537597655, 0.0461475830078125, 0.04641689682006836, 0.046388225555419924, 0.04610047912597656, 0.04614451217651367, 0.04603289413452148, 0.04613324737548828, 0.04603903961181641, 0.04620185470581055, 0.04618854522705078, 0.04628684616088867, 0.04658585739135742, 0.046611457824707034, 0.04618137741088867, 0.046388225555419924, 0.04597862243652344, 0.046069759368896485, 0.04617216110229492, 0.04619571304321289, 0.046015487670898435, 0.12743577575683593, 0.046284801483154295, 0.04600627136230469, 0.0461578254699707, 0.04592127990722656, 0.046231582641601564, 0.046134239196777345, 0.04617523193359375, 0.046069759368896485, 0.04610047912597656, 0.046061569213867185, 0.04607078552246094, 0.046209022521972655, 0.04585881423950195, 0.0459048957824707, 0.04615167999267578, 0.046268417358398435, 0.04612812805175781, 0.04615167999267578, 0.046074878692626955, 0.046023681640625, 0.046192638397216795, 0.046382080078125, 0.04645171356201172, 0.04590796661376953, 0.046355457305908204, 0.04589363098144531, 0.04599091339111328, 0.04620492935180664, 0.04676403045654297, 0.04693401718139648, 0.04628684616088867, 0.04616908645629883, 0.046324798583984375, 0.046093246459960935, 0.046148609161376954, 0.046023681640625, 0.04611072158813476, 0.04615884780883789, 0.04627872085571289, 0.046179264068603516, 0.046271488189697264, 0.04598076629638672, 0.046097312927246094, 0.04636876678466797, 0.04610047912597656, 0.04605132675170898, 0.0460134391784668, 0.04602880096435547, 0.04654489517211914, 0.04620185470581055, 0.04636774444580078, 0.04655411148071289, 0.046339073181152345, 0.04632371139526367, 0.04628889465332031, 0.04616089630126953, 0.04612607955932617, 0.046063617706298826, 0.046666751861572264, 0.04620800018310547, 0.04634316635131836, 0.04616294479370117, 0.1293834228515625, 0.046120960235595705, 0.04658585739135742, 0.04615679931640625, 0.046004222869873046, 0.04636159896850586, 0.046257152557373046, 0.04631449508666992, 0.04600320053100586, 0.04629401779174805, 0.046344192504882815, 0.04694527816772461, 0.04612812805175781, 0.04591820907592774, 0.04593868637084961, 0.04597555160522461, 0.04588236618041992, 0.04599091339111328, 0.045930496215820314, 0.04607590484619141, 0.04600627136230469, 0.04606771087646484, 0.046252033233642575, 0.0460840950012207, 0.046083072662353515, 0.04614963150024414, 0.047263744354248044, 0.046074878692626955, 0.04591734313964844, 0.04601327896118164, 0.046183425903320315, 0.04595404815673828, 0.04585779190063476, 0.046061569213867185, 0.045813758850097655, 0.04598067092895508, 0.045753345489501954, 0.04597862243652344, 0.045870079040527346, 0.04591513442993164, 0.046205951690673826, 0.04625920104980469, 0.04597760009765625, 0.04642201614379883, 0.04693196868896484, 0.04663808059692383, 0.04605542373657227, 0.04632985687255859, 0.046134273529052736, 0.04627251052856445, 0.046268417358398435, 0.04616294479370117, 0.04601139068603516, 0.046170112609863284, 0.046265342712402346, 0.04632783889770508, 0.04625616073608398, 0.04617824172973633, 0.04601036834716797, 0.04608102416992187, 0.04599705505371094, 0.0462366714477539, 0.04605235290527344, 0.12838911437988282, 0.046166015625, 0.04619059371948242, 0.04621721649169922, 0.04603289413452148, 0.04617113494873047, 0.04603903961181641, 0.046266368865966793, 0.04611174392700195, 0.046150657653808595, 0.04604108810424805, 0.04611072158813476, 0.04637081527709961, 0.046137344360351565, 0.04603084945678711, 0.04611481475830078, 0.04602880096435547, 0.046238719940185545, 0.04616294479370117, 0.04616396713256836, 0.046203903198242184, 0.0461844482421875, 0.046137344360351565, 0.04612716674804687, 0.04622739028930664, 0.046271488189697264, 0.046120960235595705, 0.046089214324951173, 0.04615577697753906, 0.046219264984130856, 0.04609024047851563, 0.046140415191650394, 0.046089214324951173, 0.04733747100830078, 0.047145984649658204, 0.046473217010498044, 0.04608102416992187, 0.04627251052856445, 0.04632473754882813, 0.04624998474121094, 0.046091262817382815, 0.04611072158813476, 0.04615679931640625, 0.046117889404296876, 0.0464087028503418, 0.04618137741088867, 0.046063617706298826, 0.0460871696472168, 0.04605952072143555, 0.04614451217651367, 0.04609228897094726, 0.04626739120483398, 0.046050304412841796, 0.04609638214111328, 0.046034942626953124, 0.046148609161376954, 0.04610559844970703, 0.04605033493041992, 0.046024673461914065, 0.04611481475830078, 0.04597452926635742, 0.04616089630126953, 0.046031871795654294]",tokens/s,21.1198357029052,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,meta-llama/Llama-2-7b-hf,meta-llama/Llama-2-7b-hf,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.1,,,,MB,1295.515648,14563.147776,0.0,13916.700672,13738.740736,s,10,0.9719790725708009,0.09719790725708008,0.0002467090630834595,0.09713264083862305,0.09734582443237305,0.09761553535461426,0.09783130409240723,"[0.09788524627685546, 0.09697257232666015, 0.09724601745605468, 0.09706687927246094, 0.09704972839355469, 0.09717404937744141, 0.09714019012451172, 0.097285888671875, 0.09712509155273437, 0.09703340911865234]",tokens/s,2633.801562444159,kWh,1.1487160460720042e-06,6.294444560269554e-07,6.279986145883548e-06,8.058146647982508e-06,tokens/kWh,31769091.72583672,MB,1295.515648,14563.147776,0.0,13916.700672,13889.779712,s,10,22.808874267578126,2.2808874267578125,0.0033197490949248166,2.2796990966796873,2.2831444091796875,2.2866671752929686,2.2894853881835937,"[2.29018994140625, 2.2792412109375, 2.281403564453125, 2.278939697265625, 2.282361572265625, 2.27857080078125, 2.280372314453125, 2.278765625, 2.280156982421875, 2.27887255859375]",tokens/s,27.62082830609133,kWh,2.6858589152173573e-05,1.471768967219689e-05,0.00011710322367171637,0.00015867950249608684,tokens/kWh,397026.70482946356,,s,629,23.64582501220704,0.037592726569486534,0.011615114725788953,0.03618611145019531,0.036350975036621096,0.03641610260009766,0.13384734680175783,"[0.03671244812011719, 0.03670937728881836, 0.03684864044189453, 0.036789249420166016, 0.03671039962768555, 0.03680153656005859, 0.03674521636962891, 0.0367534065246582, 0.03668479919433594, 0.03653324890136719, 0.03638169479370117, 0.03667148971557617, 0.03632230377197266, 0.036296703338623046, 0.03610521697998047, 0.03611545562744141, 0.03607244873046875, 0.03615027236938476, 0.03607244873046875, 0.03613183975219727, 0.03607551956176758, 0.03616665649414062, 0.03613798522949219, 0.03616563034057617, 0.036109310150146484, 0.03619839859008789, 0.036212734222412106, 0.03623833465576172, 0.03614822387695313, 0.03623833465576172, 0.03615129470825195, 0.03623731231689453, 0.03620044708251953, 0.03625574493408203, 0.036139007568359374, 0.03621785736083984, 0.036160511016845705, 0.03622809600830078, 0.03616767883300781, 0.03626803207397461, 0.03623526382446289, 0.0362977294921875, 0.03620761489868164, 0.03626393508911133, 0.03616665649414062, 0.036296703338623046, 0.036193279266357424, 0.03640217590332031, 0.036393985748291016, 0.03705344009399414, 0.03623731231689453, 0.03630284881591797, 0.03621376037597656, 0.036307968139648435, 0.03625676727294922, 0.03641241455078125, 0.03630489730834961, 0.03640524673461914, 0.03638784027099609, 0.03646464157104492, 0.03624038314819336, 0.0363612174987793, 0.13373951721191407, 0.0359628791809082, 0.03603558349609375, 0.035988479614257815, 0.03606425476074219, 0.03600281524658203, 0.03609804916381836, 0.03602841567993164, 0.03614310455322266, 0.036013057708740234, 0.03610521697998047, 0.03601919937133789, 0.03688550567626953, 0.036125694274902344, 0.03615641784667969, 0.03600896072387695, 0.03621376037597656, 0.03601715087890625, 0.036421630859375, 0.03603558349609375, 0.036144126892089845, 0.03604172897338867, 0.03614720153808594, 0.03603968048095703, 0.036135936737060545, 0.036050945281982424, 0.03618099212646484, 0.03606425476074219, 0.036208641052246096, 0.0360816650390625, 0.03621478271484375, 0.03607756805419922, 0.03616976165771484, 0.036099040985107425, 0.036206592559814454, 0.036083713531494144, 0.036225025177001956, 0.036125694274902344, 0.036222976684570314, 0.03615129470825195, 0.0363612174987793, 0.03615334320068359, 0.03625062561035156, 0.03610419082641601, 0.03623833465576172, 0.0361338882446289, 0.036350975036621096, 0.03615545654296875, 0.036314048767089845, 0.0361420783996582, 0.036274177551269535, 0.036135936737060545, 0.03625676727294922, 0.036149246215820316, 0.03633663940429688, 0.03625164794921875, 0.03632128143310547, 0.0362158088684082, 0.036345855712890625, 0.03619123077392578, 0.036296703338623046, 0.036195327758789066, 0.03633152008056641, 0.13437440490722657, 0.03602841567993164, 0.036144126892089845, 0.03600998306274414, 0.03608268737792969, 0.0359741439819336, 0.03607961654663086, 0.03605401611328125, 0.03620044708251953, 0.03605401611328125, 0.03617587280273438, 0.03604582214355469, 0.036139007568359374, 0.03600691223144531, 0.03611443328857422, 0.036040702819824216, 0.0361451530456543, 0.03606118392944336, 0.036206592559814454, 0.036070400238037106, 0.03619123077392578, 0.03610214233398437, 0.03619430541992188, 0.036149246215820316, 0.03620249557495117, 0.03612160110473633, 0.03639712142944336, 0.03632428741455078, 0.0364595832824707, 0.036346817016601564, 0.03626291275024414, 0.03611238479614258, 0.036229118347167966, 0.03611033630371094, 0.036173824310302735, 0.03612160110473633, 0.03625888061523438, 0.03617683029174805, 0.03623731231689453, 0.03618611145019531, 0.03627724838256836, 0.03614003372192383, 0.036241409301757815, 0.03620761489868164, 0.03624345779418945, 0.03618201446533203, 0.03630284881591797, 0.036229118347167966, 0.036313087463378906, 0.03626598358154297, 0.03640115356445312, 0.03620454406738281, 0.03629875183105469, 0.036206592559814454, 0.036313087463378906, 0.03623731231689453, 0.03637350463867187, 0.03623731231689453, 0.03634483337402344, 0.03619123077392578, 0.036435966491699216, 0.03627110290527344, 0.036370433807373044, 0.1338429412841797, 0.03596799850463867, 0.03599462509155273, 0.03596908950805664, 0.03604268646240234, 0.03597619247436523, 0.03601510238647461, 0.036018177032470705, 0.036103168487548826, 0.03603459167480469, 0.03607036972045898, 0.03602431869506836, 0.03612057495117187, 0.03608063888549805, 0.036101119995117184, 0.03602227020263672, 0.036141056060791016, 0.036037631988525394, 0.03614720153808594, 0.03613081741333008, 0.03620556640625, 0.036111358642578126, 0.03616358566284179, 0.03612057495117187, 0.0362239990234375, 0.0362158088684082, 0.036185089111328124, 0.036089855194091795, 0.03618815994262695, 0.03610009765625, 0.0361973762512207, 0.036070400238037106, 0.0361420783996582, 0.03613798522949219, 0.036193279266357424, 0.03614720153808594, 0.03616767883300781, 0.03611443328857422, 0.036192256927490236, 0.03616767883300781, 0.036176929473876955, 0.03610620880126953, 0.03621478271484375, 0.03615129470825195, 0.03632537460327148, 0.03624959945678711, 0.0363765754699707, 0.03626803207397461, 0.03626496124267578, 0.03623526382446289, 0.036393985748291016, 0.036178943634033206, 0.03628236770629883, 0.03624857711791992, 0.036350975036621096, 0.036209663391113284, 0.03629260635375976, 0.03623731231689453, 0.036345855712890625, 0.03622195053100586, 0.03631206512451172, 0.03624755096435547, 0.03637964630126953, 0.13388902282714843, 0.03619123077392578, 0.03631411361694336, 0.03607244873046875, 0.03605401611328125, 0.03600588989257812, 0.03607449722290039, 0.036040702819824216, 0.03611852645874023, 0.03611443328857422, 0.036111358642578126, 0.03606220626831055, 0.03612057495117187, 0.03607756805419922, 0.03613081741333008, 0.036117534637451175, 0.03639497756958008, 0.036547584533691405, 0.03621785736083984, 0.03610419082641601, 0.036154369354248046, 0.03612364959716797, 0.03615846252441406, 0.03617279815673828, 0.0361420783996582, 0.0361082878112793, 0.03620665740966797, 0.036141056060791016, 0.03622803115844726, 0.03614617538452149, 0.036190208435058595, 0.036127742767333985, 0.03626803207397461, 0.03614617538452149, 0.03618099212646484, 0.03618611145019531, 0.036219905853271485, 0.03620249557495117, 0.036187137603759766, 0.036211711883544925, 0.03625062561035156, 0.03623833465576172, 0.036222976684570314, 0.03623833465576172, 0.03626700973510742, 0.03623833465576172, 0.036279296875, 0.03629568099975586, 0.03633356857299805, 0.03628236770629883, 0.036482048034667966, 0.03628441619873047, 0.036416511535644534, 0.03630694580078125, 0.0362977294921875, 0.036305950164794924, 0.03636835098266602, 0.03628851318359375, 0.036279296875, 0.03619635009765625, 0.0363581428527832, 0.036321311950683596, 0.03662947082519531, 0.13398629760742187, 0.035969024658203126, 0.03604582214355469, 0.035972095489501955, 0.03607244873046875, 0.03602227020263672, 0.0360786247253418, 0.03598844909667969, 0.03606016159057617, 0.03602739334106445, 0.03606630325317383, 0.03599257659912109, 0.03608883285522461, 0.036067329406738284, 0.03606425476074219, 0.036127742767333985, 0.036139007568359374, 0.03609395217895508, 0.03616358566284179, 0.036036609649658206, 0.03613798522949219, 0.03601510238647461, 0.03615641784667969, 0.03606016159057617, 0.03616972732543945, 0.03605606460571289, 0.03617587280273438, 0.03610419082641601, 0.0361973762512207, 0.0361338882446289, 0.0362158088684082, 0.036162559509277346, 0.036274177551269535, 0.03614617538452149, 0.03617587280273438, 0.03609702301025391, 0.036176929473876955, 0.03612361526489258, 0.03632025527954102, 0.03613798522949219, 0.03619430541992188, 0.036178943634033206, 0.03625369644165039, 0.036155391693115234, 0.036211711883544925, 0.036162559509277346, 0.03627008056640625, 0.036219905853271485, 0.036239360809326174, 0.03620044708251953, 0.036257793426513675, 0.036185089111328124, 0.036316158294677735, 0.036187137603759766, 0.03630080032348633, 0.03619839859008789, 0.036283424377441406, 0.036208606719970705, 0.03628953552246094, 0.03623116683959961, 0.03642060852050781, 0.03631411361694336, 0.03641548919677735, 0.13395455932617187, 0.03599052810668945, 0.03614720153808594, 0.036125694274902344, 0.036152320861816405, 0.036111358642578126, 0.03606220626831055, 0.03599769592285156, 0.03607551956176758, 0.03599462509155273, 0.036070400238037106, 0.036016128540039063, 0.03618099212646484, 0.03625065612792969, 0.036178913116455075, 0.03604377746582031, 0.036155391693115234, 0.03611340713500977, 0.036157440185546875, 0.03605196762084961, 0.03618406295776367, 0.03607449722290039, 0.03620761489868164, 0.03609702301025391, 0.036225025177001956, 0.03618406295776367, 0.036192256927490236, 0.036085823059082034, 0.036200382232666015, 0.03607551956176758, 0.036193279266357424, 0.036152320861816405, 0.036239360809326174, 0.03609907150268555, 0.03619839859008789, 0.036094974517822266, 0.036195327758789066, 0.03610009765625, 0.03619839859008789, 0.03617279815673828, 0.0363694076538086, 0.03623628616333008, 0.03624038314819336, 0.03615027236938476, 0.03628134536743164, 0.03614617538452149, 0.03629260635375976, 0.03621785736083984, 0.03628953552246094, 0.036157440185546875, 0.03628851318359375, 0.036195327758789066, 0.03641241455078125, 0.036261886596679685, 0.03632025527954102, 0.03630182266235352, 0.03639295959472656, 0.036275199890136715, 0.03639910507202149, 0.03626393508911133, 0.03641753768920898, 0.03630182266235352, 0.03634182357788086, 0.13439788818359374, 0.0360079345703125, 0.03605606460571289, 0.03600384140014649, 0.03608268737792969, 0.03603251266479492, 0.03609907150268555, 0.036083713531494144, 0.03614617538452149, 0.036021247863769534, 0.036087806701660154, 0.03601510238647461, 0.03616358566284179, 0.03602227020263672, 0.036087806701660154, 0.03601715087890625, 0.03631411361694336, 0.03606022262573242, 0.03610617446899414, 0.03602431869506836, 0.036173824310302735, 0.036055038452148434, 0.03619839859008789, 0.036040702819824216, 0.036155391693115234, 0.03604889678955078, 0.036195327758789066, 0.03604889678955078, 0.03621478271484375, 0.03607756805419922, 0.03621068954467774, 0.03614003372192383, 0.03623219299316406, 0.036119552612304685, 0.03624345779418945, 0.036109310150146484, 0.036193279266357424, 0.03608268737792969, 0.03621478271484375, 0.036139007568359374, 0.03622195053100586, 0.036124671936035156, 0.036261886596679685, 0.03609600067138672, 0.036241409301757815, 0.03612876892089844, 0.03622707366943359, 0.03615641784667969, 0.03623833465576172, 0.036124671936035156, 0.03625062561035156, 0.036127742767333985, 0.03625676727294922, 0.036154369354248046, 0.036316158294677735, 0.03622092819213867, 0.036362239837646484, 0.036229118347167966, 0.03637350463867187, 0.03627315139770508, 0.036370433807373044, 0.03621887969970703, 0.03640729522705078, 0.13388287353515624, 0.03603046417236328, 0.03609804916381836, 0.03600281524658203, 0.03606425476074219, 0.03600076675415039, 0.0360970573425293, 0.036080608367919924, 0.03607449722290039, 0.03603148651123047, 0.036106239318847655, 0.03604684829711914, 0.0361267204284668, 0.03606118392944336, 0.03609088134765625, 0.03611033630371094, 0.036340736389160154, 0.03616972732543945, 0.0362158088684082, 0.03606937789916992, 0.03618304061889648, 0.036127742767333985, 0.03618409729003906, 0.03610723114013672, 0.036222976684570314, 0.036171775817871094, 0.03619635009765625, 0.03615129470825195, 0.03626496124267578, 0.0361973762512207, 0.03627212905883789, 0.0362690544128418, 0.03627724838256836, 0.03609190368652344, 0.036203521728515625, 0.036154369354248046, 0.03621068954467774, 0.0361082878112793, 0.03621478271484375, 0.03618815994262695, 0.0362158088684082, 0.03610521697998047, 0.03622604751586914, 0.0361451530456543, 0.03624038314819336, 0.0361973762512207, 0.036241409301757815, 0.03619430541992188, 0.03625267028808594, 0.036203521728515625, 0.0362608642578125, 0.0362342414855957, 0.03635507202148437, 0.036222976684570314, 0.03633152008056641, 0.03627315139770508, 0.036350975036621096, 0.03623731231689453, 0.0363694076538086, 0.03622092819213867, 0.03633561706542969, 0.03625369644165039, 0.03633359909057617, 0.13384906005859376, 0.03600076675415039, 0.036040702819824216, 0.03595673751831055, 0.03603353500366211, 0.03603251266479492, 0.03610009765625, 0.03606528091430664, 0.036094974517822266, 0.03629875183105469, 0.03629875183105469, 0.03609702301025391, 0.0361082878112793, 0.03610726547241211, 0.036070400238037106, 0.03609292984008789, 0.0361420783996582, 0.0361082878112793, 0.036109344482421875, 0.036064224243164064, 0.036116481781005856, 0.03602739334106445, 0.03611545562744141, 0.036057086944580076, 0.03612364959716797, 0.036086784362792966, 0.036195327758789066, 0.036055038452148434, 0.036173824310302735, 0.03615129470825195, 0.03620556640625, 0.036125694274902344, 0.036157440185546875, 0.03614003372192383, 0.036192256927490236, 0.036119552612304685, 0.036171775817871094, 0.03616358566284179, 0.03618815994262695, 0.0361451530456543, 0.03617587280273438, 0.03619839859008789, 0.03620249557495117, 0.036193279266357424, 0.03623628616333008, 0.036222976684570314, 0.036229118347167966, 0.03628134536743164, 0.03627008056640625, 0.03622092819213867, 0.03626700973510742, 0.03623014450073242, 0.03625676727294922, 0.03622707366943359, 0.03623014450073242, 0.03625164794921875, 0.03628953552246094, 0.03617484664916992, 0.0362762222290039, 0.03631411361694336, 0.036296703338623046, 0.03622604751586914, 0.03629875183105469]",tokens/s,26.600890418299304,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-rw-1b,tiiuae/falcon-rw-1b,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1213.095936,3538.419712,0.0,2891.972608,2759.918592,s,10,0.22366297340393068,0.022366297340393065,0.00031044612423922424,0.02226214408874512,0.022430630111694334,0.022861218643188474,0.023205689468383788,"[0.023291807174682617, 0.022239007949829102, 0.022261728286743165, 0.022259519577026366, 0.022193151473999022, 0.022251327514648436, 0.022297216415405274, 0.02226255989074707, 0.022334943771362303, 0.022271711349487303]",tokens/s,11445.792573707287,kWh,2.643789069112664e-07,1.448672797228913e-07,1.3312234114854498e-06,1.7404695981196073e-06,tokens/kWh,147086740.42717025,MB,1213.095936,3538.419712,0.0,2891.972608,2796.953088,s,10,10.460348999023438,1.0460348999023439,0.010496409539496829,1.0462456665039062,1.0595597045898437,1.0617674621582032,1.0635336682128906,"[1.0639752197265624, 1.032802490234375, 1.0487801513671875, 1.0443074951171876, 1.048183837890625, 1.055190185546875, 1.0375479736328126, 1.031301025390625, 1.059069091796875, 1.0391915283203126]",tokens/s,60.227436011821,kWh,1.2212047820330672e-05,6.691703820415936e-06,3.374263798091576e-05,5.264638962166234e-05,tokens/kWh,1196663.2555953555,,s,629,10.642643985748302,0.01691994274363798,0.00268828401669231,0.016693248748779296,0.016881459045410156,0.017036256408691407,0.03882958724975586,"[0.019746816635131836, 0.018382848739624022, 0.01701273536682129, 0.016849983215332032, 0.01669932746887207, 0.016692224502563476, 0.01680691146850586, 0.016694271087646484, 0.016714752197265623, 0.01717043113708496, 0.016660480499267577, 0.016712703704833985, 0.016696319580078126, 0.016719871520996094, 0.01678643226623535, 0.016929792404174804, 0.016880640029907225, 0.016780288696289062, 0.016722976684570314, 0.016667680740356447, 0.01679865646362305, 0.016866336822509764, 0.016791519165039064, 0.016763904571533202, 0.0167587833404541, 0.01677926445007324, 0.017565696716308594, 0.01683046340942383, 0.016694271087646484, 0.016649215698242188, 0.01680076789855957, 0.016748544692993163, 0.016890880584716796, 0.01702195167541504, 0.017115135192871094, 0.017147903442382813, 0.016978944778442383, 0.016745471954345705, 0.01680588722229004, 0.016702463150024414, 0.016724992752075195, 0.016688127517700196, 0.016785408020019533, 0.016694271087646484, 0.016761856079101564, 0.01659596824645996, 0.016747520446777343, 0.016703487396240235, 0.016726015090942382, 0.016715776443481444, 0.016660480499267577, 0.016672767639160157, 0.016738304138183592, 0.016698400497436525, 0.016764896392822266, 0.016734207153320312, 0.016719871520996094, 0.016715776443481444, 0.016722944259643553, 0.01683967971801758, 0.016669696807861328, 0.016727039337158203, 0.03879935836791992, 0.016088064193725587, 0.016125951766967773, 0.016087039947509766, 0.016087039947509766, 0.0161474552154541, 0.016137216567993166, 0.01622425651550293, 0.016097280502319337, 0.016143360137939454, 0.016110591888427735, 0.01618227195739746, 0.016087039947509766, 0.016135168075561524, 0.01609939193725586, 0.01618118476867676, 0.016144384384155275, 0.016116735458374023, 0.01604812812805176, 0.016072704315185548, 0.016060415267944335, 0.016150527954101563, 0.0160501766204834, 0.016152576446533205, 0.01623142433166504, 0.016093183517456054, 0.01621504020690918, 0.016142335891723633, 0.016142335891723633, 0.01616486358642578, 0.016107519149780272, 0.016161792755126952, 0.01607372856140137, 0.016130048751831053, 0.016067647933959962, 0.016158655166625978, 0.016133119583129883, 0.016957439422607423, 0.017089536666870117, 0.01683558464050293, 0.016681983947753908, 0.016679935455322266, 0.016657407760620118, 0.016873472213745116, 0.016767999649047852, 0.016739391326904298, 0.016672704696655275, 0.016738304138183592, 0.016660480499267577, 0.016710655212402344, 0.016677888870239257, 0.01679769515991211, 0.016688127517700196, 0.016759807586669923, 0.016949247360229493, 0.016829439163208008, 0.016728063583374024, 0.016753664016723634, 0.016701440811157226, 0.01678745651245117, 0.016648191452026367, 0.016714752197265623, 0.0166246395111084, 0.03943526458740235, 0.016693311691284178, 0.0166246395111084, 0.0167576961517334, 0.016649215698242188, 0.016711679458618164, 0.016668672561645507, 0.01620684814453125, 0.01604096031188965, 0.016103424072265626, 0.016026624679565428, 0.01613107109069824, 0.016123903274536132, 0.016136192321777345, 0.016090112686157225, 0.01617715263366699, 0.016080896377563478, 0.01618227195739746, 0.01616383934020996, 0.01639628791809082, 0.01618534469604492, 0.016620607376098634, 0.01668396759033203, 0.016914432525634765, 0.016708608627319335, 0.016821279525756835, 0.016722911834716796, 0.016842752456665038, 0.017936384201049805, 0.016990207672119142, 0.016742399215698242, 0.016995328903198242, 0.01679974365234375, 0.016899072647094726, 0.016739328384399413, 0.016884735107421875, 0.016718847274780273, 0.016689151763916017, 0.01685196876525879, 0.01682636833190918, 0.016773120880126953, 0.01680486488342285, 0.016684032440185546, 0.016734207153320312, 0.016665599822998048, 0.016699392318725585, 0.016716800689697265, 0.01679363250732422, 0.01685091209411621, 0.016731136322021483, 0.016702463150024414, 0.016702463150024414, 0.016687103271484375, 0.016688127517700196, 0.016746496200561522, 0.016740352630615234, 0.016687103271484375, 0.0168724479675293, 0.01676697540283203, 0.016708608627319335, 0.016713727951049806, 0.016679935455322266, 0.016884735107421875, 0.039583744049072264, 0.016735231399536133, 0.016842752456665038, 0.01686835289001465, 0.01699635124206543, 0.016905216217041014, 0.016713727951049806, 0.01683046340942383, 0.016661504745483398, 0.016703487396240235, 0.016673791885375978, 0.016793600082397463, 0.016720895767211915, 0.01682841682434082, 0.01681920051574707, 0.01701785659790039, 0.01680486488342285, 0.016732160568237304, 0.016656383514404297, 0.016712703704833985, 0.016737279891967775, 0.016724992752075195, 0.017552383422851564, 0.01803468894958496, 0.017076223373413087, 0.01682636833190918, 0.016740352630615234, 0.01677516746520996, 0.016680959701538087, 0.016748544692993163, 0.01665433692932129, 0.016776191711425782, 0.016656383514404297, 0.016677888870239257, 0.016722944259643553, 0.016694271087646484, 0.016719871520996094, 0.016635904312133788, 0.01658880043029785, 0.016753664016723634, 0.016187423706054686, 0.016183263778686525, 0.016101375579833984, 0.016130048751831053, 0.016136192321777345, 0.01617100715637207, 0.016133119583129883, 0.01617817687988281, 0.016125951766967773, 0.016105472564697267, 0.016135168075561524, 0.016138240814208983, 0.016153600692749022, 0.016149503707885742, 0.016149503707885742, 0.016191488265991212, 0.016100351333618163, 0.016106496810913085, 0.01607891273498535, 0.016128959655761718, 0.016160768508911134, 0.016234495162963866, 0.01611782455444336, 0.038784961700439456, 0.016092159271240233, 0.016078847885131836, 0.016096256256103517, 0.016038911819458008, 0.01607372856140137, 0.016079967498779296, 0.016116640090942384, 0.01607372856140137, 0.01614847946166992, 0.016490495681762696, 0.016752639770507814, 0.016740352630615234, 0.016748544692993163, 0.01664204788208008, 0.016764928817749023, 0.016713727951049806, 0.016732160568237304, 0.016655359268188476, 0.016690176010131837, 0.01661337661743164, 0.016695295333862305, 0.016664575576782227, 0.016708608627319335, 0.016668672561645507, 0.016719871520996094, 0.016652288436889647, 0.016652288436889647, 0.016694271087646484, 0.016711679458618164, 0.016660480499267577, 0.016717824935913086, 0.0166430721282959, 0.016740352630615234, 0.016675840377807616, 0.01665433692932129, 0.016657407760620118, 0.016653312683105468, 0.01661030387878418, 0.016744447708129884, 0.0166430721282959, 0.016686080932617187, 0.01665843200683594, 0.016695295333862305, 0.016761856079101564, 0.016723968505859374, 0.016695295333862305, 0.01679974365234375, 0.016723968505859374, 0.016702463150024414, 0.016692224502563476, 0.01675775909423828, 0.016685056686401366, 0.016703487396240235, 0.016943103790283204, 0.017343488693237305, 0.017329151153564454, 0.017098751068115235, 0.01680793571472168, 0.016756736755371093, 0.016683008193969725, 0.01676697540283203, 0.016693248748779296, 0.03884134292602539, 0.016709632873535156, 0.016669696807861328, 0.016719871520996094, 0.016689184188842774, 0.016759775161743165, 0.01664204788208008, 0.016987136840820313, 0.01676288032531738, 0.016747520446777343, 0.016665599822998048, 0.016711679458618164, 0.01663488006591797, 0.016769023895263673, 0.016825344085693358, 0.01676288032531738, 0.016709632873535156, 0.01677004814147949, 0.01663488006591797, 0.016739328384399413, 0.016669696807861328, 0.01682022476196289, 0.016744447708129884, 0.01712646484375, 0.01674131202697754, 0.016681983947753908, 0.016688127517700196, 0.016685056686401366, 0.016688127517700196, 0.016731136322021483, 0.016680959701538087, 0.016761856079101564, 0.016635904312133788, 0.016827423095703124, 0.017042400360107422, 0.01685196876525879, 0.01666662406921387, 0.016714752197265623, 0.016942079544067384, 0.016849920272827147, 0.016755712509155272, 0.01680281639099121, 0.016699392318725585, 0.016907264709472656, 0.016764928817749023, 0.01675775909423828, 0.016653312683105468, 0.016780288696289062, 0.016729087829589845, 0.016679935455322266, 0.016691200256347655, 0.016729087829589845, 0.016729087829589845, 0.016898048400878905, 0.01681510353088379, 0.016793600082397463, 0.016667648315429686, 0.016705535888671876, 0.016717887878417968, 0.016744384765625, 0.016676864624023437, 0.016750591278076172, 0.016693248748779296, 0.038989822387695314, 0.01615564727783203, 0.01643212890625, 0.01620889663696289, 0.016089088439941408, 0.01613209533691406, 0.01621299171447754, 0.01618534469604492, 0.016108543395996093, 0.016183296203613282, 0.016108543395996093, 0.016142400741577148, 0.016252864837646486, 0.016739328384399413, 0.01678643226623535, 0.016846847534179688, 0.01665126419067383, 0.016749568939208984, 0.016655359268188476, 0.016711679458618164, 0.016707584381103514, 0.016752639770507814, 0.016715776443481444, 0.016785408020019533, 0.016759807586669923, 0.016711679458618164, 0.016638975143432617, 0.01672915267944336, 0.016703424453735353, 0.016715776443481444, 0.016692224502563476, 0.016737279891967775, 0.016662527084350585, 0.016824352264404298, 0.016655328750610352, 0.016709632873535156, 0.016706560134887697, 0.016710655212402344, 0.016649215698242188, 0.01681203269958496, 0.016696319580078126, 0.016676864624023437, 0.0166297607421875, 0.016670719146728515, 0.01662054443359375, 0.01697996711730957, 0.016746528625488283, 0.016205791473388673, 0.016093183517456054, 0.016110591888427735, 0.016071680068969727, 0.016118783950805664, 0.016091136932373046, 0.016118783950805664, 0.016125951766967773, 0.01620377540588379, 0.016045055389404296, 0.016102399826049805, 0.01616486358642578, 0.01623040008544922, 0.01615667152404785, 0.016145408630371092, 0.016202751159667968, 0.03892736053466797, 0.016202848434448244, 0.01609718322753906, 0.01623664093017578, 0.016144287109375, 0.01615875244140625, 0.016097248077392576, 0.016191488265991212, 0.01601126480102539, 0.016101375579833984, 0.016119808197021485, 0.016145408630371092, 0.0160501766204834, 0.016121856689453123, 0.016079872131347657, 0.016150527954101563, 0.016052223205566405, 0.016144384384155275, 0.016013311386108398, 0.016270336151123048, 0.01619353675842285, 0.01613107109069824, 0.01620684814453125, 0.016146432876586913, 0.016064512252807618, 0.01617715263366699, 0.016037887573242187, 0.016080896377563478, 0.015911935806274414, 0.016020479202270507, 0.016078847885131836, 0.016141311645507812, 0.01617407989501953, 0.016072704315185548, 0.016047103881835938, 0.0163450870513916, 0.01659391975402832, 0.016726079940795897, 0.0165467529296875, 0.016746496200561522, 0.016579584121704103, 0.01661952018737793, 0.01660620880126953, 0.016683008193969725, 0.01662566375732422, 0.016697343826293946, 0.01657241630554199, 0.016726015090942382, 0.016737279891967775, 0.016675840377807616, 0.016628768920898436, 0.01666044807434082, 0.01662566375732422, 0.01663692855834961, 0.016673791885375978, 0.016777215957641603, 0.01665023994445801, 0.016698368072509767, 0.016705535888671876, 0.016711679458618164, 0.016918527603149415, 0.016889856338500975, 0.016701440811157226, 0.03908198547363281, 0.016928768157958983, 0.016750591278076172, 0.016894975662231446, 0.016755712509155272, 0.016773120880126953, 0.016668672561645507, 0.01684377670288086, 0.016655359268188476, 0.01742336082458496, 0.01676697540283203, 0.016841728210449217, 0.0166748161315918, 0.016662527084350585, 0.016638975143432617, 0.016793600082397463, 0.016696319580078126, 0.016792575836181642, 0.0166748161315918, 0.016683008193969725, 0.016738304138183592, 0.016737279891967775, 0.016659456253051756, 0.016973823547363282, 0.016729087829589845, 0.016684032440185546, 0.016661504745483398, 0.016670719146728515, 0.01782374382019043, 0.017117183685302736, 0.017903615951538086, 0.017076223373413087, 0.01677824020385742, 0.016805919647216797, 0.01674339294433594, 0.016678911209106445, 0.016683008193969725, 0.016695327758789062, 0.017027040481567383, 0.016885759353637696, 0.016695295333862305, 0.016716800689697265, 0.01668614387512207, 0.01672390365600586, 0.01665433692932129, 0.016684032440185546, 0.016631807327270508, 0.016915456771850586, 0.016738304138183592, 0.016769023895263673, 0.016723968505859374, 0.016729120254516602, 0.016629728317260742, 0.016749568939208984, 0.01681305694580078, 0.017292287826538084, 0.016854015350341797, 0.01683558464050293, 0.01681305694580078, 0.016695295333862305, 0.016717824935913086, 0.016825344085693358, 0.016498687744140626, 0.03892326354980469, 0.016114688873291014, 0.016193599700927735, 0.016104383468627928, 0.016054271697998047, 0.01611065673828125, 0.01612691116333008, 0.016154624938964843, 0.016071680068969727, 0.016145408630371092, 0.016097280502319337, 0.016134143829345703, 0.01620070457458496, 0.01621811294555664, 0.016194559097290038, 0.016272384643554686, 0.016123903274536132, 0.016147520065307615, 0.016159679412841795, 0.016142335891723633, 0.016067583084106444, 0.016115711212158202, 0.016098335266113283, 0.016190431594848634, 0.016086015701293945, 0.01610655975341797, 0.01611974334716797, 0.017861631393432616, 0.017024063110351564, 0.016791488647460936, 0.01675468826293945, 0.01680179214477539, 0.016688127517700196, 0.016660480499267577, 0.016681983947753908, 0.016655359268188476, 0.016639999389648438, 0.016705535888671876, 0.016690176010131837, 0.016734207153320312, 0.0166748161315918, 0.016631807327270508, 0.016660480499267577, 0.01665843200683594, 0.016662527084350585, 0.016676864624023437, 0.016685056686401366, 0.01680793571472168, 0.016745471954345705, 0.016673791885375978, 0.016702463150024414, 0.016773120880126953, 0.016793600082397463, 0.016708608627319335, 0.016667648315429686, 0.016785408020019533, 0.016761856079101564, 0.016710655212402344, 0.01679871940612793, 0.016683008193969725, 0.016667648315429686, 0.016684032440185546, 0.016709632873535156]",tokens/s,59.10185484380597,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,8,8,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/8/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66932426-661af2c02e34843438d059ad;1ae59954-e19a-4d2f-a3a2-46a5e0f5e1da) Repository Not Found for url: https://huggingface.co/8/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: 8 is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-4B,,cuda,0,42,,,,,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,217063f5c507ed7cc255df7e1f64c4333a0b4dfe,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,,MB,1343.524864,10129.768448,0.0,9483.321344,8972.087296,s,10,0.49333631896972663,0.04933363189697266,0.00032257549326324613,0.04923403167724609,0.049390260314941405,0.04984034614562988,0.050200414810180664,"[0.050290431976318356, 0.04927091217041016, 0.04929024124145508, 0.04923040008544922, 0.0491343994140625, 0.049154014587402345, 0.04923766326904297, 0.0492259521484375, 0.04928464126586914, 0.049217662811279296]",tokens/s,5189.157784584462,kWh,5.83080105691512e-07,3.1943707746429787e-07,3.332290569497561e-06,4.2348077526533706e-06,tokens/kWh,60451386.45068364,MB,1343.524864,10129.768448,0.0,9483.321344,9040.66816,s,10,21.933081787109376,2.193308178710937,0.011571171743490533,2.1954342041015624,2.2031998046875,2.2083846923828125,2.2125326025390626,"[2.202047607421875, 2.19496923828125, 2.197378173828125, 2.171572998046875, 2.17970849609375, 2.213569580078125, 2.1831435546875, 2.193405029296875, 2.195899169921875, 2.201387939453125]",tokens/s,28.72373367842302,kWh,2.5757745418903538e-05,1.4115700914303688e-05,7.932024777470472e-05,0.00011919369410791192,tokens/kWh,528551.4512450885,,s,629,22.33951544952392,0.03551592281323358,0.005946273737480436,0.03462656021118164,0.03552870559692383,0.03584266357421875,0.08417665008544922,"[0.035593215942382815, 0.03451801681518555, 0.03449343872070312, 0.03544678497314453, 0.03440435028076172, 0.03566080093383789, 0.035915775299072264, 0.0349194221496582, 0.0354969596862793, 0.03550207901000976, 0.03477503967285156, 0.03448934555053711, 0.03507814407348633, 0.034511905670166015, 0.039514080047607425, 0.03681075286865235, 0.0349378547668457, 0.03581644821166992, 0.035236862182617186, 0.03471462249755859, 0.035337215423583986, 0.034869247436523435, 0.03476172637939453, 0.03474943923950195, 0.03463270568847656, 0.03465625762939453, 0.0343818244934082, 0.03489996719360351, 0.03514572906494141, 0.03473104095458984, 0.03432137680053711, 0.03435520172119141, 0.034576385498046876, 0.03467059326171875, 0.03457331085205078, 0.03458867263793945, 0.0347740478515625, 0.03453744125366211, 0.034505729675292966, 0.03458867263793945, 0.034609153747558595, 0.03581644821166992, 0.03494099044799805, 0.034860992431640626, 0.035404800415039066, 0.03521331024169922, 0.03443711853027344, 0.03498495864868164, 0.03469728088378906, 0.034722751617431644, 0.03476070404052734, 0.034457599639892575, 0.03437055969238281, 0.03456409454345703, 0.034305023193359374, 0.03466035079956055, 0.034543617248535156, 0.03444736099243164, 0.03466547012329101, 0.034435073852539064, 0.03448115158081055, 0.03431935882568359, 0.08402022552490235, 0.034086910247802735, 0.034985984802246094, 0.034799617767333986, 0.034661376953125, 0.03434188842773438, 0.0380579833984375, 0.03575603103637695, 0.03441664123535156, 0.03440639877319336, 0.03444940948486328, 0.034374656677246096, 0.03500646209716797, 0.035337215423583986, 0.03461939239501953, 0.03466342544555664, 0.03450368118286133, 0.034530303955078126, 0.03505152130126953, 0.035122177124023435, 0.034651134490966795, 0.03464908981323242, 0.03449446487426758, 0.034743297576904295, 0.03458560180664062, 0.0345425910949707, 0.03464396667480469, 0.034667518615722655, 0.034661376953125, 0.03458969497680664, 0.03462246322631836, 0.03597926330566406, 0.035939327239990236, 0.035280895233154294, 0.03473929595947266, 0.035811233520507815, 0.034683902740478514, 0.03459481430053711, 0.03515801620483398, 0.03458355331420898, 0.03467571258544922, 0.034667518615722655, 0.03530547332763672, 0.035544063568115236, 0.035149822235107424, 0.03453440093994141, 0.0343818244934082, 0.03483955383300781, 0.03458047866821289, 0.03418828964233398, 0.03473715209960938, 0.03569868850708008, 0.03460812759399414, 0.03462249755859375, 0.03449750518798828, 0.03470131301879883, 0.03457331085205078, 0.034634750366210935, 0.03468288040161133, 0.03467264175415039, 0.034723838806152346, 0.03473715209960938, 0.0346951675415039, 0.08752845001220703, 0.0353546257019043, 0.03465932846069336, 0.03452620697021484, 0.03491635131835937, 0.03472793579101562, 0.03483443069458008, 0.03467673492431641, 0.03552870559692383, 0.03451289749145508, 0.03455692672729492, 0.034852863311767575, 0.0349378547668457, 0.03501465606689453, 0.034825214385986326, 0.034543617248535156, 0.03456921768188476, 0.034893856048583985, 0.03456099319458008, 0.03458051300048828, 0.03455382537841797, 0.03464191818237305, 0.03457535934448242, 0.03498086547851562, 0.035108863830566404, 0.0361451530456543, 0.035707904815673826, 0.03549184036254883, 0.03641753768920898, 0.03611545562744141, 0.035095550537109374, 0.03486310577392578, 0.0345272331237793, 0.034476032257080076, 0.03484979248046875, 0.034459648132324217, 0.03495219039916992, 0.03473305511474609, 0.036410430908203124, 0.03472889709472656, 0.03458150482177735, 0.034612224578857424, 0.034536449432373044, 0.034500606536865236, 0.034492416381835936, 0.034597888946533206, 0.034518047332763674, 0.03454256057739258, 0.034598911285400394, 0.03446886444091797, 0.03456921768188476, 0.03458969497680664, 0.03463372802734375, 0.03458662414550781, 0.03458355331420898, 0.03465727996826172, 0.034508800506591795, 0.03451903915405274, 0.034587646484375, 0.03454054260253906, 0.03480166244506836, 0.03470438385009766, 0.034533374786376955, 0.08428851318359375, 0.034513919830322266, 0.03434598541259765, 0.03455385589599609, 0.034546688079833986, 0.03459993743896484, 0.03461734390258789, 0.03461529541015625, 0.034541568756103515, 0.0345354232788086, 0.034563072204589845, 0.03448934555053711, 0.034563072204589845, 0.03444736099243164, 0.03452928161621094, 0.034392097473144534, 0.034123744964599606, 0.0344637451171875, 0.03425484848022461, 0.03409612655639648, 0.0340582389831543, 0.03453440093994141, 0.034460670471191404, 0.034533374786376955, 0.03483955383300781, 0.03452620697021484, 0.034549758911132815, 0.03448831939697266, 0.03451289749145508, 0.03446169662475586, 0.034561023712158204, 0.03457228851318359, 0.03458150482177735, 0.03424460983276367, 0.03450982284545898, 0.03445043182373047, 0.03447500610351562, 0.03424256134033203, 0.033995777130126956, 0.034520065307617184, 0.034435073852539064, 0.03439206314086914, 0.034533374786376955, 0.03441561508178711, 0.034541568756103515, 0.03446681594848633, 0.034444286346435544, 0.03449343872070312, 0.03446783828735352, 0.034446334838867186, 0.034476032257080076, 0.03427840042114258, 0.03433574295043945, 0.034435073852539064, 0.0344453125, 0.034543617248535156, 0.0350013427734375, 0.034546688079833986, 0.03448320007324219, 0.03446886444091797, 0.034162689208984375, 0.034412574768066403, 0.03440943908691406, 0.0841861114501953, 0.03444940948486328, 0.03454771041870117, 0.03445657730102539, 0.034490367889404294, 0.034440193176269535, 0.034457599639892575, 0.03450470352172851, 0.03451289749145508, 0.034408447265625, 0.034536449432373044, 0.03403366470336914, 0.03406335830688476, 0.034067455291748046, 0.03418828964233398, 0.0345425910949707, 0.03428147125244141, 0.03407769775390625, 0.0345456657409668, 0.03446681594848633, 0.03448012924194336, 0.034408447265625, 0.03455078506469727, 0.03450470352172851, 0.034579456329345705, 0.034411518096923825, 0.03460505676269531, 0.034477054595947264, 0.034321407318115234, 0.03448320007324219, 0.03433478546142578, 0.03471148681640625, 0.0345354232788086, 0.035743743896484374, 0.03502284622192383, 0.03410432052612305, 0.034936832427978515, 0.03556556701660156, 0.03497369766235352, 0.03534643173217773, 0.034551807403564457, 0.0351016960144043, 0.03497983932495117, 0.03479859161376953, 0.034981952667236325, 0.035052478790283205, 0.03410124969482422, 0.03450982284545898, 0.03480575942993164, 0.03414220809936523, 0.035046398162841795, 0.035383296966552735, 0.03476684951782227, 0.03471462249755859, 0.0342999038696289, 0.03421184158325195, 0.034592769622802735, 0.03460812759399414, 0.03459993743896484, 0.034716670989990234, 0.03440639877319336, 0.034543617248535156, 0.034552833557128904, 0.0844400634765625, 0.034612224578857424, 0.034563072204589845, 0.034735103607177735, 0.03449139022827148, 0.03462246322631836, 0.03480166244506836, 0.03468492889404297, 0.03527475357055664, 0.03551027297973633, 0.03468492889404297, 0.03460812759399414, 0.03545804977416992, 0.03458560180664062, 0.03490508651733398, 0.035590145111083986, 0.034699264526367186, 0.03573452758789063, 0.035659774780273434, 0.034595840454101565, 0.03486412811279297, 0.03518873596191406, 0.03446169662475586, 0.035105792999267575, 0.0353546257019043, 0.03453440093994141, 0.03528192138671875, 0.03477503967285156, 0.03451596832275391, 0.03505766296386719, 0.03507199859619141, 0.03464191818237305, 0.03544473648071289, 0.03501363372802734, 0.03498700714111328, 0.03552870559692383, 0.03527577590942383, 0.03804159927368164, 0.03605299377441406, 0.03465727996826172, 0.03470438385009766, 0.03489689636230469, 0.03514470291137695, 0.0362342414855957, 0.03530035018920898, 0.0354856948852539, 0.03542015838623047, 0.034977790832519534, 0.03488774490356445, 0.035668926239013674, 0.03586457443237305, 0.034651134490966795, 0.03555737686157227, 0.035286014556884765, 0.03457334518432617, 0.03527779388427734, 0.03460204696655273, 0.034667457580566406, 0.03576319885253906, 0.0345425910949707, 0.03542425537109375, 0.036468734741210936, 0.03526860809326172, 0.08489881896972656, 0.034646080017089846, 0.034652095794677734, 0.034767871856689454, 0.03470745468139649, 0.03449446487426758, 0.0345241584777832, 0.03461119842529297, 0.03459481430053711, 0.03488358306884766, 0.034651134490966795, 0.03456921768188476, 0.034582527160644534, 0.03451596832275391, 0.03456819152832031, 0.03461119842529297, 0.034508800506591795, 0.034446334838867186, 0.034579456329345705, 0.034492416381835936, 0.03456204986572266, 0.03448934555053711, 0.03447091293334961, 0.03457024002075195, 0.03433779144287109, 0.03420159912109375, 0.03451903915405274, 0.034484222412109376, 0.03459686279296875, 0.0342108154296875, 0.03591372680664062, 0.03594854354858398, 0.03474124908447265, 0.034520095825195315, 0.03459171295166016, 0.03450368118286133, 0.034651134490966795, 0.03466649627685547, 0.03500032043457031, 0.03480473709106445, 0.03459379196166992, 0.034536449432373044, 0.03458047866821289, 0.03459993743896484, 0.03450982284545898, 0.034525184631347655, 0.034625537872314455, 0.03464601516723633, 0.03461939239501953, 0.03504435348510742, 0.03470131301879883, 0.034574337005615234, 0.034595840454101565, 0.03452620697021484, 0.034648063659667966, 0.03462656021118164, 0.034619457244873045, 0.03469100952148438, 0.03506687927246094, 0.03470336151123047, 0.0343900146484375, 0.03447091293334961, 0.03446476745605469, 0.08421171569824219, 0.035143680572509765, 0.03488870239257812, 0.03462348937988281, 0.034597888946533206, 0.03451289749145508, 0.03451801681518555, 0.034662464141845706, 0.03449132919311523, 0.03444736099243164, 0.03463372802734375, 0.03528192138671875, 0.03537612915039062, 0.03448320007324219, 0.034563072204589845, 0.03458457565307617, 0.03458457565307617, 0.03446579360961914, 0.034530303955078126, 0.034552833557128904, 0.034459648132324217, 0.03446681594848633, 0.03439513778686523, 0.034598911285400394, 0.034508800506591795, 0.034541568756103515, 0.03629363250732422, 0.03503206253051758, 0.03460300827026367, 0.03460710525512695, 0.034726913452148435, 0.035778560638427735, 0.035727359771728515, 0.03448627090454102, 0.03455897521972656, 0.03449958419799805, 0.03452928161621094, 0.03445657730102539, 0.034223102569580076, 0.03506687927246094, 0.035399681091308595, 0.03454057693481445, 0.03474019241333008, 0.034531326293945314, 0.035436607360839846, 0.034614273071289066, 0.034686912536621095, 0.03474534225463867, 0.035388416290283206, 0.034786304473876956, 0.03473715209960938, 0.03606118392944336, 0.03520614242553711, 0.03446886444091797, 0.03477811050415039, 0.03470848083496094, 0.034614273071289066, 0.03574784088134766, 0.034614273071289066, 0.03457024002075195, 0.035448833465576174, 0.035490848541259765, 0.034625503540039064, 0.08415232086181641, 0.03524915313720703, 0.03497267150878906, 0.034590721130371094, 0.03458560180664062, 0.03473715209960938, 0.035471359252929685, 0.03520307159423828, 0.034644992828369144, 0.03461017608642578, 0.03552870559692383, 0.03443711853027344, 0.03417292785644531, 0.03449139022827148, 0.03503206253051758, 0.035146751403808595, 0.03458047866821289, 0.03451801681518555, 0.03465727996826172, 0.03471462249755859, 0.034511871337890625, 0.03450163269042969, 0.03460095977783203, 0.03477503967285156, 0.03455692672729492, 0.035829761505126956, 0.034958335876464845, 0.03480883026123047, 0.03585126495361328, 0.0345272331237793, 0.034566143035888675, 0.03462144088745117, 0.03556147384643555, 0.03504844665527344, 0.0342210578918457, 0.034561023712158204, 0.034938880920410156, 0.034770942687988284, 0.03450982284545898, 0.03460403060913086, 0.03447091293334961, 0.035487743377685545, 0.034974720001220705, 0.034506752014160154, 0.035372032165527346, 0.035269630432128905, 0.034887680053710936, 0.03459379196166992, 0.034511871337890625, 0.03486515045166016, 0.03553792190551758, 0.03486310577392578, 0.03460505676269531, 0.03499827194213867, 0.035697662353515625, 0.03479347229003906, 0.03491635131835937, 0.03500646209716797, 0.03512319946289062, 0.035133438110351564, 0.03456716918945312, 0.03460095977783203, 0.03456512069702149, 0.08439091491699219, 0.03448320007324219, 0.03466649627685547, 0.03548262405395508, 0.034546688079833986, 0.03461017608642578, 0.03456003189086914, 0.034911201477050784, 0.03470745468139649, 0.03415961456298828, 0.03463372802734375, 0.034816001892089846, 0.03465216064453125, 0.03457235336303711, 0.034595840454101565, 0.03447391891479492, 0.034909183502197266, 0.03451289749145508, 0.035757057189941405, 0.03516928100585937, 0.034994174957275394, 0.035410945892333984, 0.03464396667480469, 0.03463065719604492, 0.03471769714355469, 0.03561369705200195, 0.03503923034667969, 0.03460710525512695, 0.03520409774780273, 0.035740673065185545, 0.035130367279052735, 0.03466854476928711, 0.03464191818237305, 0.03486617660522461, 0.03491123199462891, 0.034549758911132815, 0.03455897521972656, 0.036083713531494144, 0.036354049682617184, 0.03476070404052734, 0.035378177642822264, 0.03466239929199219, 0.0353331184387207, 0.034909183502197266, 0.035040256500244144, 0.03470745468139649, 0.03530342483520508, 0.03589427185058594, 0.03476684951782227, 0.03482828903198242, 0.03483135986328125, 0.03563827133178711, 0.03473715209960938, 0.034748416900634765, 0.03461734390258789, 0.035211265563964846, 0.03510374450683594, 0.0346060791015625, 0.034628608703613284, 0.0348221435546875, 0.035714046478271484, 0.034732032775878906, 0.03469209671020508]",tokens/s,28.15638510249803,,,main,False,False,,,,,,,,,,,,,,,,,,,,,,, float16-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,internlm/internlm2-20b,internlm/internlm2-20b,cuda,0,42,,,True,,,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",internlm2,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,i,i,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/i/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669474ad-2664787d2fdd5bcc60a24be0;bf576584-f03a-4bf0-a2ee-f49d9441bbd7) Repository Not Found for url: https://huggingface.co/i/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: i is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Deci/DeciLM-7B,Deci/DeciLM-7B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1278.701568,14990.966784,0.0,14344.51968,14290.684928,s,10,0.8070263671874999,0.08070263671875001,0.0003132112196133856,0.0805906105041504,0.08091112289428712,0.08123637657165528,0.0814965795135498,"[0.08156163024902344, 0.08083884429931641, 0.0807853775024414, 0.08059661102294922, 0.08069852447509765, 0.08058460998535157, 0.08046800231933594, 0.08049507141113281, 0.08042348480224609, 0.08057421112060546]",tokens/s,3172.1392312392977,kWh,9.553098256656346e-07,5.234670793910363e-07,6.397496157419333e-06,7.876273062476004e-06,tokens/kWh,32502682.165709887,MB,1278.701568,14990.966784,0.0,14344.51968,14290.687488,s,10,22.68975317382813,2.268975317382813,0.003229486566520809,2.2680010986328125,2.269869921875,2.2741899658203124,2.2776460009765627,"[2.278510009765625, 2.2673662109375, 2.26838037109375, 2.267602294921875, 2.268310546875, 2.267691650390625, 2.268909912109375, 2.26850146484375, 2.26757080078125, 2.266909912109375]",tokens/s,27.765837520290166,kWh,2.6725956260303448e-05,1.4646700880122454e-05,0.0001179570755483812,0.00015932973268880708,tokens/kWh,395406.4250082417,,s,629,23.378355113983176,0.03716749620665048,0.009633515062260284,0.03598233413696289,0.036061388397216794,0.03619123153686524,0.11699908477783202,"[0.036760513305664065, 0.03685887908935547, 0.03673395156860351, 0.03667967987060547, 0.03672063827514648, 0.03674214553833008, 0.036770816802978515, 0.036803585052490234, 0.036749313354492184, 0.036498432159423826, 0.03635200119018555, 0.03651891326904297, 0.03659980773925781, 0.036193279266357424, 0.03606528091430664, 0.036018177032470705, 0.03598745727539063, 0.035985408782958986, 0.035996673583984375, 0.03600691223144531, 0.03600384140014649, 0.035999744415283204, 0.03598745727539063, 0.03600691223144531, 0.03599462509155273, 0.03600691223144531, 0.035996673583984375, 0.03606528091430664, 0.03597619247436523, 0.03599257659912109, 0.03599564743041992, 0.035970046997070314, 0.035980289459228515, 0.03600691223144531, 0.03605299377441406, 0.03600998306274414, 0.03603148651123047, 0.03603664016723633, 0.035977184295654295, 0.035970046997070314, 0.03598233413696289, 0.036001792907714845, 0.035975166320800785, 0.03601715087890625, 0.03604582214355469, 0.03598233413696289, 0.03600998306274414, 0.03599871826171875, 0.03598950576782227, 0.036036609649658206, 0.03601408004760742, 0.036001792907714845, 0.03603558349609375, 0.03600281524658203, 0.03598444747924805, 0.036028350830078125, 0.03599257659912109, 0.03602841567993164, 0.03605913543701172, 0.036038654327392575, 0.036021247863769534, 0.03599871826171875, 0.11729510498046875, 0.035953662872314454, 0.03593523025512695, 0.03592192077636719, 0.03594137573242188, 0.035947521209716796, 0.035955711364746096, 0.03599769592285156, 0.035934207916259765, 0.035961856842041014, 0.03593011093139648, 0.03596492767333984, 0.0359659538269043, 0.03596492767333984, 0.03598233413696289, 0.035955711364746096, 0.035955711364746096, 0.0359444465637207, 0.035983360290527344, 0.035942432403564456, 0.03593929672241211, 0.035937278747558594, 0.03597721481323242, 0.0359741439819336, 0.03597107315063477, 0.03597721481323242, 0.03595673751831055, 0.03600691223144531, 0.03597824096679687, 0.0359741439819336, 0.036038654327392575, 0.03604377746582031, 0.03598233413696289, 0.0359659538269043, 0.035955711364746096, 0.035970046997070314, 0.035998783111572265, 0.036011966705322265, 0.036001792907714845, 0.0359741439819336, 0.035983360290527344, 0.036018177032470705, 0.03599769592285156, 0.035975166320800785, 0.036004863739013675, 0.03600998306274414, 0.03604582214355469, 0.035980289459228515, 0.03603558349609375, 0.03599052810668945, 0.03597724914550781, 0.035993568420410155, 0.035983360290527344, 0.03599871826171875, 0.03602739334106445, 0.035985408782958986, 0.03601408004760742, 0.03604275131225586, 0.036078590393066406, 0.03600281524658203, 0.03604172897338867, 0.0359813117980957, 0.036001792907714845, 0.11726239776611327, 0.035926048278808596, 0.03591363143920898, 0.03597619247436523, 0.03595673751831055, 0.035999744415283204, 0.03618815994262695, 0.03611852645874023, 0.0359444465637207, 0.03593830490112305, 0.035945472717285154, 0.03595673751831055, 0.03598745727539063, 0.035929088592529294, 0.035961856842041014, 0.03596083068847656, 0.03604377746582031, 0.03605606460571289, 0.03596083068847656, 0.036004863739013675, 0.03608473587036133, 0.03600281524658203, 0.035939327239990236, 0.03599257659912109, 0.03595161437988281, 0.03597312164306641, 0.03603046417236328, 0.03597318267822266, 0.035980224609375, 0.035969024658203126, 0.03605913543701172, 0.036029441833496094, 0.03602534484863281, 0.03599769592285156, 0.036040702819824216, 0.03611852645874023, 0.03606220626831055, 0.03601100921630859, 0.03600384140014649, 0.03599564743041992, 0.03594649505615234, 0.035937278747558594, 0.035931198120117185, 0.035992511749267576, 0.035988479614257815, 0.03602227020263672, 0.035934207916259765, 0.036038654327392575, 0.03612876892089844, 0.03608575820922852, 0.036176895141601564, 0.03604275131225586, 0.035983360290527344, 0.03599052810668945, 0.03596908950805664, 0.03594438552856445, 0.035952640533447267, 0.03599359893798828, 0.03595980834960937, 0.03595980834960937, 0.035980289459228515, 0.03599257659912109, 0.035970046997070314, 0.11704524993896484, 0.0359444465637207, 0.03593318557739258, 0.03591884613037109, 0.03595475387573242, 0.03592902374267578, 0.03600076675415039, 0.036034561157226565, 0.035953662872314454, 0.03593523025512695, 0.035969024658203126, 0.03597721481323242, 0.03597721481323242, 0.03595673751831055, 0.035963905334472655, 0.03595468902587891, 0.03594035339355469, 0.035975166320800785, 0.035947521209716796, 0.03595775985717774, 0.03595673751831055, 0.035945472717285154, 0.035932159423828124, 0.03598438262939453, 0.0359444465637207, 0.03595468902587891, 0.03599359893798828, 0.035953662872314454, 0.03602431869506836, 0.0359741439819336, 0.036018177032470705, 0.03595161437988281, 0.03593427276611328, 0.03611641693115234, 0.03644825744628906, 0.03607449722290039, 0.035952640533447267, 0.035958782196044925, 0.03604582214355469, 0.036055038452148434, 0.03594144058227539, 0.036054977416992186, 0.036201473236083984, 0.03597619247436523, 0.03601408004760742, 0.035969024658203126, 0.0359741439819336, 0.03599052810668945, 0.035975200653076175, 0.035958751678466794, 0.0359659538269043, 0.036001792907714845, 0.035958782196044925, 0.03598748779296875, 0.03597820663452148, 0.03597824096679687, 0.03597824096679687, 0.03597926330566406, 0.0359741439819336, 0.03595775985717774, 0.035958782196044925, 0.03600384140014649, 0.0359741439819336, 0.11741388702392579, 0.03613798522949219, 0.035953662872314454, 0.035983360290527344, 0.03594035339355469, 0.03595673751831055, 0.03595673751831055, 0.03600998306274414, 0.035931137084960936, 0.035963905334472655, 0.03599462509155273, 0.035996673583984375, 0.03599871826171875, 0.03597824096679687, 0.035983360290527344, 0.03597619247436523, 0.03599052810668945, 0.03599871826171875, 0.0359741439819336, 0.03605299377441406, 0.0361082878112793, 0.03601919937133789, 0.035947521209716796, 0.035975166320800785, 0.03598438262939453, 0.035969024658203126, 0.03600281524658203, 0.03597318267822266, 0.035979198455810546, 0.03600384140014649, 0.03599564743041992, 0.03592704010009766, 0.03598438262939453, 0.035969024658203126, 0.03599359893798828, 0.03597721481323242, 0.036086784362792966, 0.03616665649414062, 0.03598438262939453, 0.03596083068847656, 0.035996673583984375, 0.035980289459228515, 0.035962944030761716, 0.03598844909667969, 0.03597206497192383, 0.03601715087890625, 0.035983360290527344, 0.03597926330566406, 0.03599359893798828, 0.03599359893798828, 0.03599871826171875, 0.0360079345703125, 0.036040702819824216, 0.03599052810668945, 0.03598745727539063, 0.03600281524658203, 0.035983360290527344, 0.03600076675415039, 0.03597619247436523, 0.03597619247436523, 0.035996673583984375, 0.03601919937133789, 0.03598745727539063, 0.11711385345458984, 0.035985408782958986, 0.03596492767333984, 0.03595161437988281, 0.03600076675415039, 0.03598643112182617, 0.03601715087890625, 0.03598233413696289, 0.035972095489501955, 0.03598233413696289, 0.03599462509155273, 0.035970046997070314, 0.0359516487121582, 0.035963871002197265, 0.03599359893798828, 0.035966976165771485, 0.03595980834960937, 0.0359628791809082, 0.035966976165771485, 0.03599359893798828, 0.03598643112182617, 0.03596799850463867, 0.03595673751831055, 0.03595673751831055, 0.035966976165771485, 0.03600076675415039, 0.03599257659912109, 0.03595468902587891, 0.035980289459228515, 0.035988479614257815, 0.035969024658203126, 0.03604172897338867, 0.03599564743041992, 0.036002849578857424, 0.03598742294311524, 0.035972095489501955, 0.0359813117980957, 0.03597926330566406, 0.03600281524658203, 0.03597824096679687, 0.0359741439819336, 0.03599257659912109, 0.035975166320800785, 0.035975166320800785, 0.03597824096679687, 0.03597619247436523, 0.036001792907714845, 0.03603353500366211, 0.03598233413696289, 0.03600691223144531, 0.036013057708740234, 0.03601715087890625, 0.035996673583984375, 0.03599871826171875, 0.03601715087890625, 0.036016128540039063, 0.035996673583984375, 0.03598745727539063, 0.03600281524658203, 0.03603251266479492, 0.035988479614257815, 0.03601824188232422, 0.0359884147644043, 0.11697254180908204, 0.03599359893798828, 0.03594342422485351, 0.035942401885986325, 0.035934207916259765, 0.03595980834960937, 0.03596492767333984, 0.035963905334472655, 0.035915775299072264, 0.03599155044555664, 0.036034561157226565, 0.035985408782958986, 0.03598745727539063, 0.035980289459228515, 0.03605606460571289, 0.036187137603759766, 0.0362342414855957, 0.03601100921630859, 0.035970046997070314, 0.036111358642578126, 0.035955711364746096, 0.03602022552490235, 0.03611340713500977, 0.036311038970947264, 0.03607244873046875, 0.0359813117980957, 0.03597619247436523, 0.03596499252319336, 0.035957695007324215, 0.035961856842041014, 0.03597619247436523, 0.03595673751831055, 0.035958782196044925, 0.03599564743041992, 0.03597107315063477, 0.0359741439819336, 0.03599769592285156, 0.03602227020263672, 0.03603968048095703, 0.03596799850463867, 0.03597107315063477, 0.03595775985717774, 0.03599871826171875, 0.03601718521118164, 0.03596796798706055, 0.03600384140014649, 0.03596700668334961, 0.03600073623657227, 0.03603148651123047, 0.03602431869506836, 0.0360079345703125, 0.035988479614257815, 0.03599155044555664, 0.0360263671875, 0.035972095489501955, 0.036029441833496094, 0.03608473587036133, 0.03603251266479492, 0.03606118392944336, 0.03597721481323242, 0.03599052810668945, 0.03601203155517578, 0.03596799850463867, 0.11700940704345703, 0.03597824096679687, 0.03598438262939453, 0.03592704010009766, 0.03604991912841797, 0.03594854354858398, 0.035912704467773435, 0.03593625640869141, 0.03592601776123047, 0.035958782196044925, 0.03597318267822266, 0.035937217712402346, 0.035975166320800785, 0.03594956970214844, 0.035931137084960936, 0.03594854354858398, 0.03597619247436523, 0.03596799850463867, 0.03595980834960937, 0.035955711364746096, 0.03597312164306641, 0.035952640533447267, 0.0359444465637207, 0.035999744415283204, 0.03599462509155273, 0.0359813117980957, 0.035980350494384764, 0.035940288543701175, 0.035986465454101564, 0.035968990325927735, 0.03595468902587891, 0.03595161437988281, 0.0359659538269043, 0.035999744415283204, 0.03600691223144531, 0.03598643112182617, 0.0359813117980957, 0.035980289459228515, 0.03596799850463867, 0.03600998306274414, 0.036004863739013675, 0.03600384140014649, 0.03600281524658203, 0.03598950576782227, 0.03601100921630859, 0.03597721481323242, 0.03600076675415039, 0.03598950576782227, 0.03598438262939453, 0.03600998306274414, 0.0361082878112793, 0.03633356857299805, 0.03619430541992188, 0.03598950576782227, 0.036070400238037106, 0.03597926330566406, 0.03603046417236328, 0.035975166320800785, 0.036013057708740234, 0.036560897827148435, 0.036103168487548826, 0.03600588989257812, 0.035975166320800785, 0.11688243103027343, 0.0359444465637207, 0.035939327239990236, 0.03594137573242188, 0.035914752960205076, 0.03591987228393555, 0.035934207916259765, 0.03592499160766602, 0.035917823791503906, 0.03594956970214844, 0.035966976165771485, 0.035952640533447267, 0.035985408782958986, 0.03598233413696289, 0.036038654327392575, 0.0361973762512207, 0.03616767883300781, 0.03604172897338867, 0.03603968048095703, 0.036195327758789066, 0.03599769592285156, 0.03599359893798828, 0.035947521209716796, 0.035952640533447267, 0.0359444465637207, 0.03594649505615234, 0.0359741439819336, 0.035972095489501955, 0.035963905334472655, 0.035947521209716796, 0.03599052810668945, 0.0359813117980957, 0.03595673751831055, 0.035970046997070314, 0.035945472717285154, 0.03594854354858398, 0.03599871826171875, 0.03603353500366211, 0.03598950576782227, 0.035988479614257815, 0.035953662872314454, 0.035953662872314454, 0.035952640533447267, 0.035963905334472655, 0.035963905334472655, 0.03602534484863281, 0.03601919937133789, 0.0359741439819336, 0.036111358642578126, 0.036038654327392575, 0.035980289459228515, 0.035969024658203126, 0.03598643112182617, 0.035958782196044925, 0.03597721481323242, 0.03598233413696289, 0.035958782196044925, 0.035975166320800785, 0.03599359893798828, 0.03597926330566406, 0.03597824096679687, 0.035988479614257815, 0.03604684829711914, 0.11707801818847656, 0.03591884613037109, 0.03607449722290039, 0.03597107315063477, 0.03597312164306641, 0.03602431869506836, 0.03591987228393555, 0.03594342422485351, 0.0359598388671875, 0.03597206497192383, 0.035937278747558594, 0.035914752960205076, 0.03594956970214844, 0.035942401885986325, 0.035931137084960936, 0.0359444465637207, 0.03592499160766602, 0.035939327239990236, 0.035953662872314454, 0.03596799850463867, 0.03602739334106445, 0.03597824096679687, 0.035958782196044925, 0.035972095489501955, 0.03594342422485351, 0.03598031997680664, 0.036061153411865235, 0.036073471069335936, 0.03594342422485351, 0.03595468902587891, 0.03597721481323242, 0.036001792907714845, 0.035961856842041014, 0.035966976165771485, 0.03595980834960937, 0.035963905334472655, 0.03600384140014649, 0.03599871826171875, 0.03596083068847656, 0.03595673751831055, 0.03592704010009766, 0.035985408782958986, 0.0359659538269043, 0.03594956970214844, 0.035963905334472655, 0.03594649505615234, 0.03598233413696289, 0.03597619247436523, 0.03596806335449219, 0.035995582580566406, 0.0359628791809082, 0.03595161437988281, 0.03594035339355469, 0.03595673751831055, 0.035966976165771485, 0.035985408782958986, 0.03598233413696289, 0.035970046997070314, 0.03604172897338867, 0.03613081741333008, 0.036089855194091795, 0.03600998306274414, 0.035996673583984375]",tokens/s,26.90522908618922,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,stabilityai/stablelm-3b-4e1t,stabilityai/stablelm-3b-4e1t,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1271.402496,6715.604992,0.0,6069.157888,5763.863552,s,10,0.3767350730895996,0.03767350730895996,0.0002823909925832922,0.03758452796936035,0.0377231086730957,0.038119553756713866,0.038436709823608396,"[0.03851599884033203, 0.037617759704589845, 0.03753292846679687, 0.03753919982910156, 0.03757577514648437, 0.03758752059936524, 0.037561824798583984, 0.03758687973022461, 0.03758217620849609, 0.037635009765625]",tokens/s,6795.226096167454,kWh,4.456816485592904e-07,2.442112888758925e-07,2.6355387960603813e-06,3.3254317334955644e-06,tokens/kWh,76982485.43833518,MB,1271.402496,6736.576512,0.0,6090.129408,5863.102976,s,10,16.57344763183594,1.657344763183594,0.022532741874149394,1.6548178100585937,1.6739061035156249,1.6943570434570312,1.7107177954101562,"[1.7148079833984375, 1.6353643798828126, 1.659327392578125, 1.6373353271484374, 1.6441790771484375, 1.6629339599609374, 1.636151123046875, 1.6503082275390626, 1.6693614501953125, 1.6636787109375]",tokens/s,38.0126099285361,kWh,1.9196905523101242e-05,1.0520058894677998e-05,6.260756895393975e-05,9.232453337171899e-05,tokens/kWh,682375.5040964904,,s,629,16.88425879478455,0.02684301875164475,0.004532659493577283,0.025988096237182616,0.02702561264038086,0.027627724075317384,0.06399864852905274,"[0.027968576431274414, 0.027725759506225585, 0.02796031951904297, 0.028034048080444338, 0.02771865653991699, 0.027649023056030272, 0.02793267250061035, 0.027732992172241212, 0.0274913272857666, 0.02797987174987793, 0.029194143295288084, 0.027807743072509765, 0.027992063522338868, 0.027846656799316406, 0.02779545593261719, 0.027686912536621092, 0.027472896575927733, 0.027666431427001953, 0.02775449562072754, 0.027056127548217773, 0.02697420883178711, 0.02698956871032715, 0.026990591049194337, 0.02695680046081543, 0.02711039924621582, 0.027052032470703126, 0.026900480270385742, 0.02688102340698242, 0.026916864395141602, 0.026908672332763672, 0.026895360946655275, 0.02695680046081543, 0.027114496231079102, 0.027016191482543944, 0.026874944686889647, 0.026910655975341796, 0.027064319610595702, 0.02692198371887207, 0.027107328414916993, 0.027044864654541017, 0.027026432037353516, 0.02692300796508789, 0.026894336700439454, 0.027009023666381835, 0.026864639282226564, 0.028201984405517577, 0.02750668716430664, 0.02694655990600586, 0.025955327987670897, 0.02632908821105957, 0.02693120002746582, 0.027032575607299804, 0.026998783111572267, 0.02691481590270996, 0.02706329536437988, 0.02690764808654785, 0.026985471725463867, 0.027116544723510744, 0.025988096237182616, 0.025851903915405275, 0.02666908836364746, 0.026903520584106444, 0.06462464141845703, 0.025882623672485353, 0.02574950408935547, 0.02591948890686035, 0.025874431610107423, 0.02588057518005371, 0.025881599426269532, 0.02596249580383301, 0.025868288040161135, 0.025778175354003906, 0.02591334342956543, 0.025818111419677735, 0.025840639114379883, 0.025821184158325194, 0.025860095977783205, 0.0259102725982666, 0.02588467216491699, 0.025806880950927733, 0.025829343795776366, 0.025867263793945314, 0.025975807189941406, 0.02592870330810547, 0.025817087173461914, 0.025989120483398437, 0.025814016342163085, 0.025985023498535157, 0.02594304084777832, 0.025891839981079103, 0.0259051513671875, 0.02572390365600586, 0.02576383972167969, 0.02559187126159668, 0.025595840454101564, 0.025545759201049803, 0.02561020851135254, 0.02561846351623535, 0.02555392074584961, 0.025678815841674803, 0.025529375076293947, 0.025639904022216796, 0.02554982376098633, 0.025730112075805663, 0.02565830421447754, 0.0257126407623291, 0.025564159393310547, 0.02556729507446289, 0.02561529541015625, 0.025613311767578126, 0.025498815536499023, 0.025694015502929688, 0.026619903564453123, 0.026430591583251953, 0.0265665283203125, 0.02588979148864746, 0.026480640411376953, 0.026549247741699217, 0.026615808486938477, 0.026550271987915038, 0.026704896926879884, 0.02684316825866699, 0.026659936904907228, 0.026874752044677735, 0.02696703910827637, 0.0641075210571289, 0.026983423233032225, 0.027025407791137695, 0.026806272506713868, 0.02671001625061035, 0.026846208572387696, 0.026727424621582032, 0.025977855682373048, 0.025824256896972656, 0.02588467216491699, 0.02587750434875488, 0.02594099235534668, 0.025784320831298828, 0.02591846466064453, 0.025878528594970703, 0.025886720657348632, 0.025665536880493164, 0.025854976654052734, 0.025673728942871094, 0.02589798355102539, 0.025851903915405275, 0.025753599166870117, 0.02591846466064453, 0.025952255249023438, 0.0259102725982666, 0.025771007537841797, 0.026149887084960938, 0.026426368713378907, 0.02610688018798828, 0.026050559997558592, 0.026875904083251953, 0.026936319351196288, 0.027869184494018553, 0.026848255157470705, 0.0267458553314209, 0.02696294403076172, 0.025861120223999022, 0.025894912719726562, 0.02591436767578125, 0.026895360946655275, 0.02710527992248535, 0.02696396827697754, 0.02669158363342285, 0.02691689682006836, 0.026758111953735352, 0.026746879577636717, 0.02674995231628418, 0.026916864395141602, 0.026831872940063478, 0.026850303649902343, 0.026856447219848634, 0.02680012893676758, 0.026858495712280273, 0.0261079044342041, 0.026004480361938476, 0.025988096237182616, 0.026032127380371094, 0.02592153549194336, 0.02596659278869629, 0.02593894386291504, 0.026027008056640624, 0.02589695930480957, 0.025977855682373048, 0.06398463821411132, 0.02592460823059082, 0.025827327728271485, 0.025986047744750978, 0.025873407363891602, 0.025860095977783205, 0.025956352233886718, 0.02595840072631836, 0.0259368953704834, 0.025865215301513672, 0.026033184051513673, 0.026004447937011718, 0.026073087692260744, 0.026798080444335938, 0.026926080703735353, 0.025972736358642577, 0.025955327987670897, 0.025838592529296874, 0.0259102725982666, 0.02588876724243164, 0.026024959564208985, 0.02593484878540039, 0.025967615127563477, 0.025999359130859375, 0.02594304084777832, 0.0259051513671875, 0.025931776046752928, 0.025956352233886718, 0.02594304084777832, 0.02595123291015625, 0.025903104782104492, 0.025991167068481445, 0.02593280029296875, 0.025999391555786133, 0.025926624298095703, 0.025939968109130858, 0.025882623672485353, 0.026004480361938476, 0.026013696670532226, 0.026059776306152343, 0.025969663619995118, 0.02590105628967285, 0.02591744041442871, 0.025899007797241212, 0.026068992614746093, 0.026010623931884767, 0.02593894386291504, 0.026012672424316406, 0.02588876724243164, 0.02594611167907715, 0.025994239807128908, 0.025869312286376952, 0.025983999252319336, 0.026018815994262694, 0.025939968109130858, 0.02589286422729492, 0.026015743255615235, 0.026077184677124023, 0.02594099235534668, 0.025948160171508788, 0.02591641616821289, 0.025968639373779297, 0.026024959564208985, 0.06411366271972656, 0.025973760604858398, 0.025852928161621092, 0.025790464401245116, 0.025871360778808594, 0.025933855056762694, 0.025854944229125976, 0.02592665672302246, 0.025866239547729493, 0.025815040588378906, 0.027209728240966798, 0.026860544204711914, 0.026735616683959962, 0.026764287948608398, 0.02672332763671875, 0.02676940727233887, 0.026062847137451172, 0.025806848526000976, 0.025805824279785155, 0.025875455856323244, 0.025790464401245116, 0.025780223846435548, 0.025818111419677735, 0.025965568542480468, 0.02576486396789551, 0.02568294334411621, 0.026184703826904295, 0.02592563247680664, 0.025869375228881837, 0.02578835105895996, 0.025866239547729493, 0.025818111419677735, 0.025758720397949218, 0.025772031784057618, 0.025861120223999022, 0.025832447052001953, 0.026061824798583984, 0.025769983291625977, 0.02569523239135742, 0.025818111419677735, 0.02550169563293457, 0.025554943084716796, 0.025431039810180665, 0.025825279235839844, 0.025778175354003906, 0.025799680709838867, 0.025862144470214843, 0.02595123291015625, 0.025754623413085938, 0.025858047485351563, 0.025971712112426756, 0.025649152755737304, 0.02596147155761719, 0.02680012893676758, 0.026829824447631836, 0.026831872940063478, 0.027884544372558592, 0.027100160598754884, 0.026836992263793946, 0.026060800552368164, 0.02673766326904297, 0.026786815643310546, 0.02657587242126465, 0.06429593658447266, 0.025822208404541015, 0.026005504608154296, 0.025879552841186523, 0.02591436767578125, 0.02589388847351074, 0.025886720657348632, 0.026685440063476562, 0.02811392021179199, 0.027175935745239257, 0.027075584411621095, 0.026803199768066405, 0.02592255973815918, 0.02590105628967285, 0.02589798355102539, 0.025991167068481445, 0.025969663619995118, 0.025931776046752928, 0.02588876724243164, 0.025998336791992187, 0.02593382453918457, 0.02594304084777832, 0.025956352233886718, 0.025980928421020507, 0.0259420166015625, 0.025793535232543945, 0.02594611167907715, 0.025856000900268555, 0.025911296844482422, 0.02592563247680664, 0.02590208053588867, 0.025935871124267578, 0.025907199859619142, 0.025833471298217774, 0.025842687606811524, 0.02590617561340332, 0.02586419105529785, 0.02591334342956543, 0.02589286422729492, 0.027268096923828124, 0.02734489631652832, 0.026960895538330077, 0.02679091262817383, 0.026772480010986328, 0.026787839889526367, 0.026774560928344727, 0.026874847412109375, 0.02698137664794922, 0.026788864135742187, 0.026856447219848634, 0.026805248260498047, 0.026796031951904296, 0.026797056198120117, 0.026953727722167968, 0.02689740753173828, 0.02688102340698242, 0.026862592697143556, 0.02671308708190918, 0.026830848693847657, 0.02693734359741211, 0.026864639282226564, 0.026871807098388673, 0.026892288208007813, 0.0639447021484375, 0.025900096893310548, 0.025916351318359374, 0.02627276802062988, 0.028840959548950194, 0.02736025619506836, 0.02655232048034668, 0.02595840072631836, 0.025963520050048827, 0.025881599426269532, 0.025873407363891602, 0.025883647918701173, 0.025781280517578126, 0.025645023345947267, 0.02553753662109375, 0.025599008560180665, 0.025435104370117187, 0.025594879150390625, 0.025569280624389647, 0.025462783813476563, 0.0255416316986084, 0.025531391143798828, 0.02540345573425293, 0.025526208877563475, 0.02572287940979004, 0.025404415130615234, 0.025582592010498048, 0.025487360000610353, 0.025422847747802735, 0.02544646453857422, 0.025506752014160156, 0.025469951629638672, 0.025805824279785155, 0.02569113540649414, 0.02552422332763672, 0.02560111999511719, 0.025421728134155275, 0.025531391143798828, 0.02552934455871582, 0.025597951889038087, 0.02553548812866211, 0.02545254325866699, 0.02547711944580078, 0.028635135650634767, 0.027006975173950197, 0.026628095626831053, 0.02675200080871582, 0.025948160171508788, 0.025852928161621092, 0.025878528594970703, 0.02591641616821289, 0.0259420166015625, 0.02593484878540039, 0.02596147155761719, 0.025861120223999022, 0.025872447967529296, 0.025975744247436525, 0.02596147155761719, 0.026427391052246094, 0.02655129623413086, 0.026798080444335938, 0.026851327896118164, 0.026820608139038086, 0.06406348419189453, 0.02587648010253906, 0.025870367050170897, 0.025843679428100588, 0.025919551849365233, 0.026476480484008788, 0.026786815643310546, 0.02697216033935547, 0.02688102340698242, 0.02692300796508789, 0.026817535400390623, 0.02687385559082031, 0.026795007705688476, 0.026863616943359377, 0.026828800201416016, 0.026883071899414062, 0.026861568450927735, 0.026848255157470705, 0.026806272506713868, 0.026863616943359377, 0.026847232818603517, 0.026811391830444335, 0.026874879837036132, 0.02690355110168457, 0.025949216842651366, 0.026074079513549803, 0.02591846466064453, 0.02590924835205078, 0.025979904174804686, 0.02588057518005371, 0.02596147155761719, 0.025987071990966795, 0.02611712074279785, 0.025901088714599608, 0.025650144577026367, 0.025629695892333985, 0.02552627182006836, 0.025705472946166992, 0.025833471298217774, 0.025969663619995118, 0.026031103134155274, 0.025804800033569338, 0.025857023239135742, 0.0259102725982666, 0.02590822410583496, 0.025870399475097658, 0.025679807662963867, 0.025750528335571288, 0.02546892738342285, 0.025419776916503906, 0.025588735580444336, 0.025843711853027345, 0.025869312286376952, 0.02588979148864746, 0.026036224365234374, 0.02594611167907715, 0.025834495544433594, 0.025868288040161135, 0.026162176132202147, 0.02609459114074707, 0.02589593505859375, 0.0269117431640625, 0.026811391830444335, 0.06400614166259766, 0.025833471298217774, 0.02589286422729492, 0.025794559478759766, 0.025869312286376952, 0.025794559478759766, 0.025807872772216797, 0.025863168716430664, 0.02575667190551758, 0.0259102725982666, 0.0259051513671875, 0.025796607971191408, 0.026428415298461915, 0.025772031784057618, 0.027497472763061522, 0.02689023971557617, 0.026888191223144533, 0.026415103912353514, 0.02588569641113281, 0.025903104782104492, 0.02588057518005371, 0.02627174377441406, 0.026395647048950196, 0.027595775604248047, 0.026925056457519532, 0.026836992263793946, 0.026857471466064452, 0.026933248519897462, 0.026888191223144533, 0.02691481590270996, 0.026801183700561525, 0.026884063720703125, 0.02675814437866211, 0.026863616943359377, 0.026796031951904296, 0.02687385559082031, 0.02709503936767578, 0.026883071899414062, 0.02676736068725586, 0.027271167755126953, 0.027133951187133788, 0.026870784759521486, 0.026816511154174806, 0.02676531219482422, 0.02672230339050293, 0.02587238311767578, 0.025996288299560546, 0.02593894386291504, 0.025867263793945314, 0.025875455856323244, 0.026009599685668947, 0.025850879669189454, 0.02594918441772461, 0.025996288299560546, 0.02730700874328613, 0.02693734359741211, 0.026868736267089844, 0.026884096145629883, 0.02691695976257324, 0.026996639251708983, 0.026916864395141602, 0.02696294403076172, 0.02710220718383789, 0.06400409698486328, 0.025870336532592773, 0.02589798355102539, 0.026440704345703125, 0.02674995231628418, 0.02671615982055664, 0.02685148811340332, 0.02679075241088867, 0.026763263702392577, 0.02696294403076172, 0.0267325439453125, 0.0267827205657959, 0.02667519950866699, 0.026447999954223634, 0.026559360504150392, 0.026714111328125, 0.02647769546508789, 0.026601343154907225, 0.026578975677490235, 0.026445823669433592, 0.026621919631958008, 0.026635263442993166, 0.026574848175048828, 0.02653081512451172, 0.026467327117919923, 0.026385408401489258, 0.026464256286621093, 0.026456159591674806, 0.02649283218383789, 0.02646326446533203, 0.026441696166992188, 0.0265031681060791, 0.026806304931640625, 0.026963935852050782, 0.02691276741027832, 0.02675712013244629, 0.02692095947265625, 0.026808319091796876, 0.027273216247558595, 0.027060224533081056, 0.026849279403686522, 0.026444799423217775, 0.02590617561340332, 0.02595020866394043, 0.025969663619995118, 0.02593382453918457, 0.025878528594970703, 0.026506240844726563, 0.025987071990966795, 0.02587648010253906, 0.025991167068481445, 0.026060800552368164, 0.025891839981079103, 0.026006528854370117, 0.025891839981079103, 0.025987071990966795, 0.02589084815979004, 0.02595529556274414, 0.026002431869506838, 0.025994239807128908, 0.025848831176757812, 0.02593280029296875, 0.026001407623291017]",tokens/s,37.25363414793752,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,huggyllama/llama-13b,huggyllama/llama-13b,cuda,0,42,,,True,,,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",llama,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-2.7b,,cuda,0,42,,,,,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,d35829e539df8480b726c647eeabf91e41eae047,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,,MB,1197.428736,6415.712256,0.0,5769.265152,5440.254976,s,10,0.3418375663757324,0.03418375663757324,0.0004526914877200379,0.034001792907714844,0.03436094398498535,0.03493615207672119,0.03539631855010986,"[0.03551136016845703, 0.03400342559814453, 0.03394607925415039, 0.0340835189819336, 0.033914878845214845, 0.03400016021728516, 0.03417449569702148, 0.03399910354614258, 0.03423311996459961, 0.033971424102783206]",tokens/s,7488.9370034484855,kWh,4.0399560907549787e-07,2.213038476683643e-07,2.0643256939317695e-06,2.6896251506756316e-06,tokens/kWh,95180549.57795624,MB,1197.428736,6415.712256,0.0,5769.265152,5540.555776,s,10,11.932635009765624,1.1932635009765626,0.007760920411297161,1.1950701293945312,1.202022009277344,1.2022309143066408,1.2023980383300783,"[1.2024398193359376, 1.1798909912109374, 1.2019755859375, 1.196229736328125, 1.20013232421875, 1.196576171875, 1.1914827880859375, 1.1939105224609374, 1.19030078125, 1.1796962890625]",tokens/s,52.79638566707271,kWh,1.4040069683600796e-05,7.690784595813116e-06,5.135568926426249e-05,7.30865435436764e-05,tokens/kWh,861991.7832391581,,s,629,12.219545585632329,0.01942694051769845,0.004089590142775184,0.018989055633544923,0.019164159774780275,0.019244237518310548,0.05310250915527344,"[0.019119104385375976, 0.019154943466186524, 0.019129344940185547, 0.019164159774780275, 0.019141632080078123, 0.019173376083374022, 0.01949286460876465, 0.01924710464477539, 0.019091455459594727, 0.019083263397216797, 0.019106815338134766, 0.01926655960083008, 0.01940275192260742, 0.019164159774780275, 0.019164159774780275, 0.019095552444458007, 0.019153919219970703, 0.019132415771484376, 0.019156991958618166, 0.019292160034179686, 0.019150848388671874, 0.01909760093688965, 0.019216384887695313, 0.018960384368896483, 0.019019775390625, 0.01947340774536133, 0.01967616081237793, 0.019385343551635743, 0.019418111801147463, 0.01919692802429199, 0.019150848388671874, 0.019090431213378906, 0.019134464263916014, 0.019115007400512696, 0.019084287643432618, 0.01905561637878418, 0.01906073570251465, 0.019062784194946288, 0.01920614433288574, 0.019104768753051758, 0.01906892776489258, 0.019117055892944337, 0.01906175994873047, 0.019123199462890626, 0.019094528198242186, 0.019113983154296875, 0.018952192306518553, 0.018765823364257812, 0.018729984283447267, 0.018744319915771485, 0.018770944595336913, 0.018807807922363282, 0.018725887298583984, 0.018759679794311524, 0.018860031127929687, 0.01881088066101074, 0.018893823623657227, 0.018811904907226562, 0.0188272647857666, 0.018758655548095703, 0.018774015426635742, 0.018769983291625977, 0.05307385635375977, 0.018674751281738282, 0.01867462348937988, 0.01866649627685547, 0.01864192008972168, 0.018709503173828124, 0.018655231475830078, 0.01879449653625488, 0.018720767974853517, 0.018709503173828124, 0.018707456588745116, 0.018703359603881836, 0.018711551666259766, 0.018723840713500976, 0.018767871856689454, 0.018738176345825194, 0.018710527420043945, 0.018758655548095703, 0.018687999725341797, 0.01868185615539551, 0.01875660705566406, 0.018729984283447267, 0.018689023971557618, 0.01879756736755371, 0.01865318489074707, 0.01868083190917969, 0.018667520523071288, 0.018655231475830078, 0.018685951232910156, 0.018733152389526365, 0.018689952850341796, 0.018799615859985352, 0.018685951232910156, 0.018704383850097657, 0.01860095977783203, 0.01865318489074707, 0.018670591354370117, 0.018670591354370117, 0.018702335357666015, 0.0188221435546875, 0.018680864334106446, 0.018722784042358397, 0.018761728286743166, 0.01878937530517578, 0.018704383850097657, 0.01874025535583496, 0.018645023345947264, 0.01879750442504883, 0.018782207489013672, 0.018749439239501953, 0.01875872039794922, 0.0187391357421875, 0.018723840713500976, 0.018757631301879883, 0.018757631301879883, 0.01884160041809082, 0.018795520782470702, 0.018743295669555664, 0.018754560470581053, 0.018733055114746093, 0.018775039672851563, 0.01883443260192871, 0.018779136657714843, 0.05350604629516602, 0.020114431381225584, 0.019338239669799806, 0.019102720260620116, 0.01901670455932617, 0.019072000503540038, 0.01902387237548828, 0.019180543899536134, 0.018994176864624023, 0.01901568031311035, 0.019002368927001953, 0.019146751403808594, 0.01900748825073242, 0.01901055908203125, 0.01906585693359375, 0.019159040451049804, 0.01898700714111328, 0.018981887817382814, 0.018941951751708985, 0.018926591873168946, 0.01903206443786621, 0.019082239151000976, 0.018983936309814452, 0.019212287902832033, 0.019136512756347656, 0.019066879272460938, 0.019080192565917968, 0.019102720260620116, 0.01901055908203125, 0.018957311630249024, 0.018711551666259766, 0.01882624053955078, 0.018692096710205077, 0.018791423797607423, 0.019173376083374022, 0.019105791091918945, 0.019090431213378906, 0.019054592132568358, 0.01903411293029785, 0.01923276710510254, 0.019006464004516603, 0.019162111282348633, 0.019083263397216797, 0.019062784194946288, 0.019124223709106446, 0.019058687210083008, 0.01902796745300293, 0.019198976516723632, 0.01904844856262207, 0.019140607833862306, 0.019085344314575196, 0.019142623901367187, 0.019170303344726563, 0.01926553535461426, 0.019094528198242186, 0.019195903778076173, 0.01922764778137207, 0.019118080139160155, 0.019125247955322267, 0.019125247955322267, 0.018930688858032226, 0.01879756736755371, 0.018783231735229493, 0.05360025787353516, 0.018991104125976564, 0.019013631820678712, 0.018991104125976564, 0.019022911071777344, 0.019050432205200196, 0.01903104019165039, 0.019192832946777344, 0.01902387237548828, 0.018715648651123046, 0.018694143295288086, 0.018714624404907225, 0.019001344680786132, 0.01898700714111328, 0.019130367279052735, 0.019100671768188478, 0.01903718376159668, 0.01900748825073242, 0.018722816467285155, 0.01867263984680176, 0.018713600158691408, 0.01879859161376953, 0.01900339126586914, 0.019130367279052735, 0.018688064575195312, 0.01874118423461914, 0.018689023971557618, 0.018717695236206054, 0.018734079360961914, 0.018758655548095703, 0.01908121681213379, 0.019141632080078123, 0.018763776779174804, 0.018792448043823243, 0.018710527420043945, 0.018732032775878905, 0.018712575912475587, 0.018942975997924806, 0.019145727157592773, 0.019160064697265625, 0.019123199462890626, 0.01923481559753418, 0.019087360382080077, 0.01903923225402832, 0.0188272647857666, 0.019186687469482423, 0.019079168319702147, 0.019103744506835937, 0.01906585693359375, 0.01906892776489258, 0.01905561637878418, 0.019128320693969726, 0.019182592391967773, 0.019190784454345702, 0.019094528198242186, 0.019170303344726563, 0.019083263397216797, 0.019138559341430664, 0.019124223709106446, 0.019088384628295898, 0.019153919219970703, 0.019149824142456053, 0.01908940887451172, 0.053340160369873046, 0.01898700714111328, 0.018991104125976564, 0.019120128631591796, 0.019001344680786132, 0.019046464920043946, 0.019020736694335936, 0.018993152618408202, 0.01868185615539551, 0.018721792221069337, 0.01884160041809082, 0.019113983154296875, 0.01906483268737793, 0.019014720916748048, 0.019068864822387694, 0.019143680572509765, 0.018991104125976564, 0.018997247695922852, 0.019087360382080077, 0.01904947280883789, 0.01903615951538086, 0.0190515193939209, 0.018983936309814452, 0.019141632080078123, 0.019056640625, 0.019091455459594727, 0.019080192565917968, 0.019141632080078123, 0.019128320693969726, 0.01905971145629883, 0.01925017547607422, 0.01902079963684082, 0.018724863052368163, 0.01896147155761719, 0.019042240142822266, 0.019140607833862306, 0.01903923225402832, 0.019078144073486326, 0.019080192565917968, 0.019185663223266602, 0.019163135528564454, 0.019117055892944337, 0.019119104385375976, 0.019084320068359376, 0.01910063934326172, 0.019050495147705078, 0.019021823883056642, 0.019169279098510742, 0.019090431213378906, 0.01902387237548828, 0.018932735443115235, 0.018980863571166993, 0.018945024490356444, 0.019000383377075197, 0.01899513626098633, 0.019870719909667968, 0.01921843147277832, 0.019105791091918945, 0.019120128631591796, 0.01904435157775879, 0.01880166435241699, 0.018778112411499022, 0.018803712844848632, 0.05313843154907227, 0.018669567108154296, 0.01866547203063965, 0.01866547203063965, 0.018647039413452148, 0.018703359603881836, 0.018671615600585938, 0.018738176345825194, 0.01923276710510254, 0.019766271591186522, 0.019405824661254883, 0.01920921516418457, 0.0190382080078125, 0.01881497573852539, 0.019050559997558593, 0.0191026554107666, 0.019062784194946288, 0.01906892776489258, 0.019078144073486326, 0.0190218563079834, 0.018999263763427733, 0.01908940887451172, 0.018670591354370117, 0.01880575942993164, 0.018721792221069337, 0.018722816467285155, 0.01864806365966797, 0.018740224838256835, 0.018682880401611326, 0.018749439239501953, 0.018757631301879883, 0.018956287384033203, 0.01904025650024414, 0.01909350395202637, 0.019120128631591796, 0.0190515193939209, 0.019110912322998046, 0.019112960815429687, 0.019058687210083008, 0.019151872634887695, 0.019042303085327148, 0.019066879272460938, 0.019080192565917968, 0.019062784194946288, 0.019056640625, 0.019112960815429687, 0.019056640625, 0.019163135528564454, 0.019100671768188478, 0.018745344161987306, 0.018713600158691408, 0.01883340835571289, 0.019141632080078123, 0.019121152877807617, 0.019096576690673828, 0.019188735961914064, 0.0190699520111084, 0.019138559341430664, 0.019094528198242186, 0.019118080139160155, 0.019144704818725586, 0.019126272201538085, 0.01927065658569336, 0.053359615325927735, 0.018856960296630858, 0.01906892776489258, 0.019062784194946288, 0.01925529670715332, 0.019106815338134766, 0.01901568031311035, 0.019115039825439453, 0.019043296813964845, 0.01903615951538086, 0.019009567260742186, 0.019057632446289063, 0.0190382080078125, 0.018983936309814452, 0.018993152618408202, 0.019194879531860352, 0.018970624923706055, 0.01902796745300293, 0.019050495147705078, 0.0190699520111084, 0.019075071334838867, 0.019080192565917968, 0.018989055633544923, 0.019167232513427734, 0.018959360122680666, 0.01865216064453125, 0.018671615600585938, 0.018705408096313478, 0.0186746883392334, 0.018723840713500976, 0.018704383850097657, 0.01879347229003906, 0.018675712585449217, 0.018704383850097657, 0.018754560470581053, 0.018840576171875, 0.018910207748413087, 0.01878118324279785, 0.018761728286743166, 0.01886207962036133, 0.018716672897338867, 0.018728960037231446, 0.018724863052368163, 0.01882931137084961, 0.018715648651123046, 0.018742271423339844, 0.018727935791015626, 0.0189040641784668, 0.01882111930847168, 0.018723840713500976, 0.018784255981445314, 0.018776063919067384, 0.018822208404541015, 0.018768831253051756, 0.018776063919067384, 0.0188723201751709, 0.018787328720092773, 0.019121152877807617, 0.019124223709106446, 0.019104768753051758, 0.01909350395202637, 0.019148799896240236, 0.019083263397216797, 0.05310566329956055, 0.018693119049072265, 0.01866547203063965, 0.01863372802734375, 0.018669599533081054, 0.01869308853149414, 0.01864192008972168, 0.018704383850097657, 0.01864192008972168, 0.01868185615539551, 0.018671615600585938, 0.0186562557220459, 0.018654207229614257, 0.018689023971557618, 0.01866035270690918, 0.018682880401611326, 0.018679807662963867, 0.018678783416748047, 0.018669567108154296, 0.01869824028015137, 0.01869824028015137, 0.018732032775878905, 0.018717695236206054, 0.018767871856689454, 0.018704383850097657, 0.018724863052368163, 0.018693119049072265, 0.01884364891052246, 0.01885081672668457, 0.019583999633789064, 0.01944780731201172, 0.019142656326293944, 0.01900851249694824, 0.019137535095214844, 0.019046464920043946, 0.01901254463195801, 0.01907711982727051, 0.01904640007019043, 0.01905971145629883, 0.019139583587646485, 0.02007040023803711, 0.019281919479370118, 0.019087360382080077, 0.01923993682861328, 0.019058687210083008, 0.019099647521972657, 0.01906892776489258, 0.019144704818725586, 0.019080192565917968, 0.019066879272460938, 0.01905971145629883, 0.019074047088623047, 0.019105791091918945, 0.01913043212890625, 0.019135456085205078, 0.0191549129486084, 0.01906073570251465, 0.019160064697265625, 0.019139583587646485, 0.019094528198242186, 0.019109888076782225, 0.01911302375793457, 0.019096511840820313, 0.05369036865234375, 0.01900748825073242, 0.018962432861328125, 0.01902796745300293, 0.018994176864624023, 0.01901670455932617, 0.018980863571166993, 0.019109888076782225, 0.01882316780090332, 0.018790399551391602, 0.01864089584350586, 0.01867263984680176, 0.018682880401611326, 0.01864806365966797, 0.01863167953491211, 0.018795520782470702, 0.018704416275024414, 0.01906480026245117, 0.01901260757446289, 0.019135488510131835, 0.01901670455932617, 0.01902899169921875, 0.018976831436157228, 0.01932588768005371, 0.019091455459594727, 0.01899929618835449, 0.018969600677490234, 0.019182592391967773, 0.018981887817382814, 0.01904742431640625, 0.019062784194946288, 0.019198976516723632, 0.01881497573852539, 0.018765823364257812, 0.01880268859863281, 0.018741247177124023, 0.018693119049072265, 0.018724863052368163, 0.018744319915771485, 0.01882521629333496, 0.018725887298583984, 0.018780160903930664, 0.018699264526367186, 0.018720800399780274, 0.018765792846679688, 0.01875868797302246, 0.01869513511657715, 0.01886617660522461, 0.018723840713500976, 0.018915327072143554, 0.01905254364013672, 0.019090431213378906, 0.019042303085327148, 0.019117055892944337, 0.018923519134521484, 0.01887948799133301, 0.018718719482421875, 0.018757631301879883, 0.018807807922363282, 0.018775039672851563, 0.018784255981445314, 0.018739200592041014, 0.01884364891052246, 0.05309439849853516, 0.018684928894042968, 0.018682880401611326, 0.018576383590698242, 0.018610176086425782, 0.018725887298583984, 0.018701311111450195, 0.018755584716796874, 0.01865318489074707, 0.0186562557220459, 0.0186746883392334, 0.018678783416748047, 0.01864806365966797, 0.018682880401611326, 0.0186562557220459, 0.018717695236206054, 0.01869824028015137, 0.018739200592041014, 0.018676736831665038, 0.018757631301879883, 0.018646015167236327, 0.018689023971557618, 0.01882316780090332, 0.01879654312133789, 0.018687999725341797, 0.01866854476928711, 0.01864192008972168, 0.018682880401611326, 0.018659328460693358, 0.01865216064453125, 0.018695167541503906, 0.018739200592041014, 0.018719743728637696, 0.018757631301879883, 0.018682880401611326, 0.018718719482421875, 0.018683904647827147, 0.018716672897338867, 0.01875564765930176, 0.018796480178833008, 0.018710527420043945, 0.018752511978149415, 0.018745344161987306, 0.0186562557220459, 0.018733055114746093, 0.018777088165283205, 0.01872800064086914, 0.018809791564941405, 0.01876479911804199, 0.018752511978149415, 0.018697216033935548, 0.018804800033569335, 0.018753471374511718, 0.018785280227661134, 0.01883340835571289, 0.01883545684814453, 0.018729984283447267, 0.018832384109497072, 0.018764863967895506, 0.01871766471862793, 0.01877302360534668, 0.01876371192932129, 0.018704383850097657]",tokens/s,51.47490924208955,,,main,False,False,,,,,,,,,,,,,,,,,,,,,,, float16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,microsoft/phi-1_5,microsoft/phi-1_5,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1271.840768,3876.061184,0.0,3229.61408,2982.448128,s,10,0.28811154937744143,0.028811154937744145,0.0002979086068498023,0.028700287818908692,0.028956361579895017,0.029322308444976804,0.029615065937042235,"[0.029688255310058594, 0.02870924758911133, 0.028731008529663087, 0.028668575286865235, 0.028669055938720704, 0.028681375503540038, 0.028691328048706055, 0.02868569564819336, 0.02871196746826172, 0.028875040054321288]",tokens/s,8885.447339864408,kWh,3.4075449654737655e-07,1.86653728925752e-07,1.5784194503746478e-06,2.105827675847776e-06,tokens/kWh,121567402.18400732,MB,1271.840768,3876.061184,0.0,3229.61408,3023.634432,s,10,12.861451660156249,1.2861451660156251,0.02609561377566638,1.2746920776367188,1.3334684936523438,1.3382678161621095,1.342107274169922,"[1.343067138671875, 1.3324019775390625, 1.2789603271484375, 1.2706533203125, 1.2760494384765626, 1.2697115478515626, 1.270895751953125, 1.273334716796875, 1.2690650634765626, 1.2773123779296875]",tokens/s,48.98358417438131,kWh,1.503040841139741e-05,8.23651604456098e-06,3.988086371702524e-05,6.314778817298361e-05,tokens/kWh,997659.6460896023,,s,629,13.0970500717163,0.020822019191917816,0.003485508965752573,0.0202106876373291,0.02136186866760254,0.021463833618164063,0.049208116149902344,"[0.022583263397216797, 0.022549503326416014, 0.020750335693359375, 0.020810752868652343, 0.020859968185424804, 0.02087424087524414, 0.020255680084228515, 0.02040729522705078, 0.02309017562866211, 0.022816768646240236, 0.021749759674072267, 0.021597183227539063, 0.02147020721435547, 0.02142720031738281, 0.021415935516357423, 0.021369855880737306, 0.021345279693603517, 0.021375999450683594, 0.021754880905151368, 0.02163609504699707, 0.021448703765869142, 0.021390335083007812, 0.021404735565185545, 0.021458879470825195, 0.021407743453979493, 0.021571584701538086, 0.021425151824951173, 0.02149580764770508, 0.02146713638305664, 0.021384191513061524, 0.02144972801208496, 0.021382144927978516, 0.021403648376464843, 0.021215232849121093, 0.02105548858642578, 0.021021696090698243, 0.021005311965942384, 0.021029888153076173, 0.021016576766967773, 0.020992000579833983, 0.021006336212158205, 0.021120000839233398, 0.021105663299560547, 0.021251071929931642, 0.021377023696899415, 0.021340160369873046, 0.02102579116821289, 0.020972543716430665, 0.02101043128967285, 0.02101043128967285, 0.020995071411132812, 0.02002739143371582, 0.020333568572998048, 0.021204992294311522, 0.021356544494628905, 0.021380096435546874, 0.021338111877441408, 0.021136383056640624, 0.021055551528930665, 0.02120902442932129, 0.021112831115722656, 0.021371904373168944, 0.04938854217529297, 0.020249599456787108, 0.02029465675354004, 0.020296703338623046, 0.020378623962402344, 0.020213760375976563, 0.020256767272949217, 0.020280319213867186, 0.020248575210571287, 0.020133920669555664, 0.020148191452026367, 0.02028544044494629, 0.021203968048095705, 0.02164838409423828, 0.021388288497924804, 0.02145484733581543, 0.021415935516357423, 0.021412864685058593, 0.021209087371826172, 0.021190656661987304, 0.021804031372070314, 0.022599679946899414, 0.021857280731201172, 0.021517311096191406, 0.021361663818359376, 0.02138115119934082, 0.02134524726867676, 0.021525503158569336, 0.021356544494628905, 0.02129305648803711, 0.02130227279663086, 0.021235712051391603, 0.02106675148010254, 0.02106879997253418, 0.021192703247070312, 0.021353471755981446, 0.021380096435546874, 0.02145382308959961, 0.021304319381713867, 0.021361663818359376, 0.021436416625976562, 0.021008384704589843, 0.02148249626159668, 0.02147327995300293, 0.021387264251708983, 0.021362688064575194, 0.021394432067871092, 0.021308416366577147, 0.021343231201171875, 0.021436416625976562, 0.021329919815063478, 0.021359615325927735, 0.021341184616088867, 0.021346303939819337, 0.021331968307495116, 0.0211015682220459, 0.02106572723388672, 0.0210882568359375, 0.02144767951965332, 0.021114879608154297, 0.02106777572631836, 0.02105036735534668, 0.02102579116821289, 0.04940390396118164, 0.01998847961425781, 0.020092927932739257, 0.0202926082611084, 0.020776960372924806, 0.02103193664550781, 0.021331968307495116, 0.02131455993652344, 0.021283840179443358, 0.02100534439086914, 0.02129305648803711, 0.021472223281860353, 0.02143129539489746, 0.021154815673828126, 0.02000486373901367, 0.020134912490844727, 0.02003660774230957, 0.019973119735717772, 0.02001817512512207, 0.020222976684570314, 0.020231168746948244, 0.020008960723876954, 0.02026803207397461, 0.02027724838256836, 0.02024140739440918, 0.020147199630737304, 0.020124671936035156, 0.019984384536743165, 0.019989503860473632, 0.020150272369384766, 0.02008678436279297, 0.019944448471069336, 0.020033536911010744, 0.019991552352905274, 0.02007142448425293, 0.020023296356201172, 0.0200581111907959, 0.02025574493408203, 0.020205568313598633, 0.020238336563110353, 0.020215808868408205, 0.02024140739440918, 0.020057151794433594, 0.020017087936401366, 0.02003558349609375, 0.020092927932739257, 0.02002739143371582, 0.020102144241333008, 0.020017152786254884, 0.01999564743041992, 0.0200130558013916, 0.020059135437011717, 0.020368383407592772, 0.02008064079284668, 0.020215808868408205, 0.020101119995117187, 0.020110336303710938, 0.020299776077270508, 0.02022502326965332, 0.020238367080688477, 0.020202463150024413, 0.020239360809326173, 0.020185087203979494, 0.04912332916259766, 0.019997695922851562, 0.019958784103393554, 0.019970048904418947, 0.02000588798522949, 0.020162559509277343, 0.020353023529052734, 0.020237312316894532, 0.020174848556518556, 0.021166080474853514, 0.02089779281616211, 0.020586528778076173, 0.020080608367919923, 0.02027519989013672, 0.02019430351257324, 0.02028339195251465, 0.02024550437927246, 0.020280319213867186, 0.02004278373718262, 0.02012668800354004, 0.020205568313598633, 0.02027212715148926, 0.02005606460571289, 0.020099071502685546, 0.02005401611328125, 0.020082687377929686, 0.02003046417236328, 0.020079647064208984, 0.020024288177490236, 0.02020249557495117, 0.02000486373901367, 0.02001817512512207, 0.020006912231445313, 0.020092927932739257, 0.020031488418579102, 0.020033536911010744, 0.02004787254333496, 0.02005504035949707, 0.020082687377929686, 0.02004582405090332, 0.0200447998046875, 0.020060159683227538, 0.020015104293823242, 0.020067327499389647, 0.02005401611328125, 0.020135936737060548, 0.020331520080566406, 0.020050943374633787, 0.020257791519165038, 0.02025369644165039, 0.02003865623474121, 0.020065280914306642, 0.020162559509277343, 0.02023219108581543, 0.020373504638671876, 0.020338687896728515, 0.02003046417236328, 0.02026188850402832, 0.020280319213867186, 0.020093952178955078, 0.02004991912841797, 0.020246528625488282, 0.02021785545349121, 0.04924415969848633, 0.020114431381225584, 0.02017791938781738, 0.02008064079284668, 0.02006630325317383, 0.02004684829711914, 0.02026803207397461, 0.020330495834350586, 0.020213760375976563, 0.02024038314819336, 0.020242431640625, 0.02027212715148926, 0.020207616806030275, 0.02024448013305664, 0.02045132827758789, 0.021708799362182618, 0.02126438331604004, 0.021097471237182617, 0.02086297607421875, 0.020738048553466795, 0.020739072799682616, 0.020780031204223632, 0.020333568572998048, 0.020271104812622072, 0.02018819236755371, 0.020217824935913085, 0.02023423957824707, 0.020166656494140626, 0.020170751571655272, 0.020146175384521483, 0.02030284881591797, 0.020141056060791016, 0.020239360809326173, 0.02003558349609375, 0.020017152786254884, 0.020059135437011717, 0.02004991912841797, 0.020098047256469728, 0.020146175384521483, 0.020190208435058594, 0.02024038314819336, 0.02042470359802246, 0.020123647689819335, 0.020015104293823242, 0.0200581111907959, 0.01999667167663574, 0.02001923179626465, 0.019962848663330077, 0.019981311798095702, 0.019993600845336915, 0.02002739143371582, 0.02003763198852539, 0.019979263305664064, 0.019982336044311523, 0.02024550437927246, 0.020304895401000975, 0.020190208435058594, 0.02021171188354492, 0.02016972732543945, 0.020213760375976563, 0.02027724838256836, 0.019957759857177734, 0.019951616287231445, 0.04933017730712891, 0.02021171188354492, 0.020098047256469728, 0.020006912231445313, 0.019985408782958985, 0.019974143981933593, 0.020076543807983398, 0.019977216720581056, 0.019987455368041994, 0.020170751571655272, 0.02022604751586914, 0.02029465675354004, 0.02021990394592285, 0.0202926082611084, 0.020180992126464844, 0.02023526382446289, 0.020287488937377928, 0.020197376251220703, 0.020182016372680665, 0.020075519561767577, 0.020163583755493163, 0.02028441619873047, 0.0200898551940918, 0.020050943374633787, 0.020130815505981444, 0.019975168228149414, 0.020178943634033202, 0.020198400497436524, 0.02004172706604004, 0.019919872283935547, 0.019961856842041017, 0.020015104293823242, 0.019987455368041994, 0.019978239059448243, 0.01999564743041992, 0.02021887969970703, 0.02023628807067871, 0.02023526382446289, 0.019982336044311523, 0.02003865623474121, 0.02026192092895508, 0.020079584121704103, 0.019962879180908204, 0.020495359420776366, 0.02062848091125488, 0.020282367706298828, 0.02021990394592285, 0.02025164794921875, 0.020296703338623046, 0.020229120254516602, 0.020154367446899413, 0.02027008056640625, 0.020160512924194338, 0.02021171188354492, 0.020147199630737304, 0.02005504035949707, 0.02003660774230957, 0.02027724838256836, 0.02018611145019531, 0.02025881576538086, 0.02003763198852539, 0.020176895141601564, 0.02004991912841797, 0.049114112854003904, 0.020166656494140626, 0.020264959335327147, 0.02012876892089844, 0.020024320602416993, 0.020000768661499024, 0.019993600845336915, 0.020653055191040038, 0.020813823699951172, 0.020789247512817383, 0.020799488067626954, 0.020935680389404295, 0.020728832244873048, 0.020781055450439453, 0.020339712142944336, 0.020239360809326173, 0.02006220817565918, 0.020191232681274415, 0.020062240600585937, 0.019984352111816407, 0.02006118392944336, 0.02026393508911133, 0.020312063217163084, 0.02004275131225586, 0.020213760375976563, 0.020173824310302735, 0.020167680740356447, 0.019946495056152345, 0.019964927673339843, 0.019963903427124022, 0.019966976165771484, 0.019961856842041017, 0.02006937599182129, 0.019958784103393554, 0.019998720169067383, 0.019992576599121094, 0.02004172706604004, 0.019998720169067383, 0.019974143981933593, 0.019991552352905274, 0.019994623184204103, 0.02010419273376465, 0.02003660774230957, 0.019969024658203126, 0.019981311798095702, 0.02000588798522949, 0.020032512664794923, 0.020025344848632814, 0.019978239059448243, 0.020006912231445313, 0.020089920043945313, 0.020245439529418947, 0.020264959335327147, 0.02020249557495117, 0.02024345588684082, 0.020249599456787108, 0.020238336563110353, 0.02023526382446289, 0.02025369644165039, 0.020006912231445313, 0.02025369644165039, 0.02004377555847168, 0.02001919937133789, 0.0492410888671875, 0.02004377555847168, 0.020063232421875, 0.020092927932739257, 0.020495359420776366, 0.02025267219543457, 0.02026905632019043, 0.020435968399047853, 0.02023321533203125, 0.020279296875, 0.020289535522460937, 0.020190208435058594, 0.02003558349609375, 0.020057088851928712, 0.020023296356201172, 0.02009600067138672, 0.020108287811279296, 0.020050943374633787, 0.020057088851928712, 0.0200447998046875, 0.020024320602416993, 0.02004377555847168, 0.02003558349609375, 0.020083711624145507, 0.02004377555847168, 0.019957759857177734, 0.020033536911010744, 0.019973119735717772, 0.020060159683227538, 0.019960832595825196, 0.01992192077636719, 0.019978239059448243, 0.02006937599182129, 0.020011007308959963, 0.020000768661499024, 0.02007040023803711, 0.020197376251220703, 0.02027008056640625, 0.020113407135009767, 0.020187135696411132, 0.0202608642578125, 0.02027212715148926, 0.020142080307006836, 0.020189184188842774, 0.020007936477661133, 0.020060159683227538, 0.020254720687866212, 0.020513792037963868, 0.022350847244262697, 0.02128281593322754, 0.020932640075683594, 0.020818912506103515, 0.020195327758789062, 0.02006425666809082, 0.02022400093078613, 0.020319232940673827, 0.020174848556518556, 0.02024140739440918, 0.020214784622192384, 0.020118528366088868, 0.02001408004760742, 0.02002739143371582, 0.019976192474365235, 0.04934656143188477, 0.020031488418579102, 0.01997209548950195, 0.020011007308959963, 0.019992576599121094, 0.019958784103393554, 0.020016128540039063, 0.02002739143371582, 0.019983360290527344, 0.020007936477661133, 0.020059135437011717, 0.02003046417236328, 0.020015104293823242, 0.020017152786254884, 0.01999465560913086, 0.020016096115112306, 0.020145151138305666, 0.019979263305664064, 0.02006118392944336, 0.019999744415283204, 0.019952640533447266, 0.020074495315551756, 0.02001408004760742, 0.02018611145019531, 0.020141056060791016, 0.019919872283935547, 0.0200898551940918, 0.019912704467773438, 0.019966976165771484, 0.019899391174316407, 0.020206592559814454, 0.020132863998413086, 0.019958784103393554, 0.020165632247924805, 0.020191232681274415, 0.020189184188842774, 0.020246528625488282, 0.02024550437927246, 0.020264959335327147, 0.020161535263061522, 0.020212736129760742, 0.02020351982116699, 0.020176895141601564, 0.02020351982116699, 0.02027622413635254, 0.020289535522460937, 0.020296703338623046, 0.0201943359375, 0.020255712509155272, 0.020197376251220703, 0.020204544067382812, 0.02027519989013672, 0.020153343200683595, 0.02000383949279785, 0.020170751571655272, 0.020303871154785155, 0.020254720687866212, 0.019952640533447266, 0.020189184188842774, 0.020145151138305666, 0.02030080032348633, 0.020720640182495118, 0.021169151306152344, 0.05058969497680664, 0.020896768569946288, 0.020787200927734374, 0.020810752868652343, 0.02022400093078613, 0.020137983322143553, 0.020289535522460937, 0.020329471588134765, 0.02025369644165039, 0.020147199630737304, 0.02021683120727539, 0.0202926082611084, 0.020172800064086914, 0.02024140739440918, 0.02019536018371582, 0.02032431983947754, 0.020279327392578126, 0.020260831832885743, 0.02010419273376465, 0.020150272369384766, 0.020148223876953125, 0.02020147132873535, 0.02012774467468262, 0.020149248123168945, 0.020195327758789062, 0.020173824310302735, 0.020299808502197265, 0.020226015090942382, 0.020173824310302735, 0.020135936737060548, 0.020228160858154296, 0.02017990493774414, 0.020187135696411132, 0.020337663650512695, 0.020791296005249024, 0.020320255279541014, 0.02021785545349121, 0.020183040618896485, 0.020156415939331054, 0.0202106876373291, 0.020197376251220703, 0.020191232681274415, 0.020143104553222657, 0.020175872802734376, 0.02027212715148926, 0.020179967880249023, 0.020348928451538087, 0.02020147132873535, 0.020271104812622072, 0.020206592559814454, 0.02023526382446289, 0.020273151397705077, 0.020187135696411132, 0.02022502326965332, 0.02003660774230957, 0.020125696182250977, 0.020290559768676757, 0.02024345588684082, 0.020214784622192384, 0.02021887969970703, 0.020256767272949217, 0.02025984001159668, 0.020191232681274415]",tokens/s,48.026081946373175,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,google/recurrentgemma-2b,google/recurrentgemma-2b,cuda,0,42,,,True,,,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.223-212.873.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.0,,,,1.21.4,,,,0.12.0,,,recurrent_gemma,MB,1691.762688,7434.928128,0.0,6784.28672,5879.090688,s,10,3.831963806152344,0.38319638061523437,0.003839812047765418,0.38284257507324215,0.38838264770507813,0.3898609405517578,0.39104357482910157,"[0.3913392333984375, 0.3797502746582031, 0.38331051635742186, 0.3848867492675781, 0.3808703308105469, 0.38805413818359374, 0.38342922973632815, 0.37826702880859375, 0.3796816711425781, 0.3823746337890625]",tokens/s,668.0647650924667,kWh,4.61594101469032e-06,2.5293117124833596e-06,8.84391733153802e-06,1.59891700587117e-05,tokens/kWh,16010837.27673022,MB,1751.47008,7437.02528,0.0,6786.383872,5879.093248,s,10,21.371644775390628,2.137164477539063,0.023973905228054454,2.1295450439453125,2.1668739501953125,2.1699256225585937,2.1723669604492186,"[2.15330078125, 2.1282978515625, 2.130792236328125, 2.1070888671875, 2.12294677734375, 2.172977294921875, 2.12341552734375, 2.10235400390625, 2.16619580078125, 2.164275634765625]",tokens/s,29.47831140846224,kWh,2.5073409720378897e-05,1.3741426183364267e-05,6.954214537726236e-05,0.00010835698128100554,tokens/kWh,581411.5459401747,,s,630,21.370241035461405,0.03392101751660544,0.0007929925552596019,0.03343871879577637,0.03487754364013672,0.035152075576782224,0.0359699564743042,"[0.03523891067504883, 0.03520204925537109, 0.03503308868408203, 0.034974720001220705, 0.03497881698608398, 0.0354150390625, 0.03581542587280274, 0.035110912322998046, 0.035110912322998046, 0.03465420913696289, 0.034931713104248044, 0.03503206253051758, 0.03496857452392578, 0.035023872375488284, 0.034915328979492184, 0.03510988616943359, 0.03566080093383789, 0.03566284942626953, 0.035244033813476565, 0.0355676155090332, 0.034767871856689454, 0.03461324691772461, 0.033514495849609374, 0.03336601638793945, 0.033380352020263675, 0.03331891250610351, 0.03338137435913086, 0.033053695678710936, 0.03317145538330078, 0.03326873779296875, 0.033304576873779294, 0.03325132751464844, 0.033258495330810545, 0.03331071853637695, 0.03334348678588867, 0.033448959350585936, 0.034484222412109376, 0.033454078674316406, 0.033360897064208986, 0.03313868713378906, 0.033165313720703124, 0.033339393615722655, 0.03322470474243164, 0.03324825668334961, 0.033263614654541016, 0.03322880172729492, 0.03317657470703125, 0.033173503875732424, 0.0344719352722168, 0.03477196884155274, 0.033314815521240236, 0.03347967910766601, 0.033372161865234375, 0.03322880172729492, 0.03395891189575195, 0.03478220748901367, 0.03463065719604492, 0.034576385498046876, 0.03455078506469727, 0.03464089584350586, 0.03459481430053711, 0.034592769622802735, 0.034531326293945314, 0.033535999298095705, 0.033511425018310545, 0.033301502227783206, 0.03369881439208984, 0.03525632095336914, 0.03485696029663086, 0.033279998779296875, 0.033260543823242186, 0.03323187255859375, 0.03448524856567383, 0.03463884735107422, 0.034597888946533206, 0.03350425720214844, 0.03327897644042969, 0.03323904037475586, 0.03323289489746094, 0.03332710266113281, 0.03343769454956055, 0.03339571380615235, 0.03336191940307617, 0.03341209411621094, 0.03379404830932617, 0.03463270568847656, 0.034105342864990236, 0.03341107177734375, 0.033271808624267575, 0.033309696197509765, 0.03328204727172852, 0.0332410888671875, 0.033291263580322264, 0.03324313735961914, 0.03322880172729492, 0.03322060775756836, 0.03436236953735351, 0.03361894226074219, 0.03344793701171875, 0.033276927947998046, 0.03465318298339844, 0.034549758911132815, 0.03443097686767578, 0.033291263580322264, 0.03330047988891602, 0.03323187255859375, 0.03327078247070313, 0.03335475158691406, 0.03330867385864258, 0.03335270309448242, 0.033320960998535154, 0.033314815521240236, 0.03491635131835937, 0.034732032775878906, 0.03497881698608398, 0.03352576065063476, 0.03347455978393555, 0.033928192138671875, 0.03334348678588867, 0.03384524917602539, 0.03465932846069336, 0.03467468643188477, 0.0343818244934082, 0.03445555114746094, 0.03469209671020508, 0.03458457565307617, 0.03350527954101563, 0.03368755340576172, 0.03346636962890625, 0.033298431396484376, 0.03324006271362305, 0.03337420654296875, 0.033865726470947266, 0.03466649627685547, 0.033192958831787106, 0.033127422332763674, 0.03402956771850586, 0.03337932968139649, 0.03323289489746094, 0.03326464080810547, 0.03316326522827148, 0.033260543823242186, 0.03439308929443359, 0.03627212905883789, 0.034869247436523435, 0.03469311904907227, 0.033189888000488284, 0.033140735626220705, 0.033172481536865236, 0.03317964935302734, 0.03318272018432617, 0.033242111206054685, 0.033290241241455076, 0.032906238555908206, 0.03317452621459961, 0.03328102493286133, 0.033173503875732424, 0.03326668930053711, 0.033172481536865236, 0.03317251205444336, 0.033183712005615235, 0.0331069450378418, 0.033263614654541016, 0.03381964874267578, 0.03597619247436523, 0.0348487663269043, 0.03515596771240234, 0.035517440795898435, 0.03457535934448242, 0.03458662414550781, 0.03446476745605469, 0.03450470352172851, 0.034530303955078126, 0.03457024002075195, 0.03541196823120117, 0.03482316970825195, 0.034751487731933595, 0.03454873657226563, 0.03458355331420898, 0.03316940689086914, 0.03323699188232422, 0.033173503875732424, 0.03309568023681641, 0.03322675323486328, 0.033255424499511715, 0.03362918472290039, 0.03469107055664063, 0.03322470474243164, 0.03320217514038086, 0.033588222503662106, 0.0333834228515625, 0.033276927947998046, 0.0333199348449707, 0.033317886352539065, 0.03322265625, 0.03322060775756836, 0.033260543823242186, 0.03317964935302734, 0.033132545471191405, 0.033154048919677735, 0.03315097427368164, 0.033175552368164066, 0.03318783950805664, 0.033154048919677735, 0.03323187255859375, 0.03321036911010742, 0.03327385711669922, 0.033185791015625, 0.03326873779296875, 0.0332861442565918, 0.0349224967956543, 0.03460300827026367, 0.034184192657470705, 0.03456204986572266, 0.03326976013183594, 0.034738174438476564, 0.03366604614257813, 0.033511425018310545, 0.033306625366210936, 0.03335270309448242, 0.033263614654541016, 0.03324006271362305, 0.03443302536010742, 0.03463782501220703, 0.03452108764648437, 0.033375232696533204, 0.03322470474243164, 0.033568767547607424, 0.033323009490966796, 0.03323699188232422, 0.0331253776550293, 0.03314995193481445, 0.03314176177978516, 0.03317452621459961, 0.033188865661621096, 0.03318783950805664, 0.03323801422119141, 0.03320729446411133, 0.03326873779296875, 0.03316633605957031, 0.033585151672363284, 0.03325235366821289, 0.03324006271362305, 0.0332861442565918, 0.03317145538330078, 0.033288192749023435, 0.03319705581665039, 0.0336732177734375, 0.03371724700927734, 0.03326259231567383, 0.03315507125854492, 0.03315609741210938, 0.033416191101074216, 0.03329536056518555, 0.03321139144897461, 0.033279998779296875, 0.03476070404052734, 0.03329536056518555, 0.033304576873779294, 0.03330867385864258, 0.03311718368530273, 0.0331960334777832, 0.03324620819091797, 0.033205249786376956, 0.0331673583984375, 0.03320012664794922, 0.033242111206054685, 0.03408793640136719, 0.03403878402709961, 0.033481727600097655, 0.034113536834716796, 0.035146751403808595, 0.03459993743896484, 0.03470336151123047, 0.03328921508789062, 0.03329536056518555, 0.03327897644042969, 0.033261566162109374, 0.033035263061523434, 0.03324620819091797, 0.0332564468383789, 0.03317452621459961, 0.03325235366821289, 0.033245216369628905, 0.03326563262939453, 0.03339263916015625, 0.03333529663085937, 0.033375232696533204, 0.033416191101074216, 0.03371929550170898, 0.03340185546875, 0.03326873779296875, 0.033276927947998046, 0.033334270477294925, 0.03333222579956055, 0.03331174468994141, 0.03328102493286133, 0.03325132751464844, 0.034105342864990236, 0.033486846923828126, 0.03479244613647461, 0.03482931137084961, 0.03329228973388672, 0.034598911285400394, 0.03323494338989258, 0.03330559921264648, 0.03330559921264648, 0.033288192749023435, 0.03561574554443359, 0.034830337524414064, 0.03464601516723633, 0.03473715209960938, 0.03463065719604492, 0.03340390396118164, 0.036982784271240236, 0.034830337524414064, 0.034961406707763674, 0.03605196762084961, 0.035454975128173825, 0.03474124908447265, 0.03465830230712891, 0.03447091293334961, 0.033342464447021485, 0.03465830230712891, 0.034754558563232424, 0.03469823837280273, 0.03463679885864258, 0.03464089584350586, 0.03462656021118164, 0.03455487823486328, 0.03469107055664063, 0.034601982116699216, 0.034536449432373044, 0.034590721130371094, 0.034644992828369144, 0.034385921478271485, 0.0346798095703125, 0.03467264175415039, 0.03321139144897461, 0.03331174468994141, 0.03357388687133789, 0.0336363525390625, 0.03471155166625976, 0.03479142379760742, 0.03331071853637695, 0.03434700775146484, 0.034557952880859374, 0.0358205451965332, 0.03474943923950195, 0.03469004821777344, 0.033307647705078124, 0.03323801422119141, 0.034253822326660154, 0.034781185150146485, 0.034713600158691404, 0.034710529327392575, 0.03321343994140625, 0.03330355072021484, 0.034634750366210935, 0.0345456657409668, 0.03455487823486328, 0.03467264175415039, 0.03464908981323242, 0.034563072204589845, 0.03519385528564453, 0.03479859161376953, 0.03450265502929688, 0.03465727996826172, 0.03464704132080078, 0.03431731033325195, 0.0342845458984375, 0.03319705581665039, 0.03385139083862305, 0.03457331085205078, 0.03465216064453125, 0.03471769714355469, 0.035817470550537106, 0.035885055541992186, 0.03362918472290039, 0.03336806488037109, 0.0332492790222168, 0.033380352020263675, 0.033291263580322264, 0.03333631896972656, 0.033317886352539065, 0.033259521484375, 0.0332861442565918, 0.03323801422119141, 0.03317452621459961, 0.033258495330810545, 0.033189888000488284, 0.03321548843383789, 0.03463577651977539, 0.037026817321777344, 0.03546112060546875, 0.03485184097290039, 0.035460094451904296, 0.03340083312988281, 0.03334656143188477, 0.03374899291992187, 0.033430526733398434, 0.033261566162109374, 0.033396736145019534, 0.03324518585205078, 0.03328204727172852, 0.033334270477294925, 0.03320115280151367, 0.0333496322631836, 0.033247230529785156, 0.03326259231567383, 0.03360255813598633, 0.034582527160644534, 0.03474739074707031, 0.03595468902587891, 0.03565875244140625, 0.0350750732421875, 0.034669567108154296, 0.034561023712158204, 0.032950271606445314, 0.033122303009033204, 0.033259521484375, 0.033306625366210936, 0.03315097427368164, 0.03332198333740234, 0.034902015686035154, 0.033470462799072266, 0.03360563278198242, 0.03395993423461914, 0.0332492790222168, 0.03326259231567383, 0.0332677116394043, 0.03329536056518555, 0.03321446228027344, 0.033056766510009765, 0.03319807815551758, 0.03337318420410156, 0.03326668930053711, 0.033242111206054685, 0.033298431396484376, 0.033293312072753906, 0.033225727081298825, 0.033544193267822264, 0.03336908721923828, 0.033172481536865236, 0.03319398498535156, 0.033301502227783206, 0.03379814529418945, 0.03348787307739258, 0.03319500732421875, 0.03313663864135742, 0.033081344604492184, 0.033329151153564454, 0.03325235366821289, 0.03319910430908203, 0.03322777557373047, 0.033247230529785156, 0.03319807815551758, 0.03340492630004883, 0.03320627212524414, 0.033274879455566404, 0.033372161865234375, 0.03323289489746094, 0.033260543823242186, 0.033255424499511715, 0.03319910430908203, 0.03324415969848633, 0.03322880172729492, 0.03321241760253906, 0.033323009490966796, 0.03330252838134766, 0.033263614654541016, 0.033271808624267575, 0.033291263580322264, 0.03544575881958008, 0.03501260757446289, 0.033293312072753906, 0.03331379318237305, 0.03374182510375977, 0.033508350372314456, 0.033247230529785156, 0.03348992156982422, 0.033396736145019534, 0.033293312072753906, 0.033298431396484376, 0.033880062103271484, 0.03331686401367188, 0.03328204727172852, 0.03319091033935547, 0.03326259231567383, 0.033203201293945314, 0.03325235366821289, 0.03324620819091797, 0.033189888000488284, 0.03334041595458984, 0.03336806488037109, 0.03325747299194336, 0.033205249786376956, 0.03320012664794922, 0.033279998779296875, 0.033219585418701174, 0.03327078247070313, 0.03337318420410156, 0.03335270309448242, 0.03339878463745117, 0.033688575744628906, 0.033334270477294925, 0.033463294982910154, 0.03328204727172852, 0.03324415969848633, 0.033290241241455076, 0.03322060775756836, 0.03352166366577149, 0.03333324813842774, 0.03315814590454102, 0.03331071853637695, 0.037222400665283206, 0.03355033493041992, 0.03411251068115234, 0.036615169525146485, 0.035286014556884765, 0.0351539192199707, 0.03492761611938477, 0.034874366760253905, 0.03488870239257812, 0.03473920059204102, 0.034677761077880856, 0.03468288040161133, 0.03477401733398437, 0.03473408126831055, 0.03334143829345703, 0.03432447814941406, 0.03333631896972656, 0.034900993347167966, 0.0331960334777832, 0.033344512939453126, 0.03343667221069336, 0.03343974304199219, 0.03340800094604492, 0.034726913452148435, 0.034743297576904295, 0.035149822235107424, 0.03499520111083984, 0.03478220748901367, 0.03468697738647461, 0.03483135986328125, 0.03460095977783203, 0.034819072723388675, 0.034748416900634765, 0.0347064323425293, 0.03469209671020508, 0.03495731353759766, 0.03472588729858399, 0.03467366409301758, 0.0347586555480957, 0.034781185150146485, 0.03463577651977539, 0.034912254333496096, 0.03487744140625, 0.03461939239501953, 0.034887680053710936, 0.03481702423095703, 0.03485081481933594, 0.034732032775878906, 0.03337420654296875, 0.033503231048583985, 0.034902015686035154, 0.03474943923950195, 0.03363942337036133, 0.03355955123901367, 0.03343462371826172, 0.03333017730712891, 0.03365273666381836, 0.03466649627685547, 0.03466342544555664, 0.03343974304199219, 0.03335372924804687, 0.03335168075561523, 0.03296255874633789, 0.034552833557128904, 0.0333383674621582, 0.03332710266113281, 0.033486846923828126, 0.033296382904052735, 0.033235969543457033, 0.033230846405029296, 0.034283519744873044, 0.03482112121582031, 0.03492659378051758, 0.03473612976074219, 0.034700286865234374, 0.03467264175415039, 0.03457024002075195, 0.03460710525512695, 0.0342999038696289, 0.034328575134277346, 0.03449753570556641, 0.0345456657409668, 0.03472281646728516, 0.034664447784423826, 0.0346798095703125, 0.034740222930908206, 0.034680831909179685, 0.03474227142333984, 0.034715648651123046, 0.034680831909179685, 0.03468902587890625, 0.03437670516967774, 0.03463782501220703, 0.034726913452148435, 0.034738174438476564, 0.034715648651123046, 0.03463270568847656, 0.034661376953125, 0.034628608703613284, 0.0346798095703125, 0.03399782562255859, 0.0348590087890625, 0.034685951232910156, 0.03478732681274414, 0.03468185424804687, 0.03487846374511719, 0.034631679534912106, 0.034735103607177735, 0.0346798095703125, 0.03469004821777344, 0.034651134490966795, 0.0351016960144043, 0.034871295928955076, 0.03468902587890625, 0.03456409454345703]",tokens/s,29.48024774051862,,,,,,MB,999.223296,7390.887936,0.0,6761.218048,6730.975744,s,1,7.04142138671875,7.04142138671875,0.0,7.04142138671875,7.04142138671875,7.04142138671875,7.04142138671875,[7.04142138671875],,kWh,2.8197749715218384e-06,1.5293845842411318e-06,6.92000553599581e-06,1.126916509175878e-05, float16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/gpt-j-6b,EleutherAI/gpt-j-6b,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1264.214016,13227.261952,0.0,12580.814848,12440.741376,s,10,0.8490993576049806,0.08490993576049805,0.001258145422361644,0.08446468734741211,0.0850867317199707,0.08688015022277831,0.08831488502502441,"[0.08867356872558593, 0.08462742614746094, 0.08437590026855468, 0.08446543884277344, 0.08446393585205078, 0.08468819427490235, 0.08453350067138672, 0.08438483428955078, 0.08442422485351563, 0.08446233367919923]",tokens/s,3014.959294305541,kWh,1.0001871793989152e-06,5.480521550827475e-07,5.699017164252045e-06,7.247256498733708e-06,tokens/kWh,35323711.81352973,MB,1264.214016,13227.261952,0.0,12580.814848,12543.514112,s,10,22.313226562500002,2.23132265625,0.0028140680841520567,2.2304586181640627,2.2340583007812502,2.2361828369140624,2.2378824658203125,"[2.238307373046875, 2.22813720703125, 2.230410400390625, 2.22954736328125, 2.230281005859375, 2.233586181640625, 2.2329541015625, 2.2305068359375, 2.23069091796875, 2.22880517578125]",tokens/s,28.234374720991227,kWh,2.6275194256551683e-05,1.4399654623483752e-05,0.00011294552219574736,0.00015362037107578282,tokens/kWh,410101.8605723932,,s,629,23.037688808441136,0.036625896356822195,0.010088523862937467,0.03539865493774414,0.0355887092590332,0.0357597183227539,0.12014645263671875,"[0.03583283233642578, 0.035781631469726564, 0.03575603103637695, 0.035740673065185545, 0.03604991912841797, 0.03590553665161133, 0.035958782196044925, 0.03587891387939453, 0.03614003372192383, 0.03622092819213867, 0.03565574264526367, 0.035652542114257814, 0.03592806243896484, 0.03581542587280274, 0.035716094970703126, 0.0355153923034668, 0.035402751922607424, 0.03529011154174805, 0.03524505615234375, 0.03530035018920898, 0.0353177604675293, 0.03530956649780274, 0.03539148712158203, 0.03531980895996094, 0.035294273376464846, 0.03532486343383789, 0.03533004760742187, 0.03529523086547852, 0.03530956649780274, 0.0354150390625, 0.035361793518066405, 0.03536281585693359, 0.035318782806396484, 0.03534643173217773, 0.035389438629150394, 0.03535974502563476, 0.03539251327514648, 0.03541196823120117, 0.03542118453979492, 0.035492862701416016, 0.03539251327514648, 0.03540486526489258, 0.035404735565185544, 0.03539251327514648, 0.03539353561401367, 0.03544575881958008, 0.03550003051757813, 0.03545600128173828, 0.03542835235595703, 0.03544166564941406, 0.035460094451904296, 0.03545907211303711, 0.03546931076049804, 0.035492862701416016, 0.03547238540649414, 0.03549603271484375, 0.03546825790405273, 0.035510208129882814, 0.035522560119628906, 0.035501056671142575, 0.03548364639282227, 0.03548876953125, 0.12039679718017578, 0.035179519653320314, 0.035410945892333984, 0.03536281585693359, 0.03516723251342774, 0.03520614242553711, 0.035165184020996096, 0.03524915313720703, 0.03525120162963867, 0.035198974609375, 0.035214336395263675, 0.035244033813476565, 0.035248126983642575, 0.03528396987915039, 0.03524198532104492, 0.035269630432128905, 0.03527993774414063, 0.03529721450805664, 0.03529324722290039, 0.03528185653686523, 0.03525939178466797, 0.03537926483154297, 0.03532896041870117, 0.03530137634277344, 0.03527167892456055, 0.03539251327514648, 0.03523583984375, 0.03529523086547852, 0.03535257720947266, 0.035340286254882815, 0.035350528717041016, 0.03535974502563476, 0.03536896133422852, 0.0353966064453125, 0.035419136047363284, 0.03535974502563476, 0.03536793518066406, 0.035386367797851564, 0.035361793518066405, 0.03539865493774414, 0.035369983673095705, 0.03538431930541992, 0.035337215423583986, 0.035323902130126955, 0.035386367797851564, 0.03542937469482422, 0.0354703369140625, 0.03544780731201172, 0.03542534255981445, 0.035431358337402345, 0.035517440795898435, 0.03546726226806641, 0.035422206878662106, 0.03547340774536133, 0.03543142318725586, 0.03545087814331055, 0.035509246826171875, 0.035566593170166014, 0.0356495361328125, 0.03560038375854492, 0.03547750473022461, 0.035604480743408204, 0.035454975128173825, 0.12013260650634766, 0.0352174072265625, 0.03514572906494141, 0.03525529479980469, 0.03515801620483398, 0.035538944244384765, 0.03629875183105469, 0.03563622283935547, 0.03539763259887695, 0.0354334716796875, 0.035348480224609374, 0.035227649688720705, 0.035340286254882815, 0.03527167892456055, 0.03524710464477539, 0.03527577590942383, 0.03525222396850586, 0.03531161499023437, 0.03530649566650391, 0.035330078125, 0.035378143310546874, 0.03527372741699219, 0.03533824157714844, 0.03544371032714844, 0.035343360900878903, 0.035266559600830076, 0.035302398681640625, 0.03528806304931641, 0.035332096099853515, 0.035342334747314456, 0.035304447174072266, 0.035356670379638674, 0.03536281585693359, 0.03535257720947266, 0.03535059356689453, 0.035347423553466796, 0.035366878509521485, 0.03537311935424805, 0.03537299346923828, 0.03540172958374024, 0.035422206878662106, 0.03545087814331055, 0.03539868927001953, 0.03544879913330078, 0.03544473648071289, 0.03541299057006836, 0.03544063949584961, 0.03547238540649414, 0.03547443389892578, 0.03546419143676758, 0.03542937469482422, 0.03553792190551758, 0.03548262405395508, 0.03542323303222656, 0.035454975128173825, 0.03551027297973633, 0.03550003051757813, 0.035487743377685545, 0.03549593734741211, 0.03553484725952148, 0.03552972793579102, 0.035517440795898435, 0.03545907211303711, 0.12021868896484375, 0.0352501106262207, 0.03518566513061523, 0.03519692611694336, 0.035165184020996096, 0.035209217071533204, 0.03517337417602539, 0.03537510299682617, 0.03526860809326172, 0.03525632095336914, 0.03539148712158203, 0.03525939178466797, 0.03535871887207031, 0.035334144592285156, 0.03526144027709961, 0.035266559600830076, 0.03522150421142578, 0.035310592651367184, 0.035297279357910154, 0.03525222396850586, 0.03526863861083984, 0.035321887969970704, 0.03527980804443359, 0.03530342483520508, 0.03530547332763672, 0.035296257019042966, 0.03528806304931641, 0.035381248474121094, 0.03531468963623047, 0.03535871887207031, 0.03534950256347656, 0.03528806304931641, 0.03539353561401367, 0.035351551055908204, 0.035337215423583986, 0.03547955322265625, 0.035829761505126956, 0.035714046478271484, 0.03558297729492187, 0.0353546257019043, 0.035535873413085936, 0.03547443389892578, 0.035484672546386715, 0.03539046478271484, 0.035576831817626955, 0.03542940902709961, 0.03540681457519531, 0.035442752838134764, 0.035406784057617186, 0.035514366149902346, 0.035434497833251956, 0.0355860481262207, 0.03544371032714844, 0.035448833465576174, 0.0354334716796875, 0.035490814208984374, 0.03545292663574219, 0.03544780731201172, 0.03544268798828125, 0.03547443389892578, 0.035560447692871096, 0.03565055847167969, 0.0354856948852539, 0.12045721435546874, 0.03539251327514648, 0.03519180679321289, 0.03522048187255859, 0.03534649658203125, 0.03533715057373047, 0.03517337417602539, 0.035227649688720705, 0.035286014556884765, 0.03526041412353516, 0.03527884674072266, 0.03525836944580078, 0.03524300765991211, 0.03541196823120117, 0.035931137084960936, 0.035566593170166014, 0.03545292663574219, 0.035372032165527346, 0.03530854415893555, 0.03529011154174805, 0.03528499221801758, 0.03526860809326172, 0.035331134796142576, 0.03539244842529297, 0.03535871887207031, 0.03532287979125977, 0.03527884674072266, 0.0353177604675293, 0.03532799911499023, 0.0352911376953125, 0.035348480224609374, 0.03534643173217773, 0.03541299057006836, 0.03537715148925781, 0.03535769653320313, 0.03540582275390625, 0.03539865493774414, 0.0353361930847168, 0.03538022232055664, 0.03537510299682617, 0.03553177642822265, 0.03538534545898438, 0.03544780731201172, 0.03545087814331055, 0.03537715148925781, 0.03542732620239258, 0.03540582275390625, 0.035507198333740234, 0.03542323303222656, 0.03546623992919922, 0.03546828842163086, 0.035544063568115236, 0.035507198333740234, 0.03544473648071289, 0.035438591003417966, 0.035490814208984374, 0.035542015075683595, 0.035520511627197264, 0.035490814208984374, 0.03550822448730469, 0.03551846313476562, 0.03548160171508789, 0.03547238540649414, 0.12121804809570312, 0.03549491119384766, 0.03529216003417969, 0.03521023941040039, 0.03526553726196289, 0.03518873596191406, 0.03520102310180664, 0.03524716949462891, 0.035286975860595704, 0.03539148712158203, 0.03545395278930664, 0.03530752182006836, 0.035410945892333984, 0.03553279876708984, 0.035432449340820314, 0.03535974502563476, 0.03548672103881836, 0.03534438323974609, 0.03530342483520508, 0.03533004760742187, 0.03529318237304688, 0.035326976776123044, 0.03534131240844727, 0.035404800415039066, 0.03543756866455078, 0.035323902130126955, 0.03533107376098633, 0.03531468963623047, 0.035343360900878903, 0.0354252815246582, 0.035345409393310545, 0.03537408065795898, 0.035356670379638674, 0.03541401672363281, 0.035369983673095705, 0.03537919998168945, 0.03538534545898438, 0.03552972793579102, 0.03558707046508789, 0.0355860481262207, 0.03559628677368164, 0.03551232147216797, 0.03546112060546875, 0.035514366149902346, 0.03541708755493164, 0.0354150390625, 0.03546623992919922, 0.03548160171508789, 0.0354068489074707, 0.03543040084838867, 0.03563827133178711, 0.035773441314697264, 0.035681278228759765, 0.03564031982421875, 0.03564441680908203, 0.0357017936706543, 0.03569660949707031, 0.03567103958129883, 0.03546726226806641, 0.035507198333740234, 0.03549184036254883, 0.03554099273681641, 0.035530750274658206, 0.12027801513671875, 0.03520716857910156, 0.03512934494018555, 0.035279872894287106, 0.03538739013671875, 0.03525529479980469, 0.03553279876708984, 0.03547340774536133, 0.03544582366943359, 0.035338176727294925, 0.035195903778076174, 0.03523891067504883, 0.03531161499023437, 0.03523379135131836, 0.03523993682861328, 0.03520409774780273, 0.035197952270507815, 0.03529216003417969, 0.03539046478271484, 0.03530649566650391, 0.03521228790283203, 0.035221569061279295, 0.03524396896362305, 0.03531468963623047, 0.03533004760742187, 0.03544371032714844, 0.035800064086914066, 0.036094974517822266, 0.035707904815673826, 0.03559526443481445, 0.03550515365600586, 0.03552460861206055, 0.03552972793579102, 0.03557580947875977, 0.03543552017211914, 0.035356670379638674, 0.035762176513671876, 0.035530750274658206, 0.03544166564941406, 0.03543040084838867, 0.03543552017211914, 0.03538227081298828, 0.03547340774536133, 0.03542118453979492, 0.03541401672363281, 0.035487743377685545, 0.035451904296875, 0.03543142318725586, 0.035454975128173825, 0.035434497833251956, 0.03544575881958008, 0.035469375610351565, 0.03541907119750977, 0.03546726226806641, 0.03547238540649414, 0.03547340774536133, 0.035609600067138675, 0.03549184036254883, 0.035544063568115236, 0.035576831817626955, 0.03573350524902344, 0.03563417434692383, 0.03577753448486328, 0.12013465881347657, 0.035314750671386716, 0.035409854888916015, 0.03536076736450195, 0.03513651275634765, 0.03531468963623047, 0.035402751922607424, 0.035394561767578124, 0.03528192138671875, 0.03527990341186524, 0.03526857757568359, 0.03521331024169922, 0.035264511108398434, 0.03525529479980469, 0.035269630432128905, 0.035335166931152344, 0.03529011154174805, 0.03530854415893555, 0.0352542724609375, 0.03530854415893555, 0.035286014556884765, 0.035269630432128905, 0.03529523086547852, 0.0353177604675293, 0.03529216003417969, 0.03548262405395508, 0.03530140686035156, 0.03526243209838867, 0.03533926391601563, 0.03542732620239258, 0.035329025268554685, 0.035404800415039066, 0.03532287979125977, 0.03539353561401367, 0.035348480224609374, 0.03536076736450195, 0.035373054504394534, 0.035379230499267576, 0.03577443313598633, 0.035656734466552736, 0.03546928024291992, 0.03544371032714844, 0.03540991973876953, 0.03536588668823242, 0.03548672103881836, 0.035434497833251956, 0.03544473648071289, 0.035448833465576174, 0.03541401672363281, 0.03544678497314453, 0.035432449340820314, 0.0354856948852539, 0.03550310516357422, 0.03554508972167969, 0.03555327987670898, 0.03546214294433594, 0.03566291046142578, 0.03582252883911133, 0.035620864868164064, 0.03547244644165039, 0.03550099182128906, 0.035522560119628906, 0.03558707046508789, 0.12015103912353516, 0.03515497589111328, 0.03517744064331055, 0.03520512008666992, 0.03521945571899414, 0.03515596771240234, 0.035195903778076174, 0.03522150421142578, 0.035244033813476565, 0.03525836944580078, 0.03526144027709961, 0.03528704071044922, 0.03525734329223633, 0.03520000076293945, 0.03523891067504883, 0.03523891067504883, 0.03532287979125977, 0.035286014556884765, 0.035280895233154294, 0.03526758575439453, 0.035332096099853515, 0.03606841659545899, 0.035710975646972655, 0.03552048110961914, 0.03529724884033203, 0.03546623992919922, 0.03528908920288086, 0.035356670379638674, 0.03533824157714844, 0.03530649566650391, 0.03533824157714844, 0.03539049530029297, 0.03532489776611328, 0.03535974502563476, 0.035343360900878903, 0.03596905517578125, 0.03580924987792969, 0.0355788803100586, 0.03559833526611328, 0.035609600067138675, 0.035492862701416016, 0.035348480224609374, 0.03539251327514648, 0.0354252815246582, 0.035402751922607424, 0.03552665710449219, 0.035434497833251956, 0.035536895751953124, 0.035422206878662106, 0.035419136047363284, 0.035418113708496096, 0.0354488639831543, 0.03547849655151367, 0.03548262405395508, 0.03546623992919922, 0.035471359252929685, 0.03547548675537109, 0.03547849655151367, 0.03542732620239258, 0.035476478576660156, 0.03548057556152344, 0.035506175994873046, 0.035509246826171875, 0.12015923309326172, 0.03518259048461914, 0.03516928100585937, 0.03525120162963867, 0.03515801620483398, 0.03522355270385742, 0.035163135528564454, 0.035332096099853515, 0.035304447174072266, 0.03525734329223633, 0.03528192138671875, 0.035194881439208986, 0.03523788833618164, 0.035272705078125, 0.035264511108398434, 0.03527782440185547, 0.035250175476074216, 0.035269630432128905, 0.03528806304931641, 0.035315711975097655, 0.03534438323974609, 0.035506175994873046, 0.03546217727661133, 0.03534640121459961, 0.03529830551147461, 0.03532492828369141, 0.035378177642822264, 0.03538534545898438, 0.03537510299682617, 0.035372032165527346, 0.035340286254882815, 0.035350528717041016, 0.03542835235595703, 0.03539148712158203, 0.0354068489074707, 0.03539868927001953, 0.03541705703735352, 0.03536588668823242, 0.03535257720947266, 0.03558297729492187, 0.03538534545898438, 0.03547443389892578, 0.03540582275390625, 0.03537100982666016, 0.03559731292724609, 0.03542835235595703, 0.03541916656494141, 0.03542729568481445, 0.035419136047363284, 0.035441696166992186, 0.035412960052490235, 0.03540172958374024, 0.035514366149902346, 0.03549388885498047, 0.035517440795898435, 0.03546112060546875, 0.035490814208984374, 0.0355153923034668, 0.03551027297973633, 0.035471359252929685, 0.03546931076049804, 0.0354969596862793, 0.035476478576660156]",tokens/s,27.303086053038893,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,stabilityai/stablelm-2-1_6b,stabilityai/stablelm-2-1_6b,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1276.993536,4276.617216,0.0,3632.267264,3408.132096,s,10,0.23940703964233398,0.0239407039642334,0.0003155813142295948,0.023847408294677734,0.02398629741668701,0.024434684658050536,0.02479339445114136,"[0.024883071899414064, 0.023886655807495116, 0.023772607803344725, 0.023813919067382814, 0.02384752082824707, 0.02381644821166992, 0.023857664108276368, 0.023819103240966796, 0.02386275291442871, 0.0238472957611084]",tokens/s,10693.085733086853,kWh,2.839145672499624e-07,1.5557130726576287e-07,1.4168026825865747e-06,1.8562885571023002e-06,tokens/kWh,137909593.32293713,MB,1276.993536,4278.714368,0.0,3632.267264,3474.14784,s,10,12.64006494140625,1.264006494140625,0.012221827649893114,1.2609061279296876,1.2754159545898438,1.2825172424316407,1.2881982727050783,"[1.2896185302734375, 1.2725411376953124, 1.2540771484375, 1.255328369140625, 1.2603609619140625, 1.2468157958984376, 1.273837890625, 1.25360791015625, 1.2614512939453124, 1.2724259033203125]",tokens/s,49.8415160776785,kWh,1.4844918173666186e-05,8.134760645182703e-06,3.992906372741369e-05,6.290874254626257e-05,tokens/kWh,1001450.6322975749,,s,629,12.83316420936584,0.020402486819341568,0.002910507754959045,0.01984000015258789,0.020758118057250977,0.020815462493896486,0.043985878601074224,"[0.022249343872070313, 0.020015104293823242, 0.019563520431518554, 0.020341760635375978, 0.02031718444824219, 0.020344831466674804, 0.01965260887145996, 0.019603456497192383, 0.019529727935791014, 0.019575807571411134, 0.019569664001464843, 0.01967001533508301, 0.019629056930541993, 0.01966592025756836, 0.0198656005859375, 0.01978982353210449, 0.020781055450439453, 0.020579328536987306, 0.020864000320434572, 0.020596736907958983, 0.019772415161132813, 0.02064588737487793, 0.01966592025756836, 0.01960960006713867, 0.01984921646118164, 0.019788799285888673, 0.019784704208374023, 0.020091903686523437, 0.020955135345458984, 0.02109644889831543, 0.020981760025024415, 0.020800512313842775, 0.020802560806274413, 0.020745216369628908, 0.020780031204223632, 0.02069811248779297, 0.02072166442871094, 0.020743200302124024, 0.020793312072753905, 0.020742143630981445, 0.020668415069580077, 0.020701183319091796, 0.020806655883789063, 0.020725759506225586, 0.020715520858764647, 0.020736000061035157, 0.020770816802978515, 0.02070425605773926, 0.020690944671630858, 0.020757503509521484, 0.020763647079467772, 0.020786176681518553, 0.020700159072875975, 0.020742143630981445, 0.02104012870788574, 0.020815872192382814, 0.020814847946166993, 0.020690944671630858, 0.020790271759033203, 0.020765695571899414, 0.02070528030395508, 0.020719615936279297, 0.045159423828125, 0.020746240615844725, 0.02069606399536133, 0.020760576248168947, 0.02088140869140625, 0.020649984359741212, 0.020766719818115235, 0.020776960372924806, 0.02083635139465332, 0.020743167877197266, 0.020767744064331056, 0.02075257682800293, 0.02082899284362793, 0.020764671325683593, 0.02082713508605957, 0.020784128189086915, 0.02081279945373535, 0.02082918357849121, 0.02044108772277832, 0.020757503509521484, 0.020771839141845702, 0.020753408432006838, 0.020695039749145508, 0.020798528671264648, 0.02076460838317871, 0.020791296005249024, 0.019911680221557617, 0.019793920516967774, 0.019819520950317384, 0.019708927154541016, 0.0198604793548584, 0.020007936477661133, 0.01967513656616211, 0.01958822441101074, 0.019557247161865233, 0.019571712493896484, 0.019773439407348634, 0.019877887725830077, 0.01982054328918457, 0.019959808349609375, 0.019786752700805665, 0.0198656005859375, 0.019986431121826173, 0.01989836883544922, 0.019817472457885742, 0.01984819221496582, 0.019805183410644533, 0.01986355209350586, 0.019812416076660157, 0.01974060821533203, 0.019687423706054686, 0.019771392822265626, 0.019572736740112305, 0.019579904556274414, 0.01965363121032715, 0.019728384017944335, 0.019862527847290038, 0.019963903427124022, 0.01982054328918457, 0.019800064086914062, 0.01985740852355957, 0.01984000015258789, 0.01984511947631836, 0.04400537490844727, 0.0198154239654541, 0.01982464027404785, 0.019786752700805665, 0.019811328887939454, 0.01980723190307617, 0.019862527847290038, 0.01985843276977539, 0.019908607482910155, 0.01984921646118164, 0.01982361602783203, 0.01977446365356445, 0.01985945510864258, 0.019891199111938478, 0.01987276840209961, 0.019899391174316407, 0.019770368576049805, 0.01982156753540039, 0.019805183410644533, 0.019834880828857423, 0.019784704208374023, 0.01988403129577637, 0.0198656005859375, 0.019901439666748046, 0.019812351226806642, 0.01984000015258789, 0.01984819221496582, 0.019813375473022463, 0.01984000015258789, 0.01983692741394043, 0.019786752700805665, 0.019970048904418947, 0.019812351226806642, 0.01986662483215332, 0.01986355209350586, 0.019931135177612306, 0.019812351226806642, 0.01982156753540039, 0.019843072891235353, 0.02002841567993164, 0.01993123245239258, 0.01985424041748047, 0.01985228729248047, 0.01983078384399414, 0.01983795166015625, 0.01982259178161621, 0.01983078384399414, 0.019914751052856446, 0.019779584884643556, 0.01983283233642578, 0.01985945510864258, 0.0197857608795166, 0.019861503601074217, 0.019742687225341796, 0.01984921646118164, 0.020208639144897463, 0.02145484733581543, 0.02071347236633301, 0.020761600494384767, 0.019785728454589844, 0.019737600326538086, 0.0198175048828125, 0.019766239166259764, 0.04402179336547852, 0.019805152893066405, 0.01984921646118164, 0.01983692741394043, 0.019772415161132813, 0.019819520950317384, 0.019757055282592775, 0.019852352142333985, 0.019805152893066405, 0.019734495162963866, 0.01985024070739746, 0.019764223098754884, 0.019696640014648437, 0.01979801559448242, 0.019714048385620117, 0.019794944763183595, 0.019885055541992186, 0.019758079528808595, 0.019788799285888673, 0.019794944763183595, 0.019801088333129883, 0.019805183410644533, 0.019766271591186522, 0.019842048645019532, 0.019788799285888673, 0.019826688766479493, 0.01982259178161621, 0.01983180809020996, 0.019755008697509766, 0.019811391830444336, 0.01973040008544922, 0.019833824157714845, 0.019803136825561524, 0.019776512145996093, 0.019736576080322265, 0.01981955146789551, 0.019859424591064455, 0.019834880828857423, 0.01979804801940918, 0.019857376098632813, 0.01979084777832031, 0.01982975959777832, 0.019781631469726564, 0.01983283233642578, 0.01964339256286621, 0.01979084777832031, 0.022689792633056642, 0.020799488067626954, 0.020588544845581053, 0.020601856231689454, 0.020365312576293947, 0.020551776885986327, 0.01959516716003418, 0.0195666561126709, 0.01953068733215332, 0.01961369514465332, 0.01965670394897461, 0.02033158493041992, 0.020458431243896485, 0.020420608520507814, 0.020403200149536133, 0.019582975387573243, 0.019564544677734375, 0.043911167144775394, 0.019620895385742188, 0.019941343307495116, 0.020537343978881836, 0.020418560028076172, 0.020332544326782227, 0.020331520080566406, 0.02084556770324707, 0.021608448028564452, 0.020546560287475587, 0.020493312835693358, 0.020426752090454102, 0.02029363250732422, 0.020287488937377928, 0.020379648208618165, 0.019677343368530272, 0.019605472564697267, 0.019594112396240235, 0.019537919998168944, 0.019654687881469728, 0.019588064193725586, 0.019498016357421873, 0.019522560119628905, 0.019624927520751952, 0.019579904556274414, 0.01979084777832031, 0.019851264953613282, 0.02057318305969238, 0.020373504638671876, 0.020539392471313478, 0.02059878349304199, 0.020582399368286132, 0.02066739273071289, 0.02064793586730957, 0.019825664520263672, 0.019756032943725587, 0.019766271591186522, 0.019744768142700195, 0.01978265571594238, 0.019847263336181642, 0.019767200469970703, 0.01983078384399414, 0.01983283233642578, 0.01982464027404785, 0.01982054328918457, 0.019787776947021486, 0.01987174415588379, 0.019927040100097656, 0.019817472457885742, 0.019725311279296876, 0.01985740852355957, 0.01981439971923828, 0.019811328887939454, 0.019812416076660157, 0.019889087677001954, 0.019825664520263672, 0.019861503601074217, 0.01982361602783203, 0.019802112579345704, 0.019766271591186522, 0.019804159164428712, 0.019800064086914062, 0.019944448471069336, 0.04413951873779297, 0.0198287353515625, 0.01985536003112793, 0.01982464027404785, 0.019894271850585937, 0.01990656089782715, 0.01986867141723633, 0.019862527847290038, 0.019851264953613282, 0.019772415161132813, 0.019817472457885742, 0.0198154239654541, 0.01982259178161621, 0.020009983062744142, 0.0198604793548584, 0.01986662483215332, 0.019750911712646483, 0.01979497528076172, 0.01986147117614746, 0.019714048385620117, 0.0196495361328125, 0.019610624313354492, 0.019483648300170898, 0.01962188720703125, 0.019525632858276368, 0.019851264953613282, 0.019862560272216796, 0.019749919891357423, 0.01975391960144043, 0.019927040100097656, 0.01982054328918457, 0.019801088333129883, 0.019776512145996093, 0.019765247344970704, 0.019876863479614256, 0.019769344329833984, 0.019768320083618163, 0.019588096618652344, 0.01967513656616211, 0.019644416809082032, 0.019477664947509767, 0.01954595184326172, 0.019803136825561524, 0.019951616287231445, 0.019761152267456054, 0.019745792388916016, 0.019893247604370116, 0.01980723190307617, 0.019903488159179687, 0.019723264694213868, 0.019779584884643556, 0.019783679962158202, 0.019759103775024413, 0.019817472457885742, 0.01982156753540039, 0.01984102439880371, 0.019804159164428712, 0.01979801559448242, 0.01984716796875, 0.019795967102050782, 0.01982054328918457, 0.019805183410644533, 0.019748895645141602, 0.04502729415893555, 0.021325824737548828, 0.01983283233642578, 0.020133888244628906, 0.020566015243530272, 0.020590591430664062, 0.020588544845581053, 0.020586496353149415, 0.02062848091125488, 0.020549631118774413, 0.019776512145996093, 0.02006425666809082, 0.02060697555541992, 0.020547584533691408, 0.020593664169311524, 0.019853311538696287, 0.019723264694213868, 0.019799072265625, 0.019860448837280272, 0.019984384536743165, 0.019766271591186522, 0.019819520950317384, 0.01982361602783203, 0.019769344329833984, 0.019810304641723633, 0.019779584884643556, 0.01979084777832031, 0.019793920516967774, 0.020478975296020507, 0.02063564872741699, 0.020567039489746093, 0.020756479263305663, 0.02060697555541992, 0.020619264602661135, 0.02061311912536621, 0.020594688415527345, 0.020682752609252928, 0.020731903076171874, 0.01982464027404785, 0.01985638427734375, 0.01980723190307617, 0.0198154239654541, 0.0198154239654541, 0.019862527847290038, 0.020271104812622072, 0.02064896011352539, 0.019817472457885742, 0.019869695663452147, 0.019742719650268553, 0.02017791938781738, 0.019908607482910155, 0.01989017677307129, 0.019908607482910155, 0.020599807739257812, 0.020782079696655274, 0.02072166442871094, 0.02062950325012207, 0.020734975814819336, 0.02065407943725586, 0.019766271591186522, 0.019779584884643556, 0.019804159164428712, 0.019747840881347657, 0.04393574523925781, 0.01978985595703125, 0.019811296463012697, 0.019843072891235353, 0.019775487899780272, 0.020611072540283205, 0.020874271392822264, 0.021965791702270508, 0.020808704376220705, 0.020626432418823244, 0.01979903984069824, 0.019760128021240234, 0.019763200759887696, 0.020376575469970702, 0.02089779281616211, 0.019968000411987305, 0.019762176513671875, 0.019777536392211914, 0.019748863220214845, 0.019779584884643556, 0.01964646339416504, 0.020568063735961914, 0.020274175643920898, 0.019663871765136717, 0.01945497512817383, 0.01962598419189453, 0.01960857582092285, 0.01944166374206543, 0.019507200241088866, 0.019579904556274414, 0.019503103256225587, 0.019562496185302734, 0.019533824920654298, 0.019501056671142578, 0.019543039321899415, 0.020529151916503906, 0.020356096267700196, 0.020539392471313478, 0.020371456146240235, 0.020392959594726562, 0.02025164794921875, 0.020478975296020507, 0.02026188850402832, 0.019555328369140625, 0.01945907211303711, 0.01946419143676758, 0.019521535873413084, 0.019578880310058593, 0.019582975387573243, 0.019764320373535156, 0.019569568634033203, 0.01946531105041504, 0.019436447143554688, 0.019516416549682617, 0.019520511627197267, 0.019545087814331053, 0.019494911193847657, 0.019677183151245118, 0.019548160552978516, 0.019747840881347657, 0.01948569679260254, 0.01979084777832031, 0.019869695663452147, 0.044025856018066405, 0.019795007705688476, 0.01993516731262207, 0.01983795166015625, 0.02005504035949707, 0.020685823440551757, 0.020585472106933594, 0.020668415069580077, 0.020575231552124023, 0.02068070411682129, 0.02066329574584961, 0.02059775924682617, 0.02051584053039551, 0.02081279945373535, 0.02035916709899902, 0.02065100860595703, 0.020549631118774413, 0.020612096786499022, 0.020593664169311524, 0.020736000061035157, 0.020728832244873048, 0.01984716796875, 0.019811328887939454, 0.0197806396484375, 0.019838943481445312, 0.019773439407348634, 0.019714048385620117, 0.019803136825561524, 0.019752960205078125, 0.019732479095458985, 0.01978265571594238, 0.0196997127532959, 0.019718143463134767, 0.01979903984069824, 0.0198154239654541, 0.019779584884643556, 0.019765247344970704, 0.019768320083618163, 0.01978265571594238, 0.019875904083251954, 0.019820480346679686, 0.01983692741394043, 0.01981644821166992, 0.01982771110534668, 0.019808256149291992, 0.01982464027404785, 0.019739648818969727, 0.019901439666748046, 0.01977446365356445, 0.020001792907714845, 0.019922943115234376, 0.019770368576049805, 0.01980723190307617, 0.019795967102050782, 0.019776512145996093, 0.01984819221496582, 0.0198287353515625, 0.019784704208374023, 0.01985843276977539, 0.019765247344970704, 0.019794944763183595, 0.019701759338378907, 0.019833856582641602, 0.04410367965698242, 0.01983590316772461, 0.019777536392211914, 0.019811328887939454, 0.01982054328918457, 0.019720191955566405, 0.01982156753540039, 0.02002841567993164, 0.019811328887939454, 0.01982156753540039, 0.019927040100097656, 0.01990656089782715, 0.019902463912963866, 0.02149990463256836, 0.020798463821411133, 0.020840448379516603, 0.020807680130004884, 0.020690944671630858, 0.019755008697509766, 0.019883007049560548, 0.02003660774230957, 0.019705856323242187, 0.019738624572753907, 0.019751935958862304, 0.019610624313354492, 0.020142080307006836, 0.02047091293334961, 0.020378496170043944, 0.01985228729248047, 0.019851264953613282, 0.02017791938781738, 0.020739072799682616, 0.01980723190307617, 0.019818496704101563, 0.019795967102050782, 0.020386816024780274, 0.02063155174255371, 0.020718591690063477, 0.02122854423522949, 0.020789247512817383, 0.020694015502929687, 0.02069811248779297, 0.020584447860717774, 0.020677631378173827, 0.02068377685546875, 0.020736000061035157, 0.020658176422119142, 0.01978265571594238, 0.019891231536865235, 0.019780576705932616, 0.019938304901123048, 0.019636224746704102, 0.01984409523010254, 0.019760128021240234, 0.019738624572753907, 0.01965977668762207, 0.02002339172363281, 0.020392864227294923, 0.020510719299316405, 0.020593664169311524, 0.02062745666503906, 0.020585535049438476, 0.020603872299194335]",tokens/s,49.01363293870625,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,togethercomputer/RedPajama-INCITE-Base-3B-v1,togethercomputer/RedPajama-INCITE-Base-3B-v1,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1332.113408,6833.045504,0.0,6186.5984,6021.011456,s,10,0.44095139312744136,0.04409513931274414,0.0006835938958904407,0.04387639999389649,0.04432553596496582,0.04522236881256103,0.045939835090637204,"[0.04611920166015625, 0.043928768157958986, 0.04380425643920898, 0.043882816314697266, 0.04383878326416016, 0.043926464080810544, 0.04376326370239258, 0.04369161605834961, 0.04412623977661133, 0.043869983673095705]",tokens/s,5805.628556570004,kWh,5.210874349652e-07,2.854079358726647e-07,2.711102266775277e-06,3.517597637613142e-06,tokens/kWh,72776942.2126711,MB,1332.113408,6833.045504,0.0,6186.5984,6148.783616,s,10,16.97177770996094,1.697177770996094,0.021310645414094363,1.7024412841796877,1.7154104370117187,1.7180841247558594,1.720223074951172,"[1.7207578125, 1.6907777099609376, 1.693623291015625, 1.7143214111328124, 1.69441259765625, 1.712847412109375, 1.65232470703125, 1.667426513671875, 1.710469970703125, 1.7148162841796875]",tokens/s,37.12044847430717,kWh,1.9446078882868594e-05,1.0654387857219677e-05,6.352211738642479e-05,9.362258412651305e-05,tokens/kWh,672914.5599620229,,s,629,17.33982517242429,0.027567289622296216,0.005280602665381202,0.027128896713256835,0.02738483238220215,0.02763550682067871,0.07035876037597656,"[0.027814912796020507, 0.02897407913208008, 0.027517951965332032, 0.02709401512145996, 0.027062271118164064, 0.02692095947265625, 0.027123712539672853, 0.027019264221191407, 0.026983423233032225, 0.027072511672973632, 0.027007999420166014, 0.02690355110168457, 0.02672230339050293, 0.02670694351196289, 0.02673766326904297, 0.027677696228027345, 0.02726092720031738, 0.0273623046875, 0.027219968795776366, 0.02878361511230469, 0.032279552459716795, 0.027909120559692382, 0.02760601615905762, 0.027209728240966798, 0.0271329288482666, 0.027197439193725585, 0.026824703216552736, 0.02694963264465332, 0.027001855850219726, 0.027012096405029298, 0.027295743942260742, 0.027142143249511717, 0.02691993522644043, 0.026977279663085937, 0.026838016510009766, 0.026857664108276367, 0.02691155242919922, 0.02716160011291504, 0.027357183456420898, 0.027151359558105468, 0.027107328414916993, 0.027250688552856447, 0.027076608657836915, 0.02711756706237793, 0.027217920303344727, 0.027233280181884766, 0.027640832901000976, 0.02717900848388672, 0.027108352661132814, 0.027214847564697265, 0.027175935745239257, 0.02786720085144043, 0.027667392730712892, 0.02740838432312012, 0.02715545654296875, 0.027247615814208984, 0.027214847564697265, 0.027157503128051756, 0.027222015380859374, 0.02721177673339844, 0.027230207443237304, 0.02729267120361328, 0.0716410903930664, 0.027148288726806642, 0.027143167495727538, 0.027241472244262696, 0.027215871810913086, 0.0271278076171875, 0.02702444839477539, 0.027300800323486328, 0.027173887252807616, 0.02731929588317871, 0.02672230339050293, 0.026162176132202147, 0.026059776306152343, 0.026067968368530273, 0.026053632736206055, 0.026181631088256836, 0.026039295196533203, 0.02608332824707031, 0.026137599945068358, 0.02611712074279785, 0.026426368713378907, 0.026048511505126954, 0.026059776306152343, 0.026194944381713867, 0.02593894386291504, 0.02689955139160156, 0.02720150375366211, 0.027190208435058594, 0.02713599967956543, 0.027258880615234377, 0.02709708786010742, 0.027278335571289062, 0.02710937690734863, 0.0271329288482666, 0.02712166404724121, 0.02709503936767578, 0.02669158363342285, 0.02687283134460449, 0.027100160598754884, 0.027229183197021483, 0.027042816162109375, 0.027115520477294923, 0.02718720054626465, 0.027123712539672853, 0.02719545555114746, 0.027174848556518555, 0.027092992782592775, 0.027244543075561522, 0.027042816162109375, 0.027146240234375, 0.027007999420166014, 0.02710527992248535, 0.026013696670532226, 0.025879583358764648, 0.026879968643188475, 0.026918912887573244, 0.026811391830444335, 0.02719436836242676, 0.02687385559082031, 0.027140096664428712, 0.026852352142333984, 0.026846208572387696, 0.026785791397094725, 0.07029145812988281, 0.025659391403198242, 0.02570751953125, 0.025561088562011718, 0.025758880615234375, 0.025675615310668944, 0.02568297576904297, 0.025656288146972656, 0.025649152755737304, 0.025643007278442383, 0.025819135665893556, 0.02574745559692383, 0.025999359130859375, 0.025989120483398437, 0.0267458553314209, 0.028062719345092774, 0.02757427215576172, 0.02739200019836426, 0.02715443229675293, 0.027196416854858397, 0.02736950492858887, 0.02717283248901367, 0.027098112106323242, 0.027247615814208984, 0.027209728240966798, 0.027240447998046875, 0.02716057586669922, 0.02712063980102539, 0.027379711151123046, 0.027213823318481444, 0.027706432342529296, 0.027205568313598633, 0.02749951934814453, 0.027165695190429686, 0.027131904602050783, 0.02712985610961914, 0.027229183197021483, 0.027299840927124022, 0.027348991394042968, 0.027244543075561522, 0.027099136352539063, 0.027131904602050783, 0.027181055068969725, 0.027182079315185546, 0.027140096664428712, 0.02709401512145996, 0.02716262435913086, 0.02730188751220703, 0.027200511932373047, 0.02714521598815918, 0.027197439193725585, 0.02694758415222168, 0.02679193687438965, 0.02688102340698242, 0.027244543075561522, 0.027234304428100587, 0.027114496231079102, 0.026802175521850585, 0.026910720825195314, 0.027224063873291016, 0.02705510330200195, 0.02711039924621582, 0.027191295623779296, 0.07150185394287109, 0.027250656127929686, 0.028490751266479493, 0.02711859130859375, 0.027174911499023437, 0.02716979217529297, 0.027059200286865235, 0.02719539260864258, 0.027131904602050783, 0.027218944549560548, 0.02720256042480469, 0.027193344116210938, 0.02713907241821289, 0.027201536178588868, 0.02711347198486328, 0.02728447914123535, 0.0271011848449707, 0.027244543075561522, 0.02709503936767578, 0.027166719436645507, 0.02709503936767578, 0.027191295623779296, 0.027185152053833008, 0.027328512191772462, 0.02716262435913086, 0.02714521598815918, 0.027246591567993163, 0.027133951187133788, 0.027052032470703126, 0.027188224792480467, 0.026934272766113283, 0.02707967948913574, 0.026879999160766603, 0.027083808898925782, 0.027247583389282227, 0.027197439193725585, 0.027098112106323242, 0.02716364860534668, 0.027140096664428712, 0.02756710433959961, 0.027042816162109375, 0.027053056716918947, 0.026988576889038086, 0.026855424880981447, 0.027275360107421875, 0.02696486473083496, 0.027226112365722657, 0.02775859260559082, 0.027235328674316408, 0.027250688552856447, 0.027099136352539063, 0.02754457664489746, 0.02755788803100586, 0.027371519088745116, 0.027148288726806642, 0.02735103988647461, 0.027158527374267577, 0.02719539260864258, 0.027148288726806642, 0.027287551879882813, 0.02714726448059082, 0.02721491241455078, 0.027155391693115233, 0.07042867279052735, 0.02612326431274414, 0.02608230400085449, 0.02598297691345215, 0.026007551193237305, 0.02595327949523926, 0.02591231918334961, 0.0261345272064209, 0.025977855682373048, 0.02607923126220703, 0.02615705680847168, 0.026201087951660155, 0.026359807968139647, 0.026096704483032227, 0.02599622344970703, 0.02618060874938965, 0.025973760604858398, 0.026053632736206055, 0.025960447311401368, 0.02646937561035156, 0.027265024185180665, 0.027204608917236327, 0.027081727981567383, 0.0273438720703125, 0.02710937690734863, 0.027198463439941405, 0.027130880355834962, 0.02718720054626465, 0.027059200286865235, 0.02735103988647461, 0.027100160598754884, 0.02731520080566406, 0.02716057586669922, 0.02721177673339844, 0.027233280181884766, 0.027216896057128907, 0.027166719436645507, 0.02719436836242676, 0.02713702392578125, 0.02729471969604492, 0.027054079055786134, 0.02714521598815918, 0.02738483238220215, 0.02734694480895996, 0.027257856369018556, 0.027265024185180665, 0.027114496231079102, 0.027465728759765624, 0.027640832901000976, 0.027287551879882813, 0.02716262435913086, 0.027146240234375, 0.02712678337097168, 0.0269486083984375, 0.028556287765502928, 0.027830272674560546, 0.02735513687133789, 0.027633663177490234, 0.027441152572631834, 0.026998783111572267, 0.026987680435180662, 0.027102048873901368, 0.02697318458557129, 0.07149788665771484, 0.02680918312072754, 0.026940479278564453, 0.026708927154541016, 0.02692915153503418, 0.02700492858886719, 0.02691584014892578, 0.027040767669677734, 0.026851327896118164, 0.02693939208984375, 0.026952703475952147, 0.026796159744262697, 0.026875776290893556, 0.02710220718383789, 0.026809471130371094, 0.02700992012023926, 0.02692198371887207, 0.026856447219848634, 0.026825727462768553, 0.026886144638061524, 0.027636735916137696, 0.026849311828613283, 0.02678268814086914, 0.027181055068969725, 0.027267072677612306, 0.027091968536376954, 0.027128896713256835, 0.027300800323486328, 0.02740019226074219, 0.027235328674316408, 0.027283456802368163, 0.027130880355834962, 0.02716876792907715, 0.02736947250366211, 0.027206655502319335, 0.02714419174194336, 0.027233280181884766, 0.027286527633666992, 0.02722003173828125, 0.02752505683898926, 0.027234304428100587, 0.02718617630004883, 0.027225088119506836, 0.027274240493774415, 0.02710425567626953, 0.02726911926269531, 0.027175935745239257, 0.02753331184387207, 0.0285347843170166, 0.027893760681152343, 0.027475967407226562, 0.027454463958740235, 0.02735923194885254, 0.027208703994750977, 0.027222015380859374, 0.027414527893066407, 0.027270143508911132, 0.02721075248718262, 0.027654144287109376, 0.027290624618530275, 0.027261951446533202, 0.02718003273010254, 0.027241472244262696, 0.07036624145507812, 0.026102752685546876, 0.02614169692993164, 0.02609459114074707, 0.026254335403442384, 0.026048511505126954, 0.02590412712097168, 0.026064895629882814, 0.025989120483398437, 0.02612326431274414, 0.025975807189941406, 0.026250240325927734, 0.02592255973815918, 0.02670182418823242, 0.027108352661132814, 0.02612838363647461, 0.025991167068481445, 0.026009599685668947, 0.026049535751342775, 0.026063871383666993, 0.02595327949523926, 0.026032127380371094, 0.026019840240478515, 0.02612224006652832, 0.02591436767578125, 0.026013696670532226, 0.026093568801879883, 0.026171392440795898, 0.0259102725982666, 0.026045440673828125, 0.02591641616821289, 0.02651033592224121, 0.02596249580383301, 0.02595327949523926, 0.02595123291015625, 0.02593382453918457, 0.026430463790893553, 0.026814464569091798, 0.026991615295410155, 0.02754252815246582, 0.026471424102783202, 0.026024959564208985, 0.026017791748046876, 0.02596659278869629, 0.026035200119018553, 0.025965568542480468, 0.025997312545776367, 0.026002431869506838, 0.025948160171508788, 0.026003456115722655, 0.026259456634521484, 0.027115520477294923, 0.027024383544921874, 0.027193344116210938, 0.027107328414916993, 0.027212799072265623, 0.02710323143005371, 0.026047487258911133, 0.025989120483398437, 0.025944063186645508, 0.025844736099243162, 0.025673791885375975, 0.025616384506225585, 0.07033952331542968, 0.026059776306152343, 0.026072063446044923, 0.026012672424316406, 0.026437631607055666, 0.02592870330810547, 0.025882623672485353, 0.0261345272064209, 0.025772031784057618, 0.02592460823059082, 0.025689088821411132, 0.02572083282470703, 0.025817087173461914, 0.026067968368530273, 0.026028032302856444, 0.02617651176452637, 0.026023935317993165, 0.025994239807128908, 0.02594304084777832, 0.026037248611450195, 0.025971712112426756, 0.026034175872802736, 0.0259102725982666, 0.026078208923339844, 0.025981952667236328, 0.026057727813720705, 0.026001440048217774, 0.026085344314575196, 0.026014720916748047, 0.026028032302856444, 0.02596249580383301, 0.026199039459228517, 0.026006528854370117, 0.025985023498535157, 0.025971712112426756, 0.026020864486694335, 0.025977855682373048, 0.025944063186645508, 0.025997312545776367, 0.0267827205657959, 0.027012096405029298, 0.02718720054626465, 0.0275732479095459, 0.02734182357788086, 0.027130880355834962, 0.02710937690734863, 0.02711756706237793, 0.02724870491027832, 0.027140031814575194, 0.02730291175842285, 0.02718617630004883, 0.027436031341552734, 0.027217952728271485, 0.027163616180419924, 0.02714112091064453, 0.02730803108215332, 0.027143167495727538, 0.027133951187133788, 0.027133951187133788, 0.02713599967956543, 0.027602943420410156, 0.027213823318481444, 0.027104288101196288, 0.07280636596679688, 0.02756608009338379, 0.027447359085083008, 0.027171775817871092, 0.02738483238220215, 0.027236352920532225, 0.027200511932373047, 0.027386880874633788, 0.027226112365722657, 0.02710937690734863, 0.027123712539672853, 0.02710121536254883, 0.027154399871826173, 0.027200511932373047, 0.02736025619506836, 0.027639808654785155, 0.02713497543334961, 0.027322368621826174, 0.027227136611938478, 0.027281408309936524, 0.02712063980102539, 0.027131904602050783, 0.027157503128051756, 0.027413503646850586, 0.027240447998046875, 0.027374591827392578, 0.027262975692749023, 0.027215871810913086, 0.02711244773864746, 0.0270960636138916, 0.027100160598754884, 0.02711244773864746, 0.0269752311706543, 0.02698854446411133, 0.027049983978271484, 0.026860544204711914, 0.026838016510009766, 0.02695782470703125, 0.026976255416870116, 0.027067424774169922, 0.02699875259399414, 0.026978303909301758, 0.026952703475952147, 0.026856447219848634, 0.027084800720214845, 0.02690764808654785, 0.02693529510498047, 0.02695577621459961, 0.026825727462768553, 0.02681964874267578, 0.026899391174316407, 0.02716160011291504, 0.027412511825561522, 0.027270111083984375, 0.02696601676940918, 0.026945535659790038, 0.027148288726806642, 0.02692198371887207, 0.02696499252319336, 0.026942464828491212, 0.027204608917236327, 0.027159551620483398, 0.0271779842376709, 0.07124393463134765, 0.026900320053100585, 0.027198463439941405, 0.02709401512145996, 0.02699673652648926, 0.02715340805053711, 0.02719436836242676, 0.02736128044128418, 0.027084800720214845, 0.0272936954498291, 0.02716057586669922, 0.02715648078918457, 0.02672640037536621, 0.027028480529785157, 0.027262975692749023, 0.027364351272583007, 0.02709401512145996, 0.027172927856445313, 0.02710419273376465, 0.027295743942260742, 0.0271329288482666, 0.027214847564697265, 0.027223039627075195, 0.027454463958740235, 0.027196416854858397, 0.027317279815673827, 0.0272260799407959, 0.02733568000793457, 0.0271329288482666, 0.027632640838623046, 0.027228160858154295, 0.02735923194885254, 0.0273305606842041, 0.02733363151550293, 0.02727628707885742, 0.027205631256103514, 0.02734694480895996, 0.027286527633666992, 0.027815935134887695, 0.02772684860229492, 0.027167743682861328, 0.027199487686157226, 0.02710220718383789, 0.027193344116210938, 0.0271646728515625, 0.027201536178588868, 0.0269803524017334, 0.02718003273010254, 0.027183103561401366, 0.027229183197021483, 0.02709503936767578, 0.02712063980102539, 0.027130880355834962, 0.027052032470703126, 0.026846208572387696, 0.027243520736694334, 0.02709503936767578, 0.02736844825744629, 0.02721177673339844, 0.027215871810913086, 0.02756096076965332, 0.0273623046875, 0.02711961555480957]",tokens/s,36.27487553913198,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,meta-llama/Meta-Llama-3-8B,meta-llama/Meta-Llama-3-8B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.1,,,,MB,1284.698112,17201.364992,0.0,16554.917888,16436.726784,s,10,0.8754648361206055,0.08754648361206055,0.0003590358347961992,0.0874463233947754,0.08776272735595703,0.08813846664428711,0.08843905807495117,"[0.08851420593261719, 0.08721401977539063, 0.08756735992431641, 0.08767657470703125, 0.0873911361694336, 0.08731068420410157, 0.08744409942626953, 0.08721897888183594, 0.08767922973632812, 0.08744854736328125]",tokens/s,2924.160850758979,kWh,1.03865675991274e-06,5.690339180594156e-07,7.118195265701671e-06,8.725885943673826e-06,tokens/kWh,29337995.207878835,MB,1284.698112,17201.364992,0.0,16554.917888,16436.729344,s,10,24.932140380859376,2.4932140380859376,0.0028762497653695113,2.4920468749999998,2.4961453369140627,2.4978859008789063,2.4992783520507813,"[2.49962646484375, 2.49169921875, 2.49004296875, 2.49003662109375, 2.49160107421875, 2.495758544921875, 2.491458251953125, 2.49239453125, 2.495639892578125, 2.4938828125]",tokens/s,25.268588672140503,kWh,2.936688299445149e-05,1.6094310801831613e-05,0.0001328600833834982,0.0001783212771797813,tokens/kWh,353294.9123983908,,s,629,25.6786350517273,0.04082453903295277,0.010449669636073064,0.0395489273071289,0.03967221603393555,0.039785677337646486,0.12735913970947266,"[0.04021145629882812, 0.04008348846435547, 0.0400670394897461, 0.039992321014404295, 0.04005683135986328, 0.03996876907348633, 0.04016230392456055, 0.04013568115234375, 0.04012239837646484, 0.04000252914428711, 0.03984998321533203, 0.039861248016357424, 0.03968819046020508, 0.03948134231567383, 0.03955916976928711, 0.03948134231567383, 0.03950694274902344, 0.039449600219726565, 0.039553024291992187, 0.039711742401123046, 0.03980492782592773, 0.03960115051269531, 0.0394967041015625, 0.03946188735961914, 0.03951308822631836, 0.039482368469238284, 0.03953049468994141, 0.03947212982177734, 0.03951615905761719, 0.03948646545410156, 0.03948031997680664, 0.03948748779296875, 0.039537662506103514, 0.03950796890258789, 0.03956633758544922, 0.03951513671875, 0.039556095123291016, 0.03954380798339844, 0.039599105834960936, 0.03957964706420898, 0.039602176666259765, 0.03958169555664062, 0.03955199813842773, 0.03953561782836914, 0.03955712127685547, 0.039570430755615234, 0.03958681488037109, 0.03952435302734375, 0.03954380798339844, 0.03955712127685547, 0.039654399871826174, 0.03975372695922851, 0.03958988952636719, 0.039608318328857424, 0.03959296035766602, 0.03962777709960937, 0.039626750946044925, 0.03959500885009765, 0.039602176666259765, 0.039634944915771485, 0.03958988952636719, 0.03961753463745117, 0.12736000061035158, 0.039468032836914066, 0.039417888641357424, 0.03941987228393555, 0.03940249633789063, 0.039444480895996094, 0.03940556716918946, 0.039452671051025394, 0.039387134552001955, 0.03950185775756836, 0.039405536651611325, 0.039465984344482424, 0.03940454483032227, 0.03948646545410156, 0.03942095947265625, 0.039443424224853516, 0.03945676803588867, 0.039488510131835936, 0.039444480895996094, 0.03948953628540039, 0.039462913513183595, 0.03952025604248047, 0.03948441696166992, 0.03951513671875, 0.03949363327026367, 0.039512065887451174, 0.03949772644042969, 0.03955814361572266, 0.039498752593994144, 0.039534591674804685, 0.03958784103393555, 0.03990425491333008, 0.03975065612792969, 0.03954380798339844, 0.0395335693359375, 0.03951103973388672, 0.03952230453491211, 0.039547904968261716, 0.03955916976928711, 0.03953664016723633, 0.03952435302734375, 0.039602176666259765, 0.039559200286865236, 0.03960623931884766, 0.039556095123291016, 0.039567359924316405, 0.039534591674804685, 0.03963289642333984, 0.03968000030517578, 0.03955814361572266, 0.03962572860717774, 0.039583744049072264, 0.039593982696533206, 0.03955712127685547, 0.039599105834960936, 0.03967078399658203, 0.039863296508789066, 0.03966156768798828, 0.03962265777587891, 0.03958681488037109, 0.039624702453613284, 0.03967385482788086, 0.03970969772338867, 0.12759552001953126, 0.03950899124145508, 0.0393994255065918, 0.03944345474243164, 0.03938304138183594, 0.039465984344482424, 0.03948134231567383, 0.03947315216064453, 0.03943423843383789, 0.03950899124145508, 0.03942911911010742, 0.03944140625, 0.03942195129394531, 0.03946396636962891, 0.039430110931396485, 0.03948857498168945, 0.03943008041381836, 0.03951615905761719, 0.0393891830444336, 0.03953868865966797, 0.03948339080810547, 0.03952435302734375, 0.03948748779296875, 0.03951923370361328, 0.03941584014892578, 0.039530464172363285, 0.03951411056518555, 0.03960319900512695, 0.039462913513183595, 0.03950796890258789, 0.03951411056518555, 0.03954073715209961, 0.03948646545410156, 0.03951411056518555, 0.03948953628540039, 0.0395335693359375, 0.039478271484375, 0.039532543182373044, 0.03950592041015625, 0.03953868865966797, 0.0395417594909668, 0.03953868865966797, 0.039567359924316405, 0.0395601921081543, 0.03954585647583008, 0.03954585647583008, 0.0395417594909668, 0.039621631622314454, 0.03954995346069336, 0.03958784103393555, 0.03957350540161133, 0.03954585647583008, 0.03958272171020508, 0.0396124153137207, 0.039615550994873044, 0.03960416030883789, 0.03959296035766602, 0.039585792541503906, 0.03959091186523438, 0.039602176666259765, 0.03960627365112305, 0.039564289093017575, 0.03955513763427734, 0.1276824951171875, 0.039430145263671876, 0.03946393585205078, 0.039572479248046875, 0.03941273498535156, 0.03941068649291992, 0.03942399978637695, 0.03945574569702148, 0.0393891830444336, 0.03946700668334961, 0.03942911911010742, 0.03947724914550781, 0.03951513671875, 0.039446529388427735, 0.039416831970214845, 0.03949158477783203, 0.03942502212524414, 0.0394967041015625, 0.03945676803588867, 0.03947212982177734, 0.03946086502075195, 0.039521278381347655, 0.03943219375610352, 0.039465984344482424, 0.039468032836914066, 0.039553024291992187, 0.039501823425292966, 0.03956633758544922, 0.03946905517578125, 0.039504894256591795, 0.03949772644042969, 0.039523326873779296, 0.03949264144897461, 0.03950486373901367, 0.03950694274902344, 0.039512065887451174, 0.03949977493286133, 0.039539710998535156, 0.039569408416748046, 0.03953667068481445, 0.03948028945922852, 0.039547904968261716, 0.03954687881469727, 0.03958784103393555, 0.03959296035766602, 0.03958169555664062, 0.03957145690917969, 0.039543838500976564, 0.039620574951171876, 0.03954073715209961, 0.039556095123291016, 0.03957964706420898, 0.03956224060058594, 0.039597057342529295, 0.03957555389404297, 0.03953868865966797, 0.039564289093017575, 0.03960422515869141, 0.039599105834960936, 0.03966054534912109, 0.03962777709960937, 0.03958681488037109, 0.03962777709960937, 0.12738457489013671, 0.039436286926269534, 0.03940454483032227, 0.03948031997680664, 0.03933184051513672, 0.039478271484375, 0.03937795257568359, 0.03948028945922852, 0.0395786247253418, 0.03945471954345703, 0.03947315216064453, 0.03951411056518555, 0.03953664016723633, 0.03969843292236328, 0.03950592041015625, 0.03950284957885742, 0.03944755172729492, 0.03957968139648437, 0.03945059204101563, 0.03952742385864258, 0.03945574569702148, 0.03954483032226563, 0.03947417449951172, 0.03957657623291016, 0.03947110366821289, 0.03954585647583008, 0.039508033752441406, 0.03957139205932617, 0.03948339080810547, 0.03955814361572266, 0.03950387191772461, 0.03952435302734375, 0.03947520065307617, 0.039531520843505856, 0.03949260711669922, 0.03951411056518555, 0.039444480895996094, 0.039495681762695314, 0.03952947235107422, 0.03952537536621094, 0.03948748779296875, 0.039556095123291016, 0.03956224060058594, 0.039564289093017575, 0.03952844619750977, 0.03957145690917969, 0.039550975799560545, 0.03956531143188476, 0.03962777709960937, 0.0395489273071289, 0.0395786247253418, 0.04031795120239258, 0.03968000030517578, 0.039569408416748046, 0.0396124153137207, 0.03965235137939453, 0.03955814361572266, 0.039657470703125, 0.03959807968139648, 0.03959603118896484, 0.03962879943847656, 0.039572479248046875, 0.039570430755615234, 0.12735692596435547, 0.03945580673217773, 0.039385025024414065, 0.03945779037475586, 0.03940761566162109, 0.03946188735961914, 0.03939328002929687, 0.03948953628540039, 0.03957145690917969, 0.0395417594909668, 0.039465984344482424, 0.03948953628540039, 0.03949465560913086, 0.039531520843505856, 0.03950899124145508, 0.03949260711669922, 0.03948748779296875, 0.03951308822631836, 0.03948953628540039, 0.03950899124145508, 0.03951411056518555, 0.03956531143188476, 0.04016844940185547, 0.04020019149780273, 0.039754753112792966, 0.03962060928344727, 0.039613441467285154, 0.03954687881469727, 0.03951103973388672, 0.03956224060058594, 0.0395417594909668, 0.039569408416748046, 0.039554046630859374, 0.03956326293945313, 0.03960422515869141, 0.03954995346069336, 0.039585792541503906, 0.03954278564453125, 0.039646209716796874, 0.03961446380615234, 0.039725055694580076, 0.03959807968139648, 0.03960934448242188, 0.03969740676879883, 0.039646209716796874, 0.039605247497558595, 0.039569408416748046, 0.040048641204833986, 0.04009574508666992, 0.03974041748046875, 0.03965542221069336, 0.03971583938598633, 0.03968511962890625, 0.03954995346069336, 0.039723007202148435, 0.039667713165283204, 0.03961958312988281, 0.03960934448242188, 0.039708671569824217, 0.039626750946044925, 0.039618560791015625, 0.039583744049072264, 0.03964825439453125, 0.12745728302001952, 0.039523326873779296, 0.039416831970214845, 0.03957350540161133, 0.039534591674804685, 0.03950592041015625, 0.03949055862426758, 0.03955814361572266, 0.0395335693359375, 0.03952742385864258, 0.03948134231567383, 0.039488510131835936, 0.039583744049072264, 0.03948751831054687, 0.039428062438964843, 0.03947417449951172, 0.039482368469238284, 0.03950796890258789, 0.03947622299194336, 0.03964518356323242, 0.039508033752441406, 0.03952326583862305, 0.03947110366821289, 0.03956224060058594, 0.039512065887451174, 0.039553024291992187, 0.03962060928344727, 0.039585792541503906, 0.03949055862426758, 0.03952844619750977, 0.039504894256591795, 0.03952230453491211, 0.039488544464111326, 0.03956118392944336, 0.039501823425292966, 0.039553024291992187, 0.03951615905761719, 0.0395417594909668, 0.03952640151977539, 0.0395417594909668, 0.03951615905761719, 0.039561279296875, 0.039538623809814454, 0.039623680114746096, 0.039640064239501956, 0.039593982696533206, 0.03956326293945313, 0.03956739044189453, 0.03955094528198242, 0.03952844619750977, 0.03955507278442383, 0.039583744049072264, 0.03958784103393555, 0.03954483032226563, 0.03954483032226563, 0.0395601921081543, 0.03957350540161133, 0.039567359924316405, 0.03959913635253906, 0.039585758209228515, 0.039610366821289066, 0.039580673217773435, 0.03959807968139648, 0.1272258529663086, 0.03946905517578125, 0.039378944396972655, 0.0394598388671875, 0.03952640151977539, 0.03945676803588867, 0.03946700668334961, 0.039462913513183595, 0.03947520065307617, 0.039512065887451174, 0.03945881652832031, 0.03951615905761719, 0.039430145263671876, 0.03944755172729492, 0.03944140625, 0.03948441696166992, 0.03951513671875, 0.039532543182373044, 0.039501823425292966, 0.039553024291992187, 0.03948134231567383, 0.03948134231567383, 0.0394967041015625, 0.039521278381347655, 0.03954073715209961, 0.03951615905761719, 0.03949260711669922, 0.03951308822631836, 0.03951308822631836, 0.03957964706420898, 0.03973222351074219, 0.039523326873779296, 0.0395417594909668, 0.039588863372802735, 0.0395335693359375, 0.03956838226318359, 0.039534591674804685, 0.03957452774047852, 0.03953664016723633, 0.039548961639404294, 0.03952022552490234, 0.03958272171020508, 0.03960934448242188, 0.03959296035766602, 0.03954995346069336, 0.03962681579589844, 0.03952633666992188, 0.03955199813842773, 0.039613441467285154, 0.03967180633544922, 0.03963289642333984, 0.03966566467285156, 0.0396308479309082, 0.03962777709960937, 0.03969126510620117, 0.03966873550415039, 0.03966668701171875, 0.03965235137939453, 0.03969740676879883, 0.03966259384155273, 0.03972915267944336, 0.03964108657836914, 0.03969126510620117, 0.1283440704345703, 0.039600128173828124, 0.03941580963134766, 0.03945676803588867, 0.039482368469238284, 0.040406017303466796, 0.03961548614501953, 0.03950387191772461, 0.03948748779296875, 0.03954073715209961, 0.0395489273071289, 0.039479297637939455, 0.03948441696166992, 0.03973324966430664, 0.03965030288696289, 0.039651329040527344, 0.03951513671875, 0.03957657623291016, 0.03952947235107422, 0.03949055862426758, 0.03946905517578125, 0.03951308822631836, 0.03955712127685547, 0.03948646545410156, 0.039488510131835936, 0.03953664016723633, 0.03952230453491211, 0.03955507278442383, 0.03963596725463867, 0.03975171279907227, 0.03956118392944336, 0.03961958312988281, 0.03955814361572266, 0.039602176666259765, 0.039646209716796874, 0.03957452774047852, 0.03959500885009765, 0.039618560791015625, 0.03961958312988281, 0.03972403335571289, 0.03966156768798828, 0.039657470703125, 0.03961139297485351, 0.0395786247253418, 0.03952742385864258, 0.039597057342529295, 0.039564289093017575, 0.03964416122436523, 0.039534591674804685, 0.03956838226318359, 0.03957657623291016, 0.03962265777587891, 0.03965542221069336, 0.039618560791015625, 0.039626750946044925, 0.03962572860717774, 0.03963391876220703, 0.03967385482788086, 0.03969331359863281, 0.03962265777587891, 0.03970457458496094, 0.039597057342529295, 0.039623680114746096, 0.12753612518310548, 0.03954278564453125, 0.03945676803588867, 0.03941888046264649, 0.03940556716918946, 0.03955507278442383, 0.039498752593994144, 0.03951308822631836, 0.039470111846923825, 0.039525344848632814, 0.03942297744750976, 0.03947622299194336, 0.03941888046264649, 0.039680065155029295, 0.03965024185180664, 0.04024524688720703, 0.039616512298583983, 0.03961446380615234, 0.03946700668334961, 0.03956531143188476, 0.03955507278442383, 0.03956531143188476, 0.039564289093017575, 0.03952742385864258, 0.039547904968261716, 0.03959091186523438, 0.03954483032226563, 0.0395335693359375, 0.03950796890258789, 0.03951923370361328, 0.03950387191772461, 0.0395335693359375, 0.0395489273071289, 0.03955199813842773, 0.03953664016723633, 0.03954585647583008, 0.03949465560913086, 0.0395417594909668, 0.03950796890258789, 0.03975680160522461, 0.040223743438720705, 0.039646209716796874, 0.03958681488037109, 0.039580673217773435, 0.03954483032226563, 0.039559200286865236, 0.03953968048095703, 0.039570430755615234, 0.03956121444702149, 0.03955916976928711, 0.03960627365112305, 0.03968819046020508, 0.03959296035766602, 0.039588863372802735, 0.03958784103393555, 0.03957657623291016, 0.0395786247253418, 0.03956633758544922, 0.039623680114746096, 0.03955916976928711, 0.039607295989990236, 0.03965030288696289, 0.03966668701171875]",tokens/s,24.495071437128033,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen2-beta-14B,Qwen/Qwen2-beta-14B,cuda,0,42,,,True,,,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",qwen2,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-32B,Qwen/Qwen1.5-32B,cuda,0,42,,,True,,,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",qwen2,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,0,0,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/0/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6693258f-10411ad85191c5873f9409b4;ea50e8a2-8b18-4392-a18b-44eb116ba69a) Repository Not Found for url: https://huggingface.co/0/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: 0 is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/gpt-neo-125m,EleutherAI/gpt-neo-125m,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1218.727936,1084.751872,0.0,438.304768,353.979904,s,17,0.17676496124267577,0.010397938896627986,0.0002656030797380195,0.010435104370117188,0.010544889450073243,0.010700460815429687,0.011088187866210937,"[0.01118511962890625, 0.010507040023803711, 0.010435104370117188, 0.01052195167541504, 0.010475520133972169, 0.010147135734558105, 0.010040032386779784, 0.01006704044342041, 0.010018560409545898, 0.010437088012695312, 0.010174528121948242, 0.010447936058044434, 0.010380703926086426, 0.010429439544677734, 0.01050857639312744, 0.01040988826751709, 0.010579296112060547]",tokens/s,24620.263933558974,kWh,1.2269846908468764e-07,6.723301995007885e-08,3.0485342260643176e-07,4.947849116411982e-07,tokens/kWh,517396537.31729555,MB,1218.727936,1084.751872,0.0,438.304768,363.872768,s,17,10.014291809082032,0.5890759887695313,0.012479596868940167,0.590957275390625,0.5950199584960938,0.6013368530273437,0.6193056030273437,"[0.6237977905273437, 0.590957275390625, 0.5945521850585938, 0.5914871215820312, 0.5913626098632813, 0.5676865234375, 0.5692130737304687, 0.5682417602539063, 0.5834576416015625, 0.5922348022460937, 0.590658203125, 0.5901029052734375, 0.5926476440429688, 0.5957216186523437, 0.5913841552734375, 0.5908838500976562, 0.5899026489257813]",tokens/s,106.94715317050203,kWh,6.689330642079901e-06,3.665462908454834e-06,1.1662038866948911e-05,2.2016832417483647e-05,tokens/kWh,2861447.0422171843,,s,1070,10.169835510253916,0.009504519168461597,0.0013356373713787864,0.009369600296020507,0.009513165187835694,0.009919590377807617,0.019873134956359836,"[0.010174464225769043, 0.010402815818786621, 0.009853952407836914, 0.00963379192352295, 0.009620479583740234, 0.009950207710266114, 0.00992255973815918, 0.009778176307678223, 0.009975872039794921, 0.009969599723815918, 0.010093567848205566, 0.00994611167907715, 0.009973759651184083, 0.009989119529724122, 0.010070015907287597, 0.009999360084533691, 0.009920512199401856, 0.009900032043457031, 0.010067968368530274, 0.009803775787353516, 0.009885696411132813, 0.009881600379943848, 0.009934847831726074, 0.009893888473510743, 0.010110976219177246, 0.010145792007446289, 0.010158080101013184, 0.010060799598693848, 0.009962495803833007, 0.009979904174804688, 0.009997311592102051, 0.009842687606811524, 0.009902112007141113, 0.009897952079772948, 0.010107904434204102, 0.00984166431427002, 0.009918463706970216, 0.009852928161621094, 0.009905152320861817, 0.009853952407836914, 0.010215423583984374, 0.010109951972961426, 0.010110976219177246, 0.009930751800537109, 0.009936896324157715, 0.009945088386535645, 0.0098088960647583, 0.009736191749572755, 0.009891839981079101, 0.009659392356872559, 0.009438207626342773, 0.009562111854553223, 0.00979251194000244, 0.009784319877624511, 0.009837568283081055, 0.01015500831604004, 0.009783295631408692, 0.009540608406066894, 0.009492480278015136, 0.009382911682128906, 0.009452544212341308, 0.009443327903747559, 0.020543487548828124, 0.00937987232208252, 0.009411552429199218, 0.009360383987426758, 0.009343999862670899, 0.00940544033050537, 0.009361408233642577, 0.009371647834777832, 0.009391103744506836, 0.009365504264831542, 0.009387007713317871, 0.009388031959533692, 0.009376768112182618, 0.009423871994018555, 0.009389056205749511, 0.009392127990722657, 0.009376768112182618, 0.009361408233642577, 0.009377792358398437, 0.00932147216796875, 0.009444352149963378, 0.009360383987426758, 0.009362431526184082, 0.009359359741210938, 0.009423871994018555, 0.009404416084289552, 0.009385984420776367, 0.009388031959533692, 0.009365504264831542, 0.009352191925048828, 0.009331711769104004, 0.00942796802520752, 0.009364480018615723, 0.009461759567260742, 0.009395199775695801, 0.00934505558013916, 0.00950166416168213, 0.009369664192199707, 0.009361344337463379, 0.009383935928344727, 0.009350175857543946, 0.009351136207580567, 0.009350144386291503, 0.009357312202453612, 0.009358336448669433, 0.009366527557373047, 0.009500672340393066, 0.009432064056396485, 0.009361408233642577, 0.009390080451965332, 0.009275391578674316, 0.009285632133483887, 0.009282624244689942, 0.009214912414550781, 0.009219072341918945, 0.009292799949645996, 0.009383935928344727, 0.009351167678833008, 0.009391103744506836, 0.009347071647644043, 0.009374719619750976, 0.009377792358398437, 0.009379839897155762, 0.02030182456970215, 0.009381888389587402, 0.009470975875854493, 0.009486335754394531, 0.009342975616455078, 0.009358336448669433, 0.009399295806884766, 0.009390080451965332, 0.009385984420776367, 0.009373696327209472, 0.00939724826812744, 0.00940339183807373, 0.009396224021911622, 0.009382911682128906, 0.009383935928344727, 0.00942796802520752, 0.009441280364990234, 0.009370623588562011, 0.009387007713317871, 0.009658368110656738, 0.010079232215881348, 0.009925632476806641, 0.009467904090881347, 0.009592831611633301, 0.009434111595153808, 0.009449472427368164, 0.009442303657531738, 0.009400320053100587, 0.009433088302612304, 0.009387007713317871, 0.009348095893859864, 0.009371647834777832, 0.009362431526184082, 0.009362431526184082, 0.009396224021911622, 0.009356287956237793, 0.009329664230346679, 0.009460736274719238, 0.009327615737915039, 0.009370623588562011, 0.009409536361694336, 0.009355263710021973, 0.009473024368286133, 0.009384960174560546, 0.009376768112182618, 0.009388031959533692, 0.009378815650939941, 0.009400320053100587, 0.00942080020904541, 0.009380928039550781, 0.009424832344055176, 0.009415679931640625, 0.009399295806884766, 0.009342975616455078, 0.009425919532775879, 0.009366527557373047, 0.009342975616455078, 0.009389120101928711, 0.009448384284973145, 0.009347071647644043, 0.009605119705200196, 0.00961740779876709, 0.009477120399475097, 0.020470783233642577, 0.00940544033050537, 0.009358336448669433, 0.009391103744506836, 0.009463808059692384, 0.009469951629638672, 0.009417728424072265, 0.00935321617126465, 0.00941875171661377, 0.009381888389587402, 0.00940236759185791, 0.009399295806884766, 0.009455615997314454, 0.009373696327209472, 0.009488384246826171, 0.009396224021911622, 0.009363455772399902, 0.00939724826812744, 0.009359423637390136, 0.009375679969787597, 0.009351167678833008, 0.009393152236938476, 0.00935321617126465, 0.009415679931640625, 0.009399295806884766, 0.009277440071105958, 0.00922214412689209, 0.009234432220458985, 0.009298944473266601, 0.009365504264831542, 0.009358336448669433, 0.009349120140075684, 0.009396224021911622, 0.009366527557373047, 0.009362431526184082, 0.009417728424072265, 0.009373696327209472, 0.009354240417480468, 0.009366527557373047, 0.009379839897155762, 0.009348095893859864, 0.009368576049804688, 0.009340991973876953, 0.009428928375244141, 0.009441280364990234, 0.00940544033050537, 0.009363455772399902, 0.009422847747802734, 0.009360383987426758, 0.009406463623046875, 0.009347071647644043, 0.009367615699768066, 0.009376704216003418, 0.009453568458557129, 0.00933683204650879, 0.009381888389587402, 0.009358336448669433, 0.009374719619750976, 0.009417728424072265, 0.009393152236938476, 0.009342975616455078, 0.00939724826812744, 0.009441280364990234, 0.020382720947265624, 0.009414655685424805, 0.009406463623046875, 0.009447423934936524, 0.010085375785827636, 0.009591839790344238, 0.009545696258544923, 0.009430015563964844, 0.00943824005126953, 0.009404383659362792, 0.009449472427368164, 0.009448448181152343, 0.009470975875854493, 0.009430015563964844, 0.009351167678833008, 0.009408512115478516, 0.009389056205749511, 0.009417728424072265, 0.009395232200622558, 0.009298912048339844, 0.009404416084289552, 0.009364480018615723, 0.009342975616455078, 0.009357312202453612, 0.009363455772399902, 0.009365504264831542, 0.009433088302612304, 0.009368576049804688, 0.009366527557373047, 0.009355263710021973, 0.009354271888732911, 0.009400287628173827, 0.009363455772399902, 0.009342975616455078, 0.009368576049804688, 0.009334783554077148, 0.009390080451965332, 0.009308159828186035, 0.00935321617126465, 0.009347071647644043, 0.009363455772399902, 0.009396224021911622, 0.009349120140075684, 0.009343999862670899, 0.009220095634460449, 0.009189375877380371, 0.009200639724731445, 0.009273344039916993, 0.009210880279541016, 0.009242624282836913, 0.009247743606567382, 0.00939724826812744, 0.00939417552947998, 0.009323519706726074, 0.009343999862670899, 0.00933683204650879, 0.009400320053100587, 0.009372672080993653, 0.009372672080993653, 0.009339903831481934, 0.009351167678833008, 0.009315327644348144, 0.00939417552947998, 0.0196495361328125, 0.009075712203979493, 0.008954879760742187, 0.008932352066040039, 0.00890777587890625, 0.008953856468200684, 0.008931327819824218, 0.008995840072631836, 0.008992799758911133, 0.00898249626159668, 0.009034751892089844, 0.009033727645874023, 0.00900710391998291, 0.009048064231872559, 0.008984576225280762, 0.009034751892089844, 0.009009152412414552, 0.009036800384521485, 0.009053183555603026, 0.00899071979522705, 0.009096192359924317, 0.009030655860900879, 0.009014271736145019, 0.009026559829711914, 0.009038847923278808, 0.009041919708251953, 0.008922112464904786, 0.008910847663879394, 0.008920063972473144, 0.008968192100524902, 0.009003007888793945, 0.009029631614685058, 0.009049087524414063, 0.009011199951171875, 0.009011199951171875, 0.00901734447479248, 0.00900710391998291, 0.009043968200683594, 0.009163776397705077, 0.00902348804473877, 0.00902348804473877, 0.009053183555603026, 0.008992768287658692, 0.008995840072631836, 0.009019392013549805, 0.009083904266357423, 0.009009152412414552, 0.008944640159606934, 0.008954879760742187, 0.008963071823120117, 0.008885248184204102, 0.00900710391998291, 0.009000991821289063, 0.009083871841430663, 0.008989727973937989, 0.00905622386932373, 0.009034751892089844, 0.009042943954467773, 0.009013248443603515, 0.008936448097229004, 0.008980480194091797, 0.008972288131713867, 0.008904704093933105, 0.019518463134765626, 0.009029631614685058, 0.009055232048034668, 0.008947711944580078, 0.008931327819824218, 0.00902348804473877, 0.008993791580200196, 0.009047039985656738, 0.009050111770629882, 0.009029631614685058, 0.009020416259765626, 0.009057279586791991, 0.009029631614685058, 0.009018367767333984, 0.008992768287658692, 0.009074687957763672, 0.00914739227294922, 0.009063424110412598, 0.009030655860900879, 0.008987648010253906, 0.009034751892089844, 0.008995840072631836, 0.009009152412414552, 0.009051136016845703, 0.009018367767333984, 0.00900812816619873, 0.009111552238464356, 0.009047039985656738, 0.009030655860900879, 0.00901632022857666, 0.008986623764038085, 0.009027584075927735, 0.009003007888793945, 0.008993791580200196, 0.009000960350036622, 0.009058303833007812, 0.00900812816619873, 0.008965120315551758, 0.009005056381225587, 0.009073663711547851, 0.009020416259765626, 0.009022463798522949, 0.009049087524414063, 0.008999936103820801, 0.00900607967376709, 0.008996864318847657, 0.009029631614685058, 0.009035776138305664, 0.009246784210205078, 0.009071552276611328, 0.00900812816619873, 0.009029631614685058, 0.009013248443603515, 0.009056256294250489, 0.009001983642578124, 0.009030655860900879, 0.009019392013549805, 0.00902143955230713, 0.009001983642578124, 0.009150464057922364, 0.00902143955230713, 0.008981504440307618, 0.008975359916687011, 0.01962700843811035, 0.009022527694702149, 0.008994751930236817, 0.009084927558898925, 0.00900607967376709, 0.009035776138305664, 0.009011263847351075, 0.009078720092773437, 0.009010175704956054, 0.008991744041442871, 0.009057279586791991, 0.009052160263061524, 0.00897433567047119, 0.008979455947875976, 0.009011199951171875, 0.009044992446899413, 0.009088000297546387, 0.008970239639282226, 0.008995840072631836, 0.009000960350036622, 0.00900812816619873, 0.009000960350036622, 0.00902451229095459, 0.00900607967376709, 0.009014271736145019, 0.008996864318847657, 0.008999936103820801, 0.008986687660217286, 0.009003968238830566, 0.008999936103820801, 0.009033727645874023, 0.008987648010253906, 0.009029631614685058, 0.008951807975769043, 0.008927231788635253, 0.008928288459777832, 0.008912863731384278, 0.008929280281066895, 0.009071616172790528, 0.009040896415710448, 0.008981535911560059, 0.008976351737976075, 0.009027584075927735, 0.009036800384521485, 0.008972288131713867, 0.00905628776550293, 0.00905724811553955, 0.00902348804473877, 0.009067520141601563, 0.00900614356994629, 0.009002943992614747, 0.009011199951171875, 0.009004032135009766, 0.00901529598236084, 0.008989695549011231, 0.009018367767333984, 0.00898252773284912, 0.009028608322143555, 0.00899891185760498, 0.009028608322143555, 0.009083904266357423, 0.009041919708251953, 0.009012224197387696, 0.019522560119628905, 0.009038847923278808, 0.00901734447479248, 0.009000960350036622, 0.008977439880371094, 0.009044960021972656, 0.008987648010253906, 0.00900812816619873, 0.009005056381225587, 0.009079808235168458, 0.00902348804473877, 0.009012224197387696, 0.00898252773284912, 0.009061375617980956, 0.009029631614685058, 0.009035776138305664, 0.00901632022857666, 0.00913920021057129, 0.009040896415710448, 0.00900710391998291, 0.009029631614685058, 0.009053183555603026, 0.00903270435333252, 0.009034751892089844, 0.009047039985656738, 0.009061375617980956, 0.009014271736145019, 0.008980480194091797, 0.009043968200683594, 0.009037823677062988, 0.009003007888793945, 0.009001983642578124, 0.009140288352966309, 0.009041855812072754, 0.00900812816619873, 0.009072640419006347, 0.009036800384521485, 0.009027584075927735, 0.00901632022857666, 0.009025535583496093, 0.010729472160339355, 0.013674495697021484, 0.00954470443725586, 0.009391103744506836, 0.009396224021911622, 0.009346048355102539, 0.009339903831481934, 0.009389056205749511, 0.00937990379333496, 0.009364416122436523, 0.00951296043395996, 0.009368576049804688, 0.009348095893859864, 0.009543680191040039, 0.009389056205749511, 0.009395199775695801, 0.00943513584136963, 0.009388031959533692, 0.0094136323928833, 0.009378815650939941, 0.009331775665283202, 0.00932652759552002, 0.009390080451965332, 0.020287488937377928, 0.009373696327209472, 0.00949350357055664, 0.009415679931640625, 0.009366527557373047, 0.009341952323913574, 0.009364480018615723, 0.009583616256713867, 0.009462783813476563, 0.009382911682128906, 0.009347071647644043, 0.009425919532775879, 0.009341952323913574, 0.009373696327209472, 0.009399295806884766, 0.009492480278015136, 0.009476096153259277, 0.009373696327209472, 0.009351167678833008, 0.00942796802520752, 0.009362463951110839, 0.00937059211730957, 0.009365504264831542, 0.00941875171661377, 0.009378815650939941, 0.009377792358398437, 0.009415679931640625, 0.00941875171661377, 0.00941055965423584, 0.009432064056396485, 0.009391103744506836, 0.009368576049804688, 0.009398271560668945, 0.009341952323913574, 0.009366527557373047, 0.009351231575012208, 0.009378751754760743, 0.009483263969421387, 0.009388031959533692, 0.009472000122070312, 0.009389056205749511, 0.009359359741210938, 0.009461759567260742, 0.009408576011657715, 0.00935315227508545, 0.009359359741210938, 0.009369600296020507, 0.009400320053100587, 0.00942080020904541, 0.009075712203979493, 0.009240575790405273, 0.009600000381469726, 0.00939724826812744, 0.009457663536071777, 0.009376768112182618, 0.009357312202453612, 0.009357312202453612, 0.009400320053100587, 0.009348095893859864, 0.009451519966125489, 0.009379839897155762, 0.009407487869262696, 0.00933891201019287, 0.01971504020690918, 0.009044992446899413, 0.009003007888793945, 0.009043968200683594, 0.009048064231872559, 0.009407551765441894, 0.009396160125732422, 0.009371647834777832, 0.009417728424072265, 0.009362463951110839, 0.00939516830444336, 0.00941260814666748, 0.00933683204650879, 0.009453568458557129, 0.009374719619750976, 0.009400320053100587, 0.009451519966125489, 0.009384960174560546, 0.00940544033050537, 0.009387007713317871, 0.00936963176727295, 0.009425888061523438, 0.009384960174560546, 0.009482239723205567, 0.009340928077697755, 0.009425919532775879, 0.009387007713317871, 0.009407487869262696, 0.009546751976013184, 0.009369600296020507, 0.00942182445526123, 0.009437184333801269, 0.009404416084289552, 0.009377792358398437, 0.009366527557373047, 0.009340928077697755, 0.009333760261535644, 0.00932249641418457, 0.009343999862670899, 0.009375743865966797, 0.009364480018615723, 0.009380864143371583, 0.009341952323913574, 0.009454591751098633, 0.009359359741210938, 0.009414655685424805, 0.009381888389587402, 0.009390080451965332, 0.009395199775695801, 0.009430015563964844, 0.009605119705200196, 0.009443327903747559, 0.009348095893859864, 0.009430015563964844, 0.009389056205749511, 0.009362431526184082, 0.009352191925048828, 0.009352255821228027, 0.009367520332336426, 0.00947606372833252, 0.009347071647644043, 0.009417728424072265, 0.009387007713317871, 0.020369472503662108, 0.009386943817138673, 0.00941158390045166, 0.009489407539367676, 0.009350144386291503, 0.009445376396179199, 0.009415679931640625, 0.009382911682128906, 0.009376768112182618, 0.009334783554077148, 0.009360383987426758, 0.009356287956237793, 0.009042943954467773, 0.009034751892089844, 0.00901632022857666, 0.009125951766967774, 0.009421759605407715, 0.009341952323913574, 0.009376768112182618, 0.009378815650939941, 0.009407487869262696, 0.00940236759185791, 0.009357312202453612, 0.009388031959533692, 0.009390080451965332, 0.009354240417480468, 0.009365504264831542, 0.009408512115478516, 0.009374719619750976, 0.009379839897155762, 0.00954470443725586, 0.009374719619750976, 0.009392127990722657, 0.009352191925048828, 0.009366527557373047, 0.009401344299316406, 0.009376768112182618, 0.00940544033050537, 0.009312255859375, 0.009390080451965332, 0.009294879913330079, 0.009375712394714356, 0.009326592445373535, 0.009350144386291503, 0.009347071647644043, 0.009352191925048828, 0.009318400382995605, 0.009374719619750976, 0.009366527557373047, 0.009358336448669433, 0.009334783554077148, 0.009422847747802734, 0.009350144386291503, 0.009387007713317871, 0.009380864143371583, 0.009333760261535644, 0.009379839897155762, 0.009367551803588867, 0.009365504264831542, 0.009551872253417968, 0.009374752044677734, 0.009376735687255859, 0.009362431526184082, 0.02022502326965332, 0.009356287956237793, 0.009366527557373047, 0.009396224021911622, 0.009382911682128906, 0.009408512115478516, 0.009417792320251465, 0.009409472465515137, 0.009399295806884766, 0.009431039810180664, 0.009483263969421387, 0.009505791664123535, 0.009360383987426758, 0.009399328231811523, 0.00938595199584961, 0.00963276767730713, 0.009447423934936524, 0.009378815650939941, 0.009351167678833008, 0.009379839897155762, 0.009380864143371583, 0.009340928077697755, 0.009352191925048828, 0.009366527557373047, 0.009370623588562011, 0.00942080020904541, 0.009562111854553223, 0.009361408233642577, 0.0094269437789917, 0.009395263671875, 0.009375679969787597, 0.00939417552947998, 0.009401375770568848, 0.0093603515625, 0.00940544033050537, 0.009372703552246094, 0.009397215843200683, 0.009358367919921874, 0.00936137580871582, 0.0094136323928833, 0.009362431526184082, 0.009375743865966797, 0.009368576049804688, 0.009373696327209472, 0.009351167678833008, 0.009366527557373047, 0.00939724826812744, 0.009398271560668945, 0.009376768112182618, 0.009398271560668945, 0.009424896240234374, 0.00951916790008545, 0.00941152000427246, 0.009408512115478516, 0.009423871994018555, 0.009379839897155762, 0.009460736274719238, 0.009418815612792968, 0.009407423973083496, 0.009391103744506836, 0.009407487869262696, 0.009326592445373535, 0.009367551803588867, 0.020311040878295897, 0.009393152236938476, 0.009387007713317871, 0.00943824005126953, 0.00947094440460205, 0.009391103744506836, 0.009444352149963378, 0.009392127990722657, 0.009441280364990234, 0.009381888389587402, 0.009382911682128906, 0.009432064056396485, 0.009390080451965332, 0.009404416084289552, 0.009379839897155762, 0.009409536361694336, 0.009388031959533692, 0.009379839897155762, 0.009358336448669433, 0.009359359741210938, 0.009364480018615723, 0.009376799583435059, 0.00947913646697998, 0.009440256118774413, 0.009364480018615723, 0.009398271560668945, 0.009425919532775879, 0.009498623847961426, 0.010399744033813477, 0.01016320037841797, 0.009984000205993653, 0.009529343605041504, 0.009517056465148926, 0.00940339183807373, 0.009406463623046875, 0.009467904090881347, 0.009425919532775879, 0.009414655685424805, 0.009455615997314454, 0.009455615997314454, 0.009335807800292969, 0.009533439636230469, 0.009387007713317871, 0.009384960174560546, 0.009364480018615723, 0.009416704177856445, 0.009443327903747559, 0.009438207626342773, 0.00940236759185791, 0.00941260814666748, 0.009347071647644043, 0.009417728424072265, 0.009505791664123535, 0.009463808059692384, 0.009415679931640625, 0.009406463623046875, 0.009361408233642577, 0.009375743865966797, 0.009380864143371583, 0.009449472427368164, 0.009337856292724609, 0.009352191925048828, 0.009358336448669433, 0.020303871154785155, 0.009354240417480468, 0.00935321617126465, 0.009369600296020507, 0.00939625644683838, 0.00946889591217041, 0.009440256118774413, 0.009500672340393066, 0.009469951629638672, 0.009404416084289552, 0.009439231872558594, 0.009364480018615723, 0.009364480018615723, 0.0094269437789917, 0.009391103744506836, 0.00942080020904541, 0.009268223762512207, 0.00923033618927002, 0.009234432220458985, 0.009346048355102539, 0.009430015563964844, 0.009388031959533692, 0.009385984420776367, 0.009362431526184082, 0.0094269437789917, 0.009372672080993653, 0.009367551803588867, 0.009368576049804688, 0.009373696327209472, 0.009351167678833008, 0.009515007972717286, 0.009381888389587402, 0.009382911682128906, 0.009369600296020507, 0.00942182445526123, 0.009378815650939941, 0.009385984420776367, 0.009350175857543946, 0.00926204776763916, 0.009285632133483887, 0.0092293119430542, 0.00932863998413086, 0.00981503963470459, 0.009438207626342773, 0.009371647834777832, 0.009431039810180664, 0.009370623588562011, 0.009376799583435059, 0.009368543624877929, 0.00940236759185791, 0.009564160346984863, 0.009388031959533692, 0.009339903831481934, 0.009318400382995605, 0.009354240417480468, 0.009439231872558594, 0.009404416084289552, 0.009284607887268067, 0.009375743865966797, 0.009203712463378906, 0.00930611228942871, 0.009346048355102539, 0.009347071647644043, 0.020338687896728515, 0.00941260814666748, 0.009424896240234374, 0.009443327903747559, 0.009406463623046875, 0.009384960174560546, 0.009374719619750976, 0.009355263710021973, 0.009378815650939941, 0.009374719619750976, 0.009382911682128906, 0.0094136323928833, 0.009392127990722657, 0.009350144386291503, 0.00941977596282959, 0.009393152236938476, 0.009387007713317871, 0.009340928077697755, 0.009350144386291503, 0.009312255859375, 0.009339903831481934, 0.009335807800292969, 0.009335807800292969, 0.009400320053100587, 0.009277440071105958, 0.00921497631072998, 0.009347071647644043, 0.009434111595153808, 0.009381888389587402, 0.00932249641418457, 0.009406463623046875, 0.009371647834777832, 0.009387007713317871, 0.009362431526184082, 0.009399295806884766, 0.009387071609497071, 0.009383872032165527, 0.009370623588562011, 0.009370688438415528, 0.00935929584503174, 0.009312255859375, 0.009367551803588867, 0.009354240417480468, 0.009327679634094239, 0.009421759605407715, 0.0093306884765625, 0.009307135581970214, 0.009380864143371583, 0.009349120140075684, 0.00960416030883789, 0.009457599639892579, 0.009356287956237793, 0.009372672080993653, 0.009357312202453612, 0.00932044792175293, 0.009399295806884766, 0.009309184074401856, 0.009332736015319825, 0.009364480018615723, 0.009430015563964844, 0.009346048355102539, 0.009357312202453612, 0.009362431526184082, 0.02042982482910156, 0.009332736015319825, 0.00940339183807373, 0.009401344299316406, 0.00942796802520752, 0.00952627182006836, 0.009342975616455078, 0.009407487869262696, 0.009401344299316406, 0.009349120140075684, 0.009366527557373047, 0.009360383987426758, 0.00929695987701416, 0.009433024406433106, 0.009350144386291503, 0.009371647834777832, 0.009337856292724609, 0.009391136169433593, 0.009323488235473633, 0.009282560348510742, 0.009387007713317871, 0.009313280105590821, 0.009338879585266113, 0.009383935928344727, 0.00930611228942871, 0.009297951698303223, 0.009348064422607423, 0.009323519706726074, 0.00935321617126465, 0.009315327644348144, 0.00931123161315918, 0.009369600296020507, 0.009181183815002441, 0.009541631698608399, 0.009346048355102539, 0.009332736015319825, 0.009294848442077636, 0.009331711769104004, 0.009356287956237793, 0.009335807800292969, 0.009363455772399902, 0.009385984420776367, 0.009350144386291503, 0.009335807800292969, 0.009331711769104004, 0.00935321617126465, 0.009378815650939941, 0.009387007713317871, 0.00931123161315918, 0.009445376396179199, 0.009382911682128906, 0.009392127990722657, 0.009367551803588867, 0.009335807800292969, 0.00932249641418457, 0.00941977596282959, 0.009341952323913574, 0.00930611228942871, 0.009356287956237793, 0.00931430435180664, 0.00931942367553711, 0.009333760261535644, 0.009379839897155762]",tokens/s,105.21310781488594,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,2,2,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/2/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669324c1-00aaa9ee2851a28f07cd88fb;acb1d094-4274-418b-a2e4-0910c92339b4) Repository Not Found for url: https://huggingface.co/2/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: 2 is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/xglm-564M,,cuda,0,42,,,,,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,d35829e539df8480b726c647eeabf91e41eae047,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,,MB,1409.2288,2403.86048,0.0,1757.413376,1301.782528,s,12,0.20092431831359867,0.016743693192799888,0.0005845947355924244,0.016412447929382325,0.017577545166015625,0.01788786821365356,0.01813288827896118,"[0.018194143295288085, 0.01703993606567383, 0.017637279510498045, 0.01698348808288574, 0.016280256271362304, 0.016399551391601562, 0.016408575057983397, 0.016375999450683593, 0.01641632080078125, 0.01635270309448242, 0.01640096092224121, 0.016435104370117186]",tokens/s,15289.338920166369,kWh,1.972021156705348e-07,1.0805755274807403e-07,6.430013491252678e-07,9.482610175438766e-07,tokens/kWh,269967862.5016922,MB,1409.2288,2403.86048,0.0,1757.413376,1324.7616,s,12,10.47739764404297,0.8731164703369142,0.012463386110935145,0.871078643798828,0.8861030456542969,0.8963321441650391,0.9051717022705079,"[0.907381591796875, 0.8711703491210937, 0.8872916870117188, 0.8711393432617187, 0.8620211791992187, 0.8750968017578125, 0.8642526245117188, 0.8656286010742188, 0.8710179443359375, 0.8599682006835937, 0.8670240478515625, 0.8754052734375]",tokens/s,72.15532193052073,kWh,1.0350986601054533e-05,5.671367858698863e-06,2.157481735104288e-05,3.759717181079627e-05,tokens/kWh,1675657.9541950852,,s,755,10.645324805259706,0.014099767953986363,0.002072898476441421,0.013693951606750488,0.014268006706237794,0.014471187019348145,0.03049243476867676,"[0.014705663681030273, 0.014632960319519044, 0.014416959762573242, 0.014392255783081054, 0.014154751777648926, 0.014309375762939454, 0.014519295692443847, 0.01438822364807129, 0.01427353572845459, 0.014445568084716797, 0.014393343925476074, 0.014351360321044922, 0.014326815605163573, 0.014907423973083495, 0.014781375885009766, 0.014867456436157226, 0.014431232452392579, 0.014438400268554688, 0.0142673921585083, 0.01437183952331543, 0.014529536247253418, 0.01427455997467041, 0.014268416404724121, 0.014256128311157227, 0.014230527877807618, 0.014222335815429688, 0.014404607772827148, 0.014523391723632812, 0.01447116756439209, 0.014431232452392579, 0.014433279991149902, 0.014166015625, 0.014306303977966308, 0.014012415885925293, 0.01410051155090332, 0.014724063873291015, 0.014698495864868164, 0.014879743576049804, 0.014593024253845215, 0.014564352035522461, 0.014632960319519044, 0.014234623908996581, 0.014155776023864745, 0.01419878387451172, 0.014146559715270996, 0.014079999923706055, 0.014119935989379882, 0.014319616317749024, 0.014210047721862793, 0.014202879905700684, 0.01417728042602539, 0.014112768173217773, 0.014146559715270996, 0.014128128051757812, 0.014600192070007324, 0.015132736206054688, 0.014814144134521484, 0.014343168258666992, 0.014254079818725587, 0.014228480339050293, 0.014252032279968262, 0.014241791725158692, 0.03164057540893555, 0.014221311569213867, 0.014102527618408203, 0.014171135902404786, 0.014193663597106934, 0.0141844482421875, 0.014132224082946777, 0.014152704238891601, 0.013701120376586913, 0.013706239700317382, 0.01367244815826416, 0.013661184310913087, 0.013692928314208984, 0.013634559631347656, 0.0136878080368042, 0.01365503978729248, 0.013643775939941406, 0.013627391815185547, 0.013661184310913087, 0.013625344276428223, 0.013674495697021484, 0.013752320289611816, 0.013720576286315917, 0.01368883228302002, 0.013664256095886231, 0.013664256095886231, 0.013619199752807617, 0.013683712005615235, 0.013740032196044923, 0.01365503978729248, 0.013674495697021484, 0.013652992248535157, 0.013678591728210449, 0.013606911659240722, 0.013622271537780761, 0.01363046360015869, 0.013688896179199219, 0.013662143707275391, 0.01367961597442627, 0.013623295783996582, 0.013685759544372558, 0.013658111572265624, 0.013594623565673827, 0.013684736251831055, 0.013670399665832519, 0.013578240394592284, 0.013699071884155273, 0.013613056182861329, 0.013648896217346192, 0.01365503978729248, 0.013649920463562011, 0.013689855575561523, 0.013637632369995116, 0.013644800186157227, 0.013636608123779297, 0.013639679908752441, 0.013623295783996582, 0.013640704154968262, 0.013664256095886231, 0.013670399665832519, 0.013669376373291015, 0.014028800010681153, 0.019300352096557616, 0.032573440551757815, 0.014261247634887696, 0.014099519729614257, 0.014151616096496582, 0.01417420768737793, 0.014179327964782714, 0.014352383613586426, 0.014143487930297852, 0.014120960235595703, 0.014172160148620605, 0.014102527618408203, 0.01439641571044922, 0.014232576370239258, 0.014088191986083985, 0.014140416145324707, 0.014141440391540527, 0.01417523193359375, 0.014153727531433105, 0.014125056266784668, 0.01417523193359375, 0.014141440391540527, 0.014104576110839843, 0.014156831741333007, 0.014214112281799316, 0.014169088363647461, 0.014213120460510254, 0.014171135902404786, 0.01419878387451172, 0.014145536422729492, 0.014308416366577148, 0.014143424034118652, 0.014193663597106934, 0.014216192245483398, 0.014167039871215821, 0.014216192245483398, 0.013705216407775878, 0.014040063858032227, 0.013678591728210449, 0.013702143669128418, 0.013699071884155273, 0.013637632369995116, 0.013656064033508301, 0.013709312438964843, 0.013692928314208984, 0.01430016040802002, 0.01415782356262207, 0.013697024345397948, 0.013647871971130371, 0.013656064033508301, 0.013745152473449707, 0.013841407775878906, 0.01417728042602539, 0.01426028823852539, 0.014168000221252442, 0.014231552124023437, 0.014161919593811035, 0.014187520027160644, 0.014138400077819824, 0.014204895973205567, 0.014178303718566895, 0.014166015625, 0.014193663597106934, 0.013710335731506347, 0.03161087989807129, 0.014070783615112305, 0.013612031936645508, 0.013648896217346192, 0.013717503547668456, 0.013637632369995116, 0.014560256004333496, 0.014256128311157227, 0.014317567825317384, 0.014187520027160644, 0.013700160026550293, 0.01366419219970703, 0.013799424171447755, 0.014566495895385742, 0.013682592391967773, 0.013919232368469238, 0.014119935989379882, 0.014135295867919923, 0.014183423995971679, 0.014183423995971679, 0.01415171241760254, 0.014193632125854493, 0.014161919593811035, 0.014127103805541993, 0.014155776023864745, 0.01417625617980957, 0.013743103981018067, 0.01367244815826416, 0.013675519943237305, 0.013661184310913087, 0.01376153564453125, 0.013639679908752441, 0.013647871971130371, 0.013667327880859375, 0.013669376373291015, 0.013652992248535157, 0.01366220760345459, 0.013651968002319336, 0.013676544189453126, 0.013700096130371094, 0.013608960151672364, 0.013692928314208984, 0.01368064022064209, 0.013691904067993164, 0.013652992248535157, 0.013620223999023438, 0.013623295783996582, 0.013647871971130371, 0.01365503978729248, 0.013683712005615235, 0.01370419216156006, 0.013656064033508301, 0.013607935905456543, 0.013596672058105469, 0.013624320030212403, 0.013619199752807617, 0.013692928314208984, 0.013656064033508301, 0.013703167915344238, 0.013726719856262207, 0.013697024345397948, 0.013637632369995116, 0.013667327880859375, 0.030373888015747072, 0.013656064033508301, 0.013631487846374512, 0.013651968002319336, 0.01368064022064209, 0.013697024345397948, 0.013669376373291015, 0.013636608123779297, 0.013661184310913087, 0.01363865566253662, 0.013664256095886231, 0.01364684772491455, 0.01363046360015869, 0.013675519943237305, 0.01367347240447998, 0.013789183616638183, 0.013643775939941406, 0.013700096130371094, 0.013686783790588379, 0.01368064022064209, 0.013659135818481445, 0.01367347240447998, 0.01367961597442627, 0.013668352127075196, 0.013643775939941406, 0.013624320030212403, 0.013664256095886231, 0.013708288192749024, 0.013807616233825683, 0.013690879821777344, 0.013698047637939453, 0.013656064033508301, 0.013675583839416503, 0.013654975891113282, 0.013692928314208984, 0.013677599906921387, 0.013669343948364257, 0.013613056182861329, 0.013727744102478028, 0.013675519943237305, 0.013705216407775878, 0.013718527793884277, 0.01376255989074707, 0.013684736251831055, 0.013721599578857421, 0.013749247550964355, 0.013674495697021484, 0.013710335731506347, 0.013679679870605468, 0.01367750358581543, 0.013644800186157227, 0.013690912246704101, 0.013679583549499512, 0.013713408470153808, 0.013660160064697266, 0.013658111572265624, 0.013635583877563476, 0.013644800186157227, 0.013644800186157227, 0.013685759544372558, 0.01367347240447998, 0.013633536338806153, 0.013651968002319336, 0.030510080337524413, 0.01359769630432129, 0.013637632369995116, 0.01368064022064209, 0.013633536338806153, 0.013670399665832519, 0.013633536338806153, 0.013642751693725585, 0.013647871971130371, 0.013639679908752441, 0.013618176460266113, 0.013666303634643554, 0.013957119941711426, 0.014775296211242676, 0.014713855743408203, 0.014245887756347657, 0.014179327964782714, 0.014214143753051758, 0.014208000183105468, 0.014145536422729492, 0.014137344360351562, 0.014107647895812989, 0.014128128051757812, 0.014220288276672363, 0.014130175590515137, 0.014222335815429688, 0.01420083236694336, 0.014121983528137207, 0.014140416145324707, 0.013974528312683105, 0.013905920028686524, 0.014087167739868164, 0.014088191986083985, 0.014150655746459961, 0.014471232414245605, 0.014548928260803222, 0.014204928398132324, 0.01415987205505371, 0.013767680168151856, 0.013620287895202636, 0.013629376411437988, 0.013637632369995116, 0.013721599578857421, 0.013661184310913087, 0.01370419216156006, 0.013661184310913087, 0.013692928314208984, 0.013658111572265624, 0.013641728401184081, 0.01366220760345459, 0.013623295783996582, 0.013637632369995116, 0.013644800186157227, 0.01363046360015869, 0.013632512092590332, 0.013693951606750488, 0.013657088279724122, 0.01373084831237793, 0.013635552406311036, 0.013610048294067383, 0.013654975891113282, 0.013621248245239258, 0.01368166446685791, 0.030512128829956055, 0.013631520271301269, 0.013653984069824219, 0.013642751693725585, 0.013632512092590332, 0.013700096130371094, 0.013603839874267578, 0.01366220760345459, 0.013629440307617188, 0.013697024345397948, 0.013683712005615235, 0.013813759803771973, 0.013686783790588379, 0.013682687759399414, 0.013627391815185547, 0.013643775939941406, 0.013648896217346192, 0.013674495697021484, 0.013598719596862792, 0.013697024345397948, 0.013709312438964843, 0.013669376373291015, 0.013647871971130371, 0.013661184310913087, 0.01366323184967041, 0.013621248245239258, 0.01368166446685791, 0.013708288192749024, 0.013648896217346192, 0.01367142391204834, 0.013691935539245605, 0.013931488037109375, 0.01375334358215332, 0.013651968002319336, 0.013669376373291015, 0.013744128227233888, 0.013641728401184081, 0.013642751693725585, 0.01358950424194336, 0.013714431762695312, 0.013628416061401367, 0.01363871955871582, 0.01360377597808838, 0.013645824432373046, 0.014258175849914552, 0.014231552124023437, 0.014106623649597168, 0.014112768173217773, 0.014113792419433594, 0.014199808120727539, 0.013643775939941406, 0.013726719856262207, 0.013635583877563476, 0.013625344276428223, 0.013649920463562011, 0.0136878080368042, 0.013631487846374512, 0.01359769630432129, 0.013755423545837402, 0.013459424018859863, 0.013518912315368652, 0.013725631713867188, 0.013658111572265624, 0.030483455657958985, 0.013664256095886231, 0.013645824432373046, 0.013864959716796875, 0.013658111572265624, 0.013682687759399414, 0.013627391815185547, 0.01363257598876953, 0.013691840171813964, 0.013911040306091308, 0.014115839958190919, 0.014071807861328126, 0.014092288017272948, 0.01429094409942627, 0.01417420768737793, 0.014114815711975098, 0.013666303634643554, 0.01364684772491455, 0.013710335731506347, 0.013643775939941406, 0.013657088279724122, 0.013640704154968262, 0.013684736251831055, 0.013631487846374512, 0.013724672317504882, 0.013611007690429687, 0.013648896217346192, 0.013643775939941406, 0.013708288192749024, 0.013719552040100098, 0.013682687759399414, 0.013616127967834473, 0.013668352127075196, 0.013629440307617188, 0.013669376373291015, 0.013660160064697266, 0.01363046360015869, 0.013675519943237305, 0.013636608123779297, 0.013518879890441895, 0.013538271903991698, 0.01365401554107666, 0.013634559631347656, 0.013656064033508301, 0.013670399665832519, 0.013604864120483399, 0.013684736251831055, 0.013747200012207032, 0.013735936164855958, 0.01365401554107666, 0.013676544189453126, 0.013658111572265624, 0.01377996826171875, 0.01427455997467041, 0.014128128051757812, 0.014007295608520508, 0.013503487586975099, 0.013629440307617188, 0.013716480255126954, 0.013656064033508301, 0.013667327880859375, 0.013644800186157227, 0.01366323184967041, 0.030502975463867186, 0.01367955207824707, 0.013595647811889648, 0.013651968002319336, 0.013606911659240722, 0.01367244815826416, 0.013631487846374512, 0.013664256095886231, 0.01363865566253662, 0.013643775939941406, 0.013598719596862792, 0.013618176460266113, 0.013642751693725585, 0.013689855575561523, 0.013645824432373046, 0.013623295783996582, 0.013737983703613281, 0.013752320289611816, 0.013656064033508301, 0.013596672058105469, 0.01364684772491455, 0.013616127967834473, 0.013684736251831055, 0.013656064033508301, 0.014016511917114258, 0.013742079734802246, 0.015156224250793457, 0.014486528396606446, 0.014292991638183594, 0.014235648155212402, 0.014156800270080566, 0.01360588836669922, 0.013707263946533203, 0.01367142391204834, 0.013633536338806153, 0.013642751693725585, 0.013598719596862792, 0.013723648071289063, 0.013882368087768555, 0.014111743927001954, 0.014193663597106934, 0.014116864204406738, 0.014128128051757812, 0.014109760284423829, 0.013642687797546387, 0.013706239700317382, 0.013608960151672364, 0.013640704154968262, 0.013864959716796875, 0.013759488105773926, 0.013676544189453126, 0.013721599578857421, 0.013670399665832519, 0.01368883228302002, 0.013652992248535157, 0.014693375587463378, 0.014191616058349609, 0.014254079818725587, 0.014138367652893067, 0.013615103721618652, 0.013707263946533203, 0.013793279647827148, 0.013750271797180176, 0.03043328094482422, 0.013585408210754395, 0.013641728401184081, 0.013623295783996582, 0.013744128227233888, 0.013713408470153808, 0.01372265625, 0.013668319702148438, 0.013683712005615235, 0.01364684772491455, 0.013624320030212403, 0.013634559631347656, 0.013637632369995116, 0.013637632369995116, 0.013632512092590332, 0.013657088279724122, 0.013640704154968262, 0.01363865566253662, 0.013776896476745605, 0.013639679908752441, 0.01365503978729248, 0.013621248245239258, 0.01365401554107666, 0.013637696266174316, 0.013674431800842286, 0.013685759544372558, 0.01367347240447998, 0.013616127967834473, 0.013478912353515626, 0.01344819164276123, 0.013637696266174316, 0.013530048370361328, 0.013593600273132325, 0.013678591728210449, 0.013645824432373046, 0.013633536338806153, 0.013621248245239258, 0.013647871971130371, 0.01365401554107666, 0.013643775939941406, 0.013617152214050294, 0.013726719856262207, 0.013623295783996582, 0.013659135818481445, 0.013664256095886231, 0.013709312438964843, 0.013676544189453126, 0.013590527534484862, 0.01366323184967041, 0.013670399665832519, 0.013714431762695312, 0.013631487846374512, 0.01363865566253662, 0.013693951606750488, 0.01377996826171875, 0.013677568435668945, 0.013625344276428223, 0.013642751693725585, 0.013664256095886231, 0.013628416061401367, 0.013642751693725585, 0.013484031677246093, 0.013549568176269532, 0.03052448081970215, 0.013649855613708496, 0.013633536338806153, 0.013502464294433594, 0.013503487586975099, 0.013628416061401367, 0.013903871536254882, 0.013701120376586913, 0.013594623565673827, 0.013512767791748048, 0.01355667209625244, 0.01366220760345459, 0.01377791976928711, 0.013627391815185547, 0.013657088279724122, 0.0141844482421875, 0.014118911743164063, 0.014150655746459961, 0.014105600357055664, 0.014138367652893067, 0.014142463684082032, 0.01417728042602539, 0.014208000183105468, 0.014123007774353028, 0.01367961597442627, 0.01364684772491455, 0.013851648330688476, 0.013729855537414551, 0.013717439651489258, 0.01369808006286621, 0.013657055854797363, 0.01368064022064209, 0.013719552040100098, 0.013684736251831055, 0.01368064022064209, 0.013650943756103515, 0.013739007949829102, 0.013710335731506347, 0.013690879821777344, 0.013708288192749024, 0.013667327880859375, 0.013632512092590332, 0.01370419216156006, 0.013832192420959472, 0.013745152473449707, 0.013731840133666993, 0.013822976112365723, 0.013692928314208984, 0.013713408470153808, 0.013742079734802246, 0.013706239700317382, 0.01376255989074707, 0.013720576286315917, 0.01367244815826416, 0.013732864379882812, 0.013719552040100098, 0.013750271797180176, 0.013684736251831055, 0.013742143630981446, 0.013717439651489258, 0.013693984031677247, 0.013608927726745606, 0.01360588836669922, 0.03060736083984375, 0.013993984222412109, 0.013774847984313965, 0.013719552040100098, 0.013713408470153808, 0.013724672317504882, 0.013661184310913087, 0.01376153564453125, 0.013720576286315917, 0.013742079734802246, 0.013689888000488282, 0.013784031867980957, 0.013717503547668456, 0.013714431762695312, 0.014386176109313965, 0.01429100799560547, 0.013773759841918946, 0.013821951866149903, 0.013708288192749024, 0.013699071884155273, 0.013805567741394043, 0.013757439613342285, 0.01375436782836914, 0.01429094409942627, 0.014359552383422852, 0.014256128311157227, 0.014210047721862793, 0.014270463943481445, 0.013602815628051757, 0.013836288452148437, 0.014110719680786133, 0.014208000183105468, 0.014219264030456542, 0.014232576370239258, 0.01437286376953125, 0.014454848289489746, 0.014202816009521484, 0.014205951690673829, 0.01409331226348877, 0.014435327529907227, 0.014436351776123046, 0.013774847984313965, 0.013735936164855958, 0.013751296043395997, 0.013711359977722168, 0.013729791641235351, 0.013737983703613281, 0.01374617576599121, 0.013714431762695312, 0.013716480255126954, 0.01367244815826416, 0.013723648071289063, 0.013692928314208984, 0.013703167915344238, 0.013792320251464844, 0.013847488403320313, 0.013676544189453126, 0.0136079683303833, 0.013741024017333984, 0.013666303634643554, 0.013679648399353027, 0.013644767761230468, 0.013708288192749024]",tokens/s,70.92315300956953,,,main,False,False,,,,,,,,,,,,,,,,,,,,,,, float16-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-14B,Qwen/Qwen-14B,cuda,0,42,,,True,,,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",qwen,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,huggyllama/llama-7b,huggyllama/llama-7b,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.1,,,,MB,1256.435712,14529.593344,0.0,13883.14624,13705.186304,s,10,0.9733994750976562,0.09733994750976563,0.00037133338389398903,0.0972454376220703,0.09747267303466797,0.09793678359985351,0.09830807205200195,"[0.09840089416503907, 0.0972484130859375, 0.09692841339111329, 0.09732080078125, 0.09719913482666015, 0.09736953735351563, 0.09724246215820312, 0.09730467224121093, 0.09718070220947266, 0.09720444488525391]",tokens/s,2629.9582704656464,kWh,1.1503372104293963e-06,6.30330825863461e-07,6.299953799398075e-06,8.080621835690931e-06,tokens/kWh,31680730.17218616,MB,1256.435712,14529.593344,0.0,13883.14624,13856.22528,s,10,22.825175292968748,2.282517529296875,0.0019184782227131703,2.2821934814453124,2.2837198486328125,2.2856649780273437,2.2872210815429685,"[2.281288818359375, 2.2814697265625, 2.279878662109375, 2.28247021484375, 2.28216357421875, 2.281911376953125, 2.282871826171875, 2.287610107421875, 2.28328759765625, 2.282223388671875]",tokens/s,27.60110237550159,kWh,2.6882328671416297e-05,1.4732435174316816e-05,0.00011731889509560169,0.00015893365894133484,tokens/kWh,396391.80535857665,,s,629,23.663276023864768,0.037620470626176065,0.011627260591752632,0.03621376037597656,0.03636470413208008,0.03641712570190429,0.13402455993652343,"[0.03610009765625, 0.03621376037597656, 0.03596492767333984, 0.03605401611328125, 0.036021247863769534, 0.036089855194091795, 0.03613183975219727, 0.03608063888549805, 0.03601615905761719, 0.03608163070678711, 0.03603968048095703, 0.03607961654663086, 0.03606425476074219, 0.03615334320068359, 0.03612979125976563, 0.03618611145019531, 0.036057086944580076, 0.03612057495117187, 0.03602739334106445, 0.0361267204284668, 0.03606937789916992, 0.03617484664916992, 0.03607551956176758, 0.036154369354248046, 0.03612057495117187, 0.03623321533203125, 0.03616767883300781, 0.03624652862548828, 0.03609804916381836, 0.03628441619873047, 0.03612979125976563, 0.036162559509277346, 0.036171775817871094, 0.036201473236083984, 0.036116481781005856, 0.03630694580078125, 0.03614720153808594, 0.03620761489868164, 0.036155391693115234, 0.036222976684570314, 0.03618406295776367, 0.03628339385986328, 0.036155391693115234, 0.03626803207397461, 0.03616665649414062, 0.03627315139770508, 0.03618201446533203, 0.036291584014892575, 0.03629363250732422, 0.03651686477661133, 0.03641958236694336, 0.036547584533691405, 0.03626399993896484, 0.03632940673828125, 0.03628646469116211, 0.036395008087158204, 0.03629264068603515, 0.036413406372070314, 0.03625574493408203, 0.03643801498413086, 0.036280319213867186, 0.03637145614624023, 0.1340590057373047, 0.03606220626831055, 0.03607551956176758, 0.03600384140014649, 0.03609190368652344, 0.0360816650390625, 0.03611347198486328, 0.036071361541748045, 0.03611340713500977, 0.036067329406738284, 0.03613695907592773, 0.036037631988525394, 0.0361420783996582, 0.036057086944580076, 0.036157440185546875, 0.036055038452148434, 0.036178943634033206, 0.03611443328857422, 0.036144126892089845, 0.036057086944580076, 0.03619123077392578, 0.03609702301025391, 0.03616159820556641, 0.03610822296142578, 0.036176895141601564, 0.03614617538452149, 0.03623526382446289, 0.03608883285522461, 0.0362239990234375, 0.03610726547241211, 0.03624857711791992, 0.036203521728515625, 0.036296703338623046, 0.03619430541992188, 0.036206592559814454, 0.036173824310302735, 0.036275199890136715, 0.03620249557495117, 0.03624755096435547, 0.03620044708251953, 0.03625164794921875, 0.03616563034057617, 0.03628441619873047, 0.03625881576538086, 0.03632332611083984, 0.036413440704345705, 0.03636636734008789, 0.03630179214477539, 0.03638579177856445, 0.03624652862548828, 0.03631923294067383, 0.036245502471923825, 0.036316158294677735, 0.03623116683959961, 0.0362977294921875, 0.03623321533203125, 0.0363765754699707, 0.03633561706542969, 0.036410369873046876, 0.03628441619873047, 0.036413440704345705, 0.03626291275024414, 0.0364031982421875, 0.13376205444335937, 0.036067329406738284, 0.03623219299316406, 0.0361451530456543, 0.03606835174560547, 0.036050945281982424, 0.03609088134765625, 0.036037631988525394, 0.03605299377441406, 0.03598745727539063, 0.03612364959716797, 0.035983360290527344, 0.036071422576904294, 0.03602431869506836, 0.036103168487548826, 0.03605913543701172, 0.036122623443603515, 0.03608268737792969, 0.036168704986572264, 0.03604582214355469, 0.03616460800170898, 0.03611852645874023, 0.03615334320068359, 0.03614003372192383, 0.036157440185546875, 0.03611033630371094, 0.0362239990234375, 0.03611443328857422, 0.03623833465576172, 0.03625881576538086, 0.036431873321533206, 0.03622809600830078, 0.03619123077392578, 0.03614822387695313, 0.036257793426513675, 0.03617996978759765, 0.03623116683959961, 0.03611545562744141, 0.03616563034057617, 0.03611033630371094, 0.03625881576538086, 0.03624038314819336, 0.03630080032348633, 0.03615129470825195, 0.03625164794921875, 0.03625164794921875, 0.036261886596679685, 0.03621376037597656, 0.03626700973510742, 0.036122623443603515, 0.03628134536743164, 0.03618815994262695, 0.03629260635375976, 0.03619839859008789, 0.03628339385986328, 0.036211711883544925, 0.0363059196472168, 0.036190208435058595, 0.036307968139648435, 0.036209663391113284, 0.03632844924926758, 0.03625369644165039, 0.036446208953857424, 0.1341265869140625, 0.036038654327392575, 0.036087806701660154, 0.036018177032470705, 0.03612160110473633, 0.036029441833496094, 0.0361082878112793, 0.03606118392944336, 0.03617484664916992, 0.03613183975219727, 0.03647078323364258, 0.03630387115478516, 0.03627212905883789, 0.036131904602050784, 0.03622393417358399, 0.03615027236938476, 0.03619839859008789, 0.036152320861816405, 0.03620556640625, 0.036111358642578126, 0.036173824310302735, 0.03606937789916992, 0.03614310455322266, 0.03614003372192383, 0.03625369644165039, 0.03621478271484375, 0.03623731231689453, 0.03613491058349609, 0.03629574584960937, 0.03612768173217774, 0.03625164794921875, 0.036171775817871094, 0.036302879333496095, 0.036175838470458986, 0.036170753479003906, 0.03616563034057617, 0.03629260635375976, 0.03613183975219727, 0.03625267028808594, 0.03623014450073242, 0.036310016632080076, 0.036168704986572264, 0.0362762222290039, 0.03616665649414062, 0.03648614501953125, 0.0362158088684082, 0.03628953552246094, 0.03620249557495117, 0.03630182266235352, 0.036241409301757815, 0.036318206787109376, 0.0363059196472168, 0.03643699264526367, 0.03621683120727539, 0.03628851318359375, 0.03629264068603515, 0.03633660888671875, 0.03628851318359375, 0.036347904205322266, 0.036206592559814454, 0.036362239837646484, 0.036278270721435545, 0.03643801498413086, 0.13411737060546874, 0.036025375366210935, 0.03606934356689453, 0.03601510238647461, 0.036155391693115234, 0.03602227020263672, 0.03608473587036133, 0.03605811309814453, 0.036162559509277346, 0.036078590393066406, 0.03609907150268555, 0.036050945281982424, 0.0361451530456543, 0.03615027236938476, 0.036203521728515625, 0.03611340713500977, 0.03621478271484375, 0.03617279815673828, 0.03617792129516602, 0.03618099212646484, 0.03620556640625, 0.036071422576904294, 0.03611545562744141, 0.036073566436767575, 0.03621161651611328, 0.036132865905761716, 0.036160511016845705, 0.03608473587036133, 0.03627212905883789, 0.03618201446533203, 0.03627212905883789, 0.03616153717041016, 0.03630284881591797, 0.03618304061889648, 0.0362608642578125, 0.03609907150268555, 0.03623628616333008, 0.03621376037597656, 0.03625164794921875, 0.03615027236938476, 0.036241409301757815, 0.03626291275024414, 0.03623731231689453, 0.03641145706176758, 0.03646150588989258, 0.03624038314819336, 0.03630182266235352, 0.0362762222290039, 0.036354049682617184, 0.03625676727294922, 0.03633356857299805, 0.036219905853271485, 0.03638579177856445, 0.03624959945678711, 0.03640524673461914, 0.036354049682617184, 0.036364288330078126, 0.03641958236694336, 0.03639091110229492, 0.03627315139770508, 0.03632844924926758, 0.03624345779418945, 0.03643801498413086, 0.13464781188964844, 0.03604787063598633, 0.03614617538452149, 0.03604787063598633, 0.03610419082641601, 0.03603968048095703, 0.03611340713500977, 0.03607654571533203, 0.0361451530456543, 0.03605923080444336, 0.03619318389892578, 0.03607756805419922, 0.03615334320068359, 0.03606016159057617, 0.03612876892089844, 0.0360816650390625, 0.0361451530456543, 0.03615846252441406, 0.03624857711791992, 0.036055038452148434, 0.03617996978759765, 0.03612364959716797, 0.03618918228149414, 0.03616563034057617, 0.03630284881591797, 0.03625267028808594, 0.03627724838256836, 0.03623321533203125, 0.03632128143310547, 0.03615948867797852, 0.036241409301757815, 0.036178943634033206, 0.03621683120727539, 0.036160511016845705, 0.03622611236572266, 0.03613894271850586, 0.03625888061523438, 0.03615430450439453, 0.03624755096435547, 0.03618406295776367, 0.03623526382446289, 0.036149246215820316, 0.03628339385986328, 0.036176895141601564, 0.03626598358154297, 0.03625881576538086, 0.036310016632080076, 0.03621376037597656, 0.03632537460327148, 0.036242431640625, 0.036348960876464845, 0.03623011016845703, 0.036350975036621096, 0.03623219299316406, 0.0363059196472168, 0.03623526382446289, 0.03631411361694336, 0.03629056167602539, 0.036404254913330075, 0.036235233306884766, 0.036340736389160154, 0.03622809600830078, 0.03642060852050781, 0.13402828979492187, 0.036004863739013675, 0.03602534484863281, 0.03595775985717774, 0.036044830322265624, 0.03599459075927734, 0.03605196762084961, 0.0359741439819336, 0.03605913543701172, 0.035972095489501955, 0.036078590393066406, 0.03600588989257812, 0.03608883285522461, 0.03602534484863281, 0.036057086944580076, 0.03617996978759765, 0.03640115356445312, 0.03615846252441406, 0.036187137603759766, 0.03611443328857422, 0.03639807891845703, 0.03616972732543945, 0.03623731231689453, 0.036193279266357424, 0.03622707366943359, 0.03621785736083984, 0.03622195053100586, 0.036073471069335936, 0.036192256927490236, 0.03611033630371094, 0.03620249557495117, 0.036111358642578126, 0.036208641052246096, 0.03620454406738281, 0.03634995269775391, 0.036278270721435545, 0.036310016632080076, 0.036192256927490236, 0.03622198486328125, 0.036195297241210934, 0.03633356857299805, 0.03624448013305664, 0.036595775604248044, 0.036914112091064454, 0.036413440704345705, 0.0362690544128418, 0.03637247848510742, 0.03630182266235352, 0.03640524673461914, 0.03625369644165039, 0.03636326217651367, 0.03625574493408203, 0.036421630859375, 0.03629568099975586, 0.03634380722045898, 0.03623833465576172, 0.03632025527954102, 0.03628339385986328, 0.03640524673461914, 0.03631411361694336, 0.03642675018310547, 0.03628543853759766, 0.036347904205322266, 0.13416038513183592, 0.03612979125976563, 0.03619123077392578, 0.036117504119873044, 0.03619839859008789, 0.03617996978759765, 0.03621478271484375, 0.036063297271728516, 0.036230079650878905, 0.03615337753295898, 0.036203487396240235, 0.03609395217895508, 0.03620761489868164, 0.036122623443603515, 0.03622092819213867, 0.036116481781005856, 0.036291584014892575, 0.03614726257324219, 0.03615430450439453, 0.03609088134765625, 0.03620761489868164, 0.036089855194091795, 0.036222976684570314, 0.03629363250732422, 0.04064767837524414, 0.036400127410888675, 0.03630697631835938, 0.03615740966796875, 0.03634995269775391, 0.03615846252441406, 0.03626496124267578, 0.03614720153808594, 0.036241409301757815, 0.03618819046020508, 0.036238304138183595, 0.036154369354248046, 0.036278270721435545, 0.036187137603759766, 0.036299774169921875, 0.036185089111328124, 0.03623219299316406, 0.036187137603759766, 0.036359169006347655, 0.03624345779418945, 0.03638886260986328, 0.03628646469116211, 0.036348926544189454, 0.03632230377197266, 0.03640524673461914, 0.03624448013305664, 0.036291584014892575, 0.03616358566284179, 0.036299774169921875, 0.0362158088684082, 0.03630899047851562, 0.03622403335571289, 0.03629155349731445, 0.03626598358154297, 0.036340736389160154, 0.03625471878051758, 0.03638476943969727, 0.03624038314819336, 0.03637247848510742, 0.13407232666015625, 0.03602329635620117, 0.03604377746582031, 0.035996673583984375, 0.03604787063598633, 0.0360079345703125, 0.03604889678955078, 0.03602227020263672, 0.03611340713500977, 0.036122623443603515, 0.036511745452880856, 0.036154369354248046, 0.036222976684570314, 0.036162559509277346, 0.036193279266357424, 0.036173824310302735, 0.03620044708251953, 0.03625369644165039, 0.03621785736083984, 0.036144126892089845, 0.03626291275024414, 0.036187137603759766, 0.03623628616333008, 0.036160511016845705, 0.036242431640625, 0.03621478271484375, 0.036257793426513675, 0.036245502471923825, 0.03621785736083984, 0.03614617538452149, 0.03620044708251953, 0.0361267204284668, 0.03619942474365234, 0.03618304061889648, 0.03624959945678711, 0.036149246215820316, 0.036239360809326174, 0.03623833465576172, 0.03631718444824219, 0.03624345779418945, 0.03631718444824219, 0.0362874870300293, 0.036350975036621096, 0.036203521728515625, 0.03624038314819336, 0.03626598358154297, 0.0363059196472168, 0.036222976684570314, 0.036296703338623046, 0.0362608642578125, 0.03638784027099609, 0.03618201446533203, 0.036377601623535157, 0.036278270721435545, 0.03638886260986328, 0.0369172477722168, 0.036364288330078126, 0.03633152008056641, 0.03633152008056641, 0.036229118347167966, 0.036354049682617184, 0.03620454406738281, 0.036345855712890625, 0.13401496887207032, 0.03605811309814453, 0.036078590393066406, 0.03600896072387695, 0.03609395217895508, 0.036029441833496094, 0.03610214233398437, 0.03604172897338867, 0.03609500885009766, 0.036033504486083986, 0.036109310150146484, 0.03609600067138672, 0.036327457427978514, 0.03635708618164062, 0.03621068954467774, 0.03616460800170898, 0.036241409301757815, 0.03617996978759765, 0.036222015380859375, 0.03611743927001953, 0.03622707366943359, 0.0361451530456543, 0.036241409301757815, 0.03617792129516602, 0.0362342414855957, 0.03619942474365234, 0.03640729522705078, 0.03611545562744141, 0.03621686553955078, 0.036116447448730465, 0.0362342414855957, 0.036168704986572264, 0.03624345779418945, 0.036157440185546875, 0.03620556640625, 0.03614310455322266, 0.03623321533203125, 0.03612057495117187, 0.03621068954467774, 0.0362874870300293, 0.03632025527954102, 0.036171775817871094, 0.036261886596679685, 0.036185089111328124, 0.036275199890136715, 0.03627212905883789, 0.036348926544189454, 0.03630489730834961, 0.03631206512451172, 0.036162559509277346, 0.036340736389160154, 0.03616972732543945, 0.036310016632080076, 0.0361973762512207, 0.03628646469116211, 0.03625267028808594, 0.036334590911865236, 0.036291584014892575, 0.03647180938720703, 0.036350975036621096, 0.036395008087158204, 0.03625062561035156, 0.03650764846801758]",tokens/s,26.581272997265664,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-180B,tiiuae/falcon-180B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 403 Client Error: Forbidden for url: https://huggingface.co/tiiuae/falcon-180B/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-669326f9-5e934d395b98ac5813771999;cd9b2484-e469-43a0-8f3d-6908a7a2c87e) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/tiiuae/falcon-180B/resolve/main/config.json. If you are trying to create or update content,make sure you have a token with the `write` role. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1826, in _raise_on_head_call_error raise LocalEntryNotFoundError( huggingface_hub.utils._errors.LocalEntryNotFoundError: An error happened while trying to locate the file on the Hub and we cannot find the requested files in the local cache. Please check your connection and try again or make sure your Internet connection is on. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 445, in cached_file raise EnvironmentError( OSError: We couldn't connect to 'https://huggingface.co' to load this file, couldn't find it in the cached files and it looks like tiiuae/falcon-180B is not the path to a directory containing a file named config.json. Checkout your internet connection or see how to run the library in offline mode at 'https://huggingface.co/docs/transformers/installation#offline-mode'. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/gpt-neo-1.3B,EleutherAI/gpt-neo-1.3B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1216.96256,3760.717824,0.0,3114.27072,2817.469952,s,10,0.26985433769226075,0.026985433769226075,0.0004369347338127713,0.026887856483459472,0.02711618938446045,0.027684686374664307,0.028139483966827392,"[0.028253183364868165, 0.02686617660522461, 0.02656275177001953, 0.026863008499145507, 0.02690995216369629, 0.02692323112487793, 0.026810752868652345, 0.026909536361694335, 0.026989856719970704, 0.026765888214111327]",tokens/s,9486.599407267631,kWh,3.159318457950245e-07,1.7304473625806753e-07,1.430757977641692e-06,1.919734559694784e-06,tokens/kWh,133351769.23662879,MB,1217.35168,3762.814976,0.0,3116.367872,2885.484544,s,10,11.051232299804688,1.1051232299804687,0.005412843790888853,1.1057021484375,1.1093506469726562,1.1125512390136718,1.1151117126464842,"[1.1157518310546874, 1.096914794921875, 1.09553564453125, 1.108639404296875, 1.1060224609375, 1.1053436279296875, 1.10409619140625, 1.1057786865234376, 1.1056256103515625, 1.1075240478515624]",tokens/s,57.007217196143294,kWh,1.2736028605940368e-05,6.978946035250213e-06,3.7167884012357774e-05,5.688285865354835e-05,tokens/kWh,1107539.2744184819,,s,629,11.273896976470951,0.01792352460488227,0.003236159460334246,0.017551359176635743,0.017711513137817384,0.01790873565673828,0.04453519409179689,"[0.018689023971557618, 0.017946624755859376, 0.01763737678527832, 0.01768550491333008, 0.017695743560791014, 0.01746124839782715, 0.01798246383666992, 0.017761280059814453, 0.01761075210571289, 0.017458175659179686, 0.01744691276550293, 0.017479679107666016, 0.01760767936706543, 0.0176363525390625, 0.01764352035522461, 0.017500160217285156, 0.01756159973144531, 0.017573888778686524, 0.017484800338745117, 0.017525760650634766, 0.017529855728149413, 0.017760255813598632, 0.017754112243652344, 0.017521663665771483, 0.01735577583312988, 0.017472511291503907, 0.017257471084594727, 0.017460224151611328, 0.018155519485473632, 0.01823641586303711, 0.018851839065551757, 0.018335744857788085, 0.01780531120300293, 0.017705984115600586, 0.017770496368408203, 0.01761996841430664, 0.01760358428955078, 0.01775103950500488, 0.018753536224365236, 0.017871871948242187, 0.01759334373474121, 0.017625087738037108, 0.017582080841064454, 0.017532928466796875, 0.017683456420898438, 0.01756675148010254, 0.017619935989379883, 0.017564672470092774, 0.017574911117553712, 0.017589248657226563, 0.017533952713012696, 0.017548288345336914, 0.017538047790527343, 0.017550336837768556, 0.01760665512084961, 0.017514495849609374, 0.017528831481933595, 0.017590272903442384, 0.017514495849609374, 0.017510400772094727, 0.017540096282958984, 0.01763020706176758, 0.04504883193969727, 0.01765273666381836, 0.017551359176635743, 0.01757695960998535, 0.01759846305847168, 0.01766092872619629, 0.017534975051879884, 0.017666048049926757, 0.017745920181274414, 0.017573888778686524, 0.017543167114257813, 0.01758515167236328, 0.01754627227783203, 0.017599456787109374, 0.01744076728820801, 0.01764454460144043, 0.017555456161499023, 0.01760153579711914, 0.017506303787231444, 0.017485824584960938, 0.017526784896850587, 0.017508352279663086, 0.01743974494934082, 0.017639455795288087, 0.01750217628479004, 0.017469440460205078, 0.017557504653930665, 0.01761689567565918, 0.017483776092529296, 0.01756159973144531, 0.017526784896850587, 0.017559551239013673, 0.017542144775390626, 0.017555456161499023, 0.017531967163085936, 0.017474496841430664, 0.017589248657226563, 0.017499135971069335, 0.01756159973144531, 0.017772544860839845, 0.017564672470092774, 0.0176363525390625, 0.017674240112304687, 0.017710079193115236, 0.017574911117553712, 0.01756979179382324, 0.01761894416809082, 0.01701068878173828, 0.016842752456665038, 0.016837631225585938, 0.01681920051574707, 0.016928768157958983, 0.016881664276123046, 0.01686425590515137, 0.016911359786987306, 0.016977920532226562, 0.016935935974121095, 0.016874496459960937, 0.016885759353637696, 0.016955392837524414, 0.016895999908447267, 0.016915456771850586, 0.016884735107421875, 0.043886592864990234, 0.016846847534179688, 0.01697689628601074, 0.016917503356933594, 0.016876544952392578, 0.01680384063720703, 0.016892927169799805, 0.016944128036499022, 0.01686016082763672, 0.01686016082763672, 0.0168724479675293, 0.016841728210449217, 0.016888832092285155, 0.016849920272827147, 0.016890880584716796, 0.016959487915039064, 0.016920576095581053, 0.01683660888671875, 0.016879615783691407, 0.0168222713470459, 0.017484800338745117, 0.017683456420898438, 0.01765273666381836, 0.01764249610900879, 0.01818726348876953, 0.017889280319213868, 0.017571840286254883, 0.017671167373657228, 0.01761996841430664, 0.017517568588256836, 0.017510400772094727, 0.017732608795166017, 0.017558528900146485, 0.017556480407714844, 0.017565696716308594, 0.017530879974365234, 0.017665023803710937, 0.017502208709716797, 0.017565696716308594, 0.01762611198425293, 0.017554431915283202, 0.017559551239013673, 0.017511423110961915, 0.017540096282958984, 0.017551359176635743, 0.017681407928466796, 0.01767628860473633, 0.017743871688842772, 0.0175861759185791, 0.01764454460144043, 0.0175994873046875, 0.01759231948852539, 0.01761075210571289, 0.017710079193115236, 0.017641471862792968, 0.01762713623046875, 0.0176363525390625, 0.017487871170043946, 0.01754521560668945, 0.01759539222717285, 0.017551359176635743, 0.01747865676879883, 0.01765273666381836, 0.04415180969238281, 0.017253376007080077, 0.017581087112426758, 0.017534975051879884, 0.017598432540893556, 0.017502208709716797, 0.01760870361328125, 0.01759539222717285, 0.017597440719604493, 0.017527807235717775, 0.017507328033447265, 0.017571840286254883, 0.017770496368408203, 0.01764659118652344, 0.017558528900146485, 0.017803264617919923, 0.017512447357177736, 0.017558528900146485, 0.017489919662475584, 0.017484832763671875, 0.017585119247436522, 0.017522687911987304, 0.017511423110961915, 0.01785856056213379, 0.01763430404663086, 0.017521663665771483, 0.017456127166748048, 0.017666048049926757, 0.01767731285095215, 0.017597440719604493, 0.017583103179931642, 0.017574911117553712, 0.017547264099121093, 0.01755340766906738, 0.017521663665771483, 0.017555456161499023, 0.0175063362121582, 0.018175968170166014, 0.0182108154296875, 0.018128896713256838, 0.017757183074951173, 0.017743871688842772, 0.017678335189819337, 0.017690624237060547, 0.017580032348632812, 0.01760358428955078, 0.017671167373657228, 0.017617919921875, 0.01759334373474121, 0.01754521560668945, 0.017506303787231444, 0.017691648483276368, 0.017570816040039062, 0.017582080841064454, 0.017565696716308594, 0.01761177635192871, 0.017566719055175782, 0.017507328033447265, 0.017323007583618166, 0.017312768936157227, 0.017415168762207032, 0.0176312313079834, 0.017532928466796875, 0.04468428802490235, 0.017474559783935546, 0.017516544342041016, 0.017480703353881837, 0.017511423110961915, 0.017348608016967772, 0.017106943130493164, 0.017575935363769533, 0.017504255294799806, 0.01744895935058594, 0.017514495849609374, 0.017529855728149413, 0.017485824584960938, 0.017572864532470703, 0.017476608276367187, 0.01756879997253418, 0.017613792419433595, 0.017496063232421876, 0.017555456161499023, 0.017505279541015627, 0.017518592834472657, 0.01748684883117676, 0.017570816040039062, 0.0176363525390625, 0.01756879997253418, 0.017535968780517577, 0.01754425621032715, 0.017514432907104492, 0.017501184463500977, 0.017543167114257813, 0.01747865676879883, 0.01764556884765625, 0.017474559783935546, 0.01776639938354492, 0.01855392074584961, 0.018039743423461915, 0.017580032348632812, 0.01759334373474121, 0.017967103958129883, 0.017674240112304687, 0.017551359176635743, 0.017506303787231444, 0.017479679107666016, 0.017534975051879884, 0.01763737678527832, 0.01760870361328125, 0.017574911117553712, 0.01760767936706543, 0.0175861759185791, 0.01740902328491211, 0.01725644874572754, 0.017288192749023438, 0.017306623458862306, 0.017360895156860352, 0.017570816040039062, 0.01760051155090332, 0.01751865577697754, 0.017597375869750978, 0.01764249610900879, 0.017509376525878906, 0.017543167114257813, 0.017524736404418945, 0.017581056594848633, 0.044935169219970705, 0.017501184463500977, 0.017526784896850587, 0.01749510383605957, 0.01757382392883301, 0.01745715141296387, 0.017950719833374023, 0.017918975830078124, 0.017663999557495116, 0.017452032089233398, 0.017460224151611328, 0.017570816040039062, 0.017482751846313475, 0.017443840026855468, 0.017546239852905272, 0.017548288345336914, 0.017494016647338868, 0.017509376525878906, 0.017321983337402345, 0.017304576873779298, 0.017484800338745117, 0.017484800338745117, 0.017961984634399415, 0.01765376091003418, 0.01760972785949707, 0.01759539222717285, 0.01783500862121582, 0.01747148895263672, 0.017605632781982423, 0.017624063491821287, 0.01761177635192871, 0.01757695960998535, 0.017580032348632812, 0.017552383422851564, 0.017528831481933595, 0.017507328033447265, 0.01765990447998047, 0.017466367721557616, 0.01759539222717285, 0.017674240112304687, 0.017557504653930665, 0.01764454460144043, 0.017634368896484374, 0.01749907112121582, 0.017491968154907226, 0.017529855728149413, 0.01758720016479492, 0.017729536056518554, 0.017573888778686524, 0.01745715141296387, 0.017552383422851564, 0.01748684883117676, 0.017567743301391603, 0.01744588851928711, 0.017624063491821287, 0.01779097557067871, 0.017339391708374022, 0.01722163200378418, 0.017584127426147463, 0.01720217514038086, 0.01720319938659668, 0.01714995193481445, 0.01719910430908203, 0.044906494140625, 0.017486879348754883, 0.01748886489868164, 0.017175552368164062, 0.01723904037475586, 0.018711551666259766, 0.018062335968017578, 0.0176312313079834, 0.01773465538024902, 0.017570816040039062, 0.017490943908691405, 0.017491968154907226, 0.017449983596801756, 0.01744895935058594, 0.017496063232421876, 0.017547264099121093, 0.01760972785949707, 0.017566719055175782, 0.017467391967773437, 0.01728102493286133, 0.017293312072753905, 0.017217536926269532, 0.01747865676879883, 0.017633279800415038, 0.017566719055175782, 0.017525760650634766, 0.017522687911987304, 0.017466367721557616, 0.017516544342041016, 0.017513471603393553, 0.017516544342041016, 0.017563648223876953, 0.017554431915283202, 0.01757900810241699, 0.017557504653930665, 0.017533952713012696, 0.017582080841064454, 0.017469440460205078, 0.017558528900146485, 0.017570816040039062, 0.01778892707824707, 0.017464319229125978, 0.017516544342041016, 0.017449983596801756, 0.017531904220581054, 0.017555456161499023, 0.017695743560791014, 0.017723392486572266, 0.017534975051879884, 0.017505279541015627, 0.017670143127441407, 0.017473535537719728, 0.017556480407714844, 0.017447935104370118, 0.017513471603393553, 0.01759436798095703, 0.017449983596801756, 0.01744588851928711, 0.017031167984008787, 0.016932863235473633, 0.01719398307800293, 0.01720217514038086, 0.017565696716308594, 0.04491571044921875, 0.01760870361328125, 0.017510400772094727, 0.017520639419555666, 0.017476608276367187, 0.017539072036743163, 0.017284095764160155, 0.017367040634155274, 0.01760665512084961, 0.017464319229125978, 0.017328128814697266, 0.017655807495117186, 0.017575935363769533, 0.017333248138427734, 0.01723904037475586, 0.017649663925170898, 0.01760153579711914, 0.017605632781982423, 0.017521663665771483, 0.017498111724853514, 0.017473600387573243, 0.017558464050292967, 0.017501184463500977, 0.017583103179931642, 0.01739571189880371, 0.017261568069458007, 0.01763737678527832, 0.017473535537719728, 0.017500160217285156, 0.01762713623046875, 0.01763942337036133, 0.01759129524230957, 0.017357824325561523, 0.01784012794494629, 0.017518592834472657, 0.017545248031616213, 0.017585119247436522, 0.017558528900146485, 0.017511423110961915, 0.0176680965423584, 0.01762713623046875, 0.018106367111206053, 0.017869823455810546, 0.017564672470092774, 0.01750124740600586, 0.01756051254272461, 0.017540159225463866, 0.017650623321533204, 0.017518592834472657, 0.01748684883117676, 0.01760256004333496, 0.017717248916625978, 0.017596416473388672, 0.01756876754760742, 0.01752992057800293, 0.017594303131103516, 0.017520639419555666, 0.017508384704589843, 0.017456096649169923, 0.017524736404418945, 0.017541120529174805, 0.017539072036743163, 0.017479679107666016, 0.04500582504272461, 0.017509376525878906, 0.01765795135498047, 0.017485727310180665, 0.017546239852905272, 0.017489919662475584, 0.0175994873046875, 0.01759231948852539, 0.017514495849609374, 0.01748684883117676, 0.01756985664367676, 0.017566656112670897, 0.017443840026855468, 0.017473535537719728, 0.017562623977661132, 0.01760767936706543, 0.01765273666381836, 0.017525760650634766, 0.017557504653930665, 0.017583103179931642, 0.017483776092529296, 0.017549312591552735, 0.017543167114257813, 0.01763737678527832, 0.017508352279663086, 0.017542144775390626, 0.017504255294799806, 0.017555456161499023, 0.017514495849609374, 0.017522687911987304, 0.01745305633544922, 0.017617919921875, 0.017795072555541993, 0.01757904052734375, 0.017512416839599608, 0.017518592834472657, 0.017466367721557616, 0.017574911117553712, 0.017460224151611328, 0.017699840545654297, 0.017518592834472657, 0.017543167114257813, 0.017655807495117186, 0.017548288345336914, 0.017550336837768556, 0.01764659118652344, 0.017570816040039062, 0.017693695068359376, 0.017535999298095704, 0.017506303787231444, 0.017551359176635743, 0.017546239852905272, 0.017512447357177736, 0.017511423110961915, 0.017484800338745117, 0.017622016906738282, 0.017491968154907226, 0.01739776039123535, 0.01723904037475586, 0.01724723243713379, 0.017403968811035155, 0.017583040237426757, 0.017810432434082032, 0.04481740951538086, 0.01766092872619629, 0.0175861759185791, 0.017543167114257813, 0.01749504089355469, 0.01744595146179199, 0.01750115203857422, 0.017487840652465822, 0.017513471603393553, 0.01747974395751953, 0.017409984588623046, 0.017584127426147463, 0.017466367721557616, 0.017550336837768556, 0.017533952713012696, 0.01760767936706543, 0.01738035202026367, 0.01721651268005371, 0.017161216735839844, 0.01756159973144531, 0.01765990447998047, 0.017492992401123047, 0.0174704647064209, 0.01756159973144531, 0.01758211135864258, 0.017604576110839844, 0.017641471862792968, 0.017893375396728514, 0.01787494468688965, 0.018922496795654296, 0.017939456939697264, 0.01768550491333008, 0.01759539222717285, 0.017665023803710937, 0.017566719055175782, 0.01761587142944336, 0.017552383422851564, 0.01761689567565918, 0.017505279541015627, 0.01741004753112793, 0.017287168502807617, 0.017533952713012696, 0.017543167114257813, 0.017571903228759764, 0.01751545524597168, 0.017328128814697266, 0.017459199905395507, 0.01766912078857422, 0.01758515167236328, 0.017551359176635743, 0.017480703353881837, 0.017540096282958984, 0.01765273666381836, 0.01760358428955078, 0.017597440719604493, 0.017613824844360353, 0.017547264099121093, 0.01762713623046875, 0.017559551239013673, 0.017625087738037108, 0.017543167114257813, 0.017530879974365234, 0.017511423110961915]",tokens/s,55.79259783132194,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,mistralai/Mistral-7B-v0.1,mistralai/Mistral-7B-v0.1,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.1,,,,MB,1273.683968,15987.113984,0.0,15340.66688,15114.865664,s,10,0.8408147277832031,0.08408147277832032,0.0004271306080976992,0.08409201431274414,0.08439436569213866,0.08472729606628417,0.08499364036560059,"[0.08506022644042968, 0.08346109008789063, 0.08421068572998047, 0.08381267547607422, 0.08432038116455078, 0.08363267517089844, 0.08425023651123047, 0.08397334289550781, 0.08426064300537109, 0.08383277130126954]",tokens/s,3044.665983372349,kWh,9.948000075325134e-07,5.451005580767788e-07,6.758023613697514e-06,8.297924179306807e-06,tokens/kWh,30851089.31682065,MB,1273.683968,15987.113984,0.0,15340.66688,15134.124032,s,10,23.713447265625,2.3713447265625,0.002944806273489073,2.3702906494140628,2.3727993896484376,2.3762648071289063,2.3790371411132814,"[2.379730224609375, 2.371194580078125, 2.37024951171875, 2.370331787109375, 2.369844482421875, 2.371651123046875, 2.372029296875, 2.369348876953125, 2.368902099609375, 2.370165283203125]",tokens/s,26.567204377460868,kWh,2.7922422098261967e-05,1.530250755295801e-05,0.0001232537192845031,0.0001664786489357231,tokens/kWh,378426.90580894914,,s,629,24.430713893890378,0.038840562629396475,0.010041817916429953,0.03761459350585938,0.03773158416748047,0.037872434997558595,0.12195401611328126,"[0.03847782516479492, 0.03823513412475586, 0.03829452896118164, 0.03828224182128906, 0.03832729721069336, 0.03821670532226563, 0.03828940963745117, 0.03830988693237305, 0.03845939254760742, 0.03806003189086914, 0.03791462326049805, 0.03796582412719727, 0.0377784309387207, 0.037667839050292966, 0.03751424026489258, 0.03754905700683594, 0.037663745880126956, 0.037544960021972655, 0.037544960021972655, 0.03767302322387695, 0.03784799957275391, 0.03764326477050781, 0.03758694458007812, 0.037679168701171876, 0.037698497772216795, 0.03751116943359375, 0.0375654411315918, 0.037550079345703126, 0.03754905700683594, 0.03751116943359375, 0.03759619140625, 0.03773846435546875, 0.037768192291259765, 0.03766886520385742, 0.0376258544921875, 0.037566463470458986, 0.03758796691894531, 0.037610496520996094, 0.03760332870483398, 0.03760639953613281, 0.037742656707763673, 0.03759916687011719, 0.037644287109375, 0.03759001541137695, 0.03763814544677734, 0.03762995147705078, 0.0376258544921875, 0.03771187210083008, 0.03768115234375, 0.03766067123413086, 0.0376627197265625, 0.03761459350585938, 0.037634048461914066, 0.03769753646850586, 0.03771187210083008, 0.03768012619018555, 0.03765760040283203, 0.03768729782104492, 0.037664768218994144, 0.03771596908569336, 0.03762483215332031, 0.037743614196777346, 0.1216153564453125, 0.03750297546386719, 0.03748761749267578, 0.037473281860351565, 0.03747532653808594, 0.03751935958862305, 0.037498878479003905, 0.0375203857421875, 0.037512191772460936, 0.03769145584106445, 0.03829344177246094, 0.037689342498779296, 0.037541889190673826, 0.03753779220581055, 0.03753881454467774, 0.037541889190673826, 0.037591041564941405, 0.03754086303710937, 0.03750707244873047, 0.03762176132202148, 0.037628929138183595, 0.037566463470458986, 0.037531646728515625, 0.037577728271484374, 0.03755724716186523, 0.03759206390380859, 0.037558273315429686, 0.03765248107910156, 0.03758796691894531, 0.0376258544921875, 0.03763814544677734, 0.0375654411315918, 0.03763715362548828, 0.03758486557006836, 0.03764633560180664, 0.03759417724609375, 0.03761043167114258, 0.0376545295715332, 0.037541889190673826, 0.03760332870483398, 0.03759820938110352, 0.037661697387695314, 0.03763711929321289, 0.0376258544921875, 0.037689342498779296, 0.03764326477050781, 0.037705726623535156, 0.03776204681396484, 0.03769241714477539, 0.03767295837402344, 0.037623809814453124, 0.03762790298461914, 0.03763203048706055, 0.037653472900390624, 0.037731391906738285, 0.037655487060546874, 0.03763916778564453, 0.03779891204833984, 0.0380948486328125, 0.03783679962158203, 0.03766988754272461, 0.03772723388671875, 0.03766579055786133, 0.12242124938964843, 0.037556224822998044, 0.0375203857421875, 0.037610496520996094, 0.03753676986694336, 0.037561344146728515, 0.03750809478759766, 0.03747430419921875, 0.03750400161743164, 0.03754905700683594, 0.03754086303710937, 0.03750400161743164, 0.03747436904907227, 0.03753567886352539, 0.037501953125, 0.03750400161743164, 0.03751424026489258, 0.037577728271484374, 0.03751935958862305, 0.03752755355834961, 0.03754393768310547, 0.03754086303710937, 0.037482494354248046, 0.03780915069580078, 0.038012928009033206, 0.037617664337158206, 0.03759001541137695, 0.03757056045532227, 0.037574657440185545, 0.037566463470458986, 0.03753881454467774, 0.03757779312133789, 0.03783980941772461, 0.037749759674072264, 0.037585918426513674, 0.037618686676025394, 0.03755011367797852, 0.03759203338623047, 0.03759718322753906, 0.037647361755371096, 0.037663745880126956, 0.03759513473510742, 0.037618686676025394, 0.03760844802856445, 0.03765555191040039, 0.037651454925537106, 0.037573696136474606, 0.037682113647460935, 0.037610496520996094, 0.03768320083618164, 0.037959678649902344, 0.03785830307006836, 0.037736446380615234, 0.03763097763061524, 0.037663745880126956, 0.03767091369628906, 0.03763916778564453, 0.0377077751159668, 0.03769753646850586, 0.03768320083618164, 0.03763097763061524, 0.03760639953613281, 0.03763814544677734, 0.12196147155761719, 0.03751424026489258, 0.0375470085144043, 0.037556224822998044, 0.03750912094116211, 0.03751628875732422, 0.03754086303710937, 0.03751116943359375, 0.03747225570678711, 0.0375654411315918, 0.03754086303710937, 0.03759718322753906, 0.038035457611083984, 0.03784703826904297, 0.03754086303710937, 0.037566463470458986, 0.037571582794189456, 0.03754291152954101, 0.0375203857421875, 0.037550079345703126, 0.0374917106628418, 0.037556224822998044, 0.0375470085144043, 0.037571582794189456, 0.03761356735229492, 0.03770470428466797, 0.03769343948364258, 0.03760332870483398, 0.03759001541137695, 0.03759308624267578, 0.03759820938110352, 0.03763814544677734, 0.03760537719726562, 0.03761356735229492, 0.03769651031494141, 0.03764019012451172, 0.03760332870483398, 0.03758899307250976, 0.037617664337158206, 0.037722110748291016, 0.03766681671142578, 0.03764326477050781, 0.03763097763061524, 0.037607425689697264, 0.03758796691894531, 0.037615615844726565, 0.03764223861694336, 0.03765043258666992, 0.037615615844726565, 0.037594112396240234, 0.03766579055786133, 0.03765657424926758, 0.03765657424926758, 0.03767705535888672, 0.037647361755371096, 0.03763916778564453, 0.03767603302001953, 0.03763302230834961, 0.03769548797607422, 0.037719039916992186, 0.03764633560180664, 0.037647361755371096, 0.03766681671142578, 0.12246121978759765, 0.03751011276245117, 0.0375203857421875, 0.037526527404785154, 0.03751116943359375, 0.03760025787353516, 0.03747635269165039, 0.03759718322753906, 0.03746099090576172, 0.03758694458007812, 0.037528575897216795, 0.03754393768310547, 0.03752345657348633, 0.03761663818359375, 0.03757670211791992, 0.03761356735229492, 0.03751935958862305, 0.03757875061035156, 0.03760025787353516, 0.037601280212402347, 0.03759718322753906, 0.03760947036743164, 0.037510143280029294, 0.037591041564941405, 0.037539840698242184, 0.03759718322753906, 0.03751424026489258, 0.037591041564941405, 0.03771187210083008, 0.037594112396240234, 0.03757056045532227, 0.03761663818359375, 0.037539840698242184, 0.03763302230834961, 0.03765043258666992, 0.037602302551269534, 0.03766067123413086, 0.03774771118164062, 0.03759513473510742, 0.03758899307250976, 0.037628929138183595, 0.03764223861694336, 0.03761663818359375, 0.03762995147705078, 0.03759926223754883, 0.03762992095947266, 0.03760844802856445, 0.03767398452758789, 0.037591041564941405, 0.037645313262939455, 0.037648384094238284, 0.037623809814453124, 0.03770163345336914, 0.03762483215332031, 0.03771289443969727, 0.03764326477050781, 0.03767295837402344, 0.037819393157958986, 0.03767603302001953, 0.037705726623535156, 0.03772723388671875, 0.03763916778564453, 0.0377262077331543, 0.12193484497070313, 0.037558273315429686, 0.03750912094116211, 0.03746201705932617, 0.03745894241333008, 0.03749990463256836, 0.03751424026489258, 0.03750604629516602, 0.037512191772460936, 0.037525505065917966, 0.037466110229492186, 0.037495807647705076, 0.03748147201538086, 0.037601280212402347, 0.037917697906494144, 0.03800678253173828, 0.03761459350585938, 0.03763711929321289, 0.03754086303710937, 0.03756748962402344, 0.03760639953613281, 0.03761971282958984, 0.03753881454467774, 0.037601280212402347, 0.03766579055786133, 0.03762790298461914, 0.03759619140625, 0.03758179092407227, 0.037664768218994144, 0.03763711929321289, 0.037645313262939455, 0.03766988754272461, 0.03761971282958984, 0.03758899307250976, 0.037564414978027344, 0.037628929138183595, 0.037574657440185545, 0.037602302551269534, 0.037628929138183595, 0.03781017684936523, 0.03762688064575195, 0.03765862274169922, 0.03762790298461914, 0.037720062255859374, 0.03774054336547852, 0.0376360969543457, 0.0376995849609375, 0.03768012619018555, 0.037722110748291016, 0.03762790298461914, 0.03775385665893555, 0.03764019012451172, 0.03765043258666992, 0.03764940643310547, 0.03796582412719727, 0.037718017578125, 0.037667839050292966, 0.037703678131103514, 0.03770163345336914, 0.03770982360839844, 0.0377446403503418, 0.03765248107910156, 0.037733375549316404, 0.12274483489990234, 0.037599231719970705, 0.03751628875732422, 0.03759724807739258, 0.03755513763427734, 0.037561344146728515, 0.03746406555175781, 0.03754393768310547, 0.03744870376586914, 0.037561344146728515, 0.03747020721435547, 0.03762278366088867, 0.03747225570678711, 0.037553150177001955, 0.03752243041992188, 0.03752140808105469, 0.03758489608764649, 0.037576766967773435, 0.037603263854980466, 0.03765248107910156, 0.037553150177001955, 0.037561344146728515, 0.03762483215332031, 0.037660736083984375, 0.03754182434082031, 0.037664768218994144, 0.03762483215332031, 0.037553150177001955, 0.0376627197265625, 0.037921791076660154, 0.03764019012451172, 0.03760844802856445, 0.03755523300170899, 0.03764835357666016, 0.03758899307250976, 0.03757263946533203, 0.0376022720336914, 0.03772313690185547, 0.0376258544921875, 0.03761459350585938, 0.03763411331176758, 0.03766368103027344, 0.03786137771606445, 0.037667839050292966, 0.03768012619018555, 0.03765964889526367, 0.037689342498779296, 0.037718017578125, 0.03774054336547852, 0.037664768218994144, 0.037634048461914066, 0.0376360969543457, 0.03769036865234375, 0.03767398452758789, 0.037744705200195315, 0.038224830627441406, 0.03787980651855469, 0.03767910385131836, 0.0377149429321289, 0.037700607299804685, 0.03769343948364258, 0.03766681671142578, 0.03767500686645508, 0.12211814117431641, 0.03755110549926758, 0.03753577423095703, 0.03757257461547851, 0.037594112396240234, 0.03760844802856445, 0.037528575897216795, 0.03753062438964844, 0.037495807647705076, 0.03758796691894531, 0.03751833724975586, 0.03754905700683594, 0.03750809478759766, 0.03752959823608398, 0.03755417633056641, 0.03748556900024414, 0.03765862274169922, 0.037553150177001955, 0.037553150177001955, 0.037572608947753904, 0.037531646728515625, 0.03755724716186523, 0.03759820938110352, 0.03762483215332031, 0.03753676986694336, 0.03760537719726562, 0.037561344146728515, 0.03762176132202148, 0.037623809814453124, 0.0376995849609375, 0.03776409530639648, 0.0375654411315918, 0.03754598236083984, 0.03757363128662109, 0.03760435104370117, 0.037661697387695314, 0.037572608947753904, 0.03760537719726562, 0.03756748962402344, 0.03760025787353516, 0.03760537719726562, 0.037563392639160156, 0.03765555191040039, 0.03760844802856445, 0.03763814544677734, 0.037618686676025394, 0.037609535217285155, 0.037635009765625, 0.037607425689697264, 0.03760851287841797, 0.037649345397949216, 0.03765248107910156, 0.03762278366088867, 0.03760639953613281, 0.037648384094238284, 0.0376627197265625, 0.03769343948364258, 0.03770265579223633, 0.03768115234375, 0.03766483306884766, 0.03765446472167969, 0.03766988754272461, 0.03768320083618164, 0.12238745880126953, 0.037556224822998044, 0.03748761749267578, 0.03758796691894531, 0.037477375030517575, 0.037550079345703126, 0.03747123336791992, 0.03754291152954101, 0.037507102966308596, 0.037538784027099606, 0.037454849243164064, 0.03752345657348633, 0.03746406555175781, 0.03752447891235351, 0.03748966217041016, 0.03753062438964844, 0.03748659133911133, 0.03757567977905273, 0.037746688842773435, 0.037564414978027344, 0.037531646728515625, 0.03758694458007812, 0.037571582794189456, 0.0377077751159668, 0.0375623664855957, 0.03760537719726562, 0.037607425689697264, 0.037579776763916016, 0.03759308624267578, 0.037528575897216795, 0.03754598236083984, 0.037591041564941405, 0.037574657440185545, 0.037579776763916016, 0.037572608947753904, 0.03760537719726562, 0.037582847595214845, 0.037596160888671876, 0.037564414978027344, 0.03760025787353516, 0.037580799102783204, 0.03765350341796875, 0.03773952102661133, 0.03769651031494141, 0.03764633560180664, 0.03767705535888672, 0.03765964889526367, 0.03761459350585938, 0.037631999969482424, 0.03761971282958984, 0.03764940643310547, 0.03767193603515625, 0.03764223861694336, 0.037631999969482424, 0.03763711929321289, 0.03764326477050781, 0.037663745880126956, 0.0376545295715332, 0.037663745880126956, 0.037661697387695314, 0.03766988754272461, 0.03761459350585938, 0.03769651031494141, 0.1219727325439453, 0.037515262603759765, 0.03747225570678711, 0.03765862274169922, 0.037792766571044925, 0.03757567977905273, 0.037700607299804685, 0.03754598236083984, 0.037495807647705076, 0.03750809478759766, 0.037512191772460936, 0.03757056045532227, 0.03750502395629883, 0.03753779220581055, 0.03750604629516602, 0.03752959823608398, 0.03748044967651367, 0.03758694458007812, 0.03754291152954101, 0.03763916778564453, 0.03782963180541992, 0.037572608947753904, 0.037558273315429686, 0.03758796691894531, 0.03754393768310547, 0.03762073516845703, 0.03755724716186523, 0.03760639953613281, 0.037615615844726565, 0.03758185577392578, 0.037614559173583986, 0.03758899307250976, 0.03765657424926758, 0.037645313262939455, 0.037615615844726565, 0.03760332870483398, 0.037580799102783204, 0.037602302551269534, 0.037594112396240234, 0.03763302230834961, 0.03762483215332031, 0.03761356735229492, 0.037684223175048825, 0.03760537719726562, 0.037613601684570314, 0.03761865615844726, 0.03768012619018555, 0.03762483215332031, 0.037689342498779296, 0.037678081512451174, 0.03765760040283203, 0.03765760040283203, 0.03765350341796875, 0.037645313262939455, 0.03767500686645508, 0.03766886520385742, 0.03765760040283203, 0.03766579055786133, 0.03768832015991211, 0.037678081512451174, 0.03773235321044922, 0.03767603302001953, 0.037698558807373043]",tokens/s,25.7462799790431,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,google/recurrentgemma-7b,google/recurrentgemma-7b,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/google/recurrentgemma-7b/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6694702e-45e94dfb0d2424ba19619b8f;6d73ef94-1ef0-486a-ba48-b597765df1b9) Repository Not Found for url: https://huggingface.co/google/recurrentgemma-7b/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: google/recurrentgemma-7b is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,mistralai/Mixtral-8x7B-v0.1,mistralai/Mixtral-8x7B-v0.1,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ self.load_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 250, in load_model_with_no_weights self.pretrained_model = self.automodel_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3710, in from_pretrained model = cls(config, *model_args, **model_kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/mixtral/modeling_mixtral.py"", line 1323, in __init__ self.model = MixtralModel(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/mixtral/modeling_mixtral.py"", line 1087, in __init__ [MixtralDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/mixtral/modeling_mixtral.py"", line 1087, in [MixtralDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/mixtral/modeling_mixtral.py"", line 877, in __init__ self.block_sparse_moe = MixtralSparseMoeBlock(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/mixtral/modeling_mixtral.py"", line 824, in __init__ self.experts = nn.ModuleList([MixtralBlockSparseTop2MLP(config) for _ in range(self.num_experts)]) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/mixtral/modeling_mixtral.py"", line 824, in self.experts = nn.ModuleList([MixtralBlockSparseTop2MLP(config) for _ in range(self.num_experts)]) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/mixtral/modeling_mixtral.py"", line 791, in __init__ self.w2 = nn.Linear(self.ffn_dim, self.hidden_dim, bias=False) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/linear.py"", line 98, in __init__ self.weight = Parameter(torch.empty((out_features, in_features), **factory_kwargs)) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_device.py"", line 78, in __torch_function__ return func(*args, **kwargs) torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 112.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,t,t,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/t/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6694738f-1ba9805e2122e4be3754abe1;36f6c1f2-5559-41d0-8e4d-3f0d28d3b50a) Repository Not Found for url: https://huggingface.co/t/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: t is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-72B,Qwen/Qwen-72B,cuda,0,42,,,True,,,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",qwen,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-30b,facebook/opt-30b,cuda,0,42,,,True,,,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",opt,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-MoE-A2.7B,Qwen/Qwen1.5-MoE-A2.7B,cuda,0,42,,,True,,,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",qwen2_moe,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-1.3b,EleutherAI/pythia-1.3b,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1308.086272,3848.798208,0.0,3202.351104,3105.82784,s,10,0.2721215019226074,0.027212150192260743,0.0005506786455814327,0.02701697540283203,0.027374995040893555,0.02811365737915039,0.02870458724975586,"[0.028852319717407225, 0.027018911361694337, 0.027015039443969727, 0.02704364776611328, 0.02697238349914551, 0.027210847854614258, 0.026960575103759765, 0.027014848709106445, 0.027043231964111326, 0.026989696502685546]",tokens/s,9407.562364285624,kWh,3.2189005781151185e-07,1.7636263548326822e-07,1.3816753408828266e-06,1.8799280341776065e-06,tokens/kWh,136175425.51940814,MB,1308.086272,3848.798208,0.0,3202.351104,3188.152832,s,10,12.426950073242189,1.242695007324219,0.008562259006632902,1.2417922973632813,1.2499968627929687,1.2562812072753906,1.261308682861328,"[1.2625655517578125, 1.2339195556640625, 1.240325927734375, 1.2399776611328126, 1.2432586669921875, 1.2456578369140625, 1.248600341796875, 1.2466455078125, 1.2324085693359375, 1.2335904541015625]",tokens/s,50.69626869721809,kWh,1.458196097775694e-05,7.990656032580708e-06,3.975949646091645e-05,6.233211347125409e-05,tokens/kWh,1010714.9668373417,,s,629,12.648431613922098,0.020108794298763305,0.003280659308354322,0.01960038375854492,0.02013347854614258,0.02032803840637207,0.04704813217163086,"[0.0214783992767334, 0.021115903854370118, 0.020468736648559572, 0.020332544326782227, 0.020254720687866212, 0.020347904205322266, 0.020740095138549804, 0.020281343460083007, 0.020090879440307616, 0.020147199630737304, 0.020297727584838866, 0.020521984100341797, 0.020248575210571287, 0.020165632247924805, 0.020307968139648438, 0.020371456146240235, 0.020257791519165038, 0.020256767272949217, 0.02067148780822754, 0.020296703338623046, 0.020556800842285155, 0.02008883285522461, 0.020115455627441405, 0.02023219108581543, 0.020761600494384767, 0.020948991775512696, 0.020332544326782227, 0.020207616806030275, 0.020082687377929686, 0.019530752182006835, 0.020118528366088868, 0.020188159942626953, 0.019778560638427735, 0.019520511627197267, 0.019552255630493166, 0.019744768142700195, 0.01953388786315918, 0.019320768356323244, 0.01942118453979492, 0.019566591262817384, 0.019548160552978516, 0.020168703079223634, 0.02017791938781738, 0.01982464027404785, 0.019555328369140625, 0.019459104537963866, 0.019784671783447266, 0.019486719131469727, 0.019564544677734375, 0.019489791870117186, 0.019513343811035155, 0.020002815246582033, 0.019563520431518554, 0.019555360794067382, 0.019584991455078124, 0.019551231384277345, 0.019553279876708983, 0.019561471939086913, 0.019582975387573243, 0.019479551315307618, 0.019614719390869142, 0.019607551574707033, 0.04705587387084961, 0.019513343811035155, 0.019746816635131836, 0.01989836883544922, 0.019544063568115236, 0.019530752182006835, 0.019528703689575197, 0.019704832077026366, 0.019781631469726564, 0.019590143203735352, 0.01949286460876465, 0.019535871505737306, 0.019504127502441407, 0.019942399978637695, 0.02003660774230957, 0.01942937660217285, 0.019337215423583985, 0.019526655197143555, 0.019591167449951173, 0.019478527069091797, 0.019520511627197267, 0.019573759078979493, 0.019545087814331053, 0.019629056930541993, 0.019511295318603517, 0.019347455978393553, 0.019390464782714844, 0.019573759078979493, 0.019532800674438477, 0.019591167449951173, 0.019467264175415038, 0.01963212776184082, 0.019481599807739256, 0.019610624313354492, 0.019531776428222656, 0.019518463134765626, 0.019507200241088866, 0.019497983932495116, 0.020135936737060548, 0.019662912368774415, 0.019652544021606447, 0.01960038375854492, 0.019510271072387696, 0.019577856063842772, 0.019582975387573243, 0.019571712493896484, 0.019501056671142578, 0.01966182327270508, 0.019482624053955077, 0.01964339256286621, 0.01950931167602539, 0.019533760070800783, 0.019546112060546874, 0.01946009635925293, 0.0196177921295166, 0.01960140800476074, 0.019551231384277345, 0.019531776428222656, 0.019500032424926757, 0.019696640014648437, 0.019522560119628905, 0.019571712493896484, 0.019483648300170898, 0.04720435333251953, 0.019708927154541016, 0.019531776428222656, 0.019674144744873046, 0.01991881561279297, 0.019949567794799804, 0.019680255889892577, 0.01961881637573242, 0.019574783325195313, 0.019575807571411134, 0.019936256408691407, 0.019579904556274414, 0.019325952529907226, 0.01958502388000488, 0.019467264175415038, 0.019826688766479493, 0.01947750473022461, 0.019554304122924804, 0.01972844886779785, 0.019658687591552735, 0.0196495361328125, 0.019556352615356445, 0.019562496185302734, 0.0196495361328125, 0.019896320343017578, 0.01962495994567871, 0.019536895751953123, 0.019524608612060547, 0.019512351989746095, 0.019564512252807618, 0.019567615509033204, 0.0196495361328125, 0.01961676788330078, 0.01964339256286621, 0.019512319564819337, 0.01965465545654297, 0.019556352615356445, 0.01964646339416504, 0.019570688247680663, 0.01967411231994629, 0.019721216201782226, 0.020162559509277343, 0.021012479782104493, 0.020463615417480468, 0.01987379264831543, 0.019793920516967774, 0.019688447952270507, 0.019631103515625, 0.019312639236450196, 0.01960038375854492, 0.019574783325195313, 0.01959321594238281, 0.019612672805786133, 0.02006425666809082, 0.019579904556274414, 0.019673088073730468, 0.01968435287475586, 0.01980620765686035, 0.01949286460876465, 0.019517440795898438, 0.01959321594238281, 0.01968639945983887, 0.019525632858276368, 0.047126529693603515, 0.019548160552978516, 0.019597312927246095, 0.01957072067260742, 0.01950716781616211, 0.019537919998168944, 0.01959321594238281, 0.01959017562866211, 0.019514335632324218, 0.019490816116333007, 0.019470367431640625, 0.019619808197021485, 0.020369407653808593, 0.019561471939086913, 0.019414016723632813, 0.01965977668762207, 0.019550207138061524, 0.019851264953613282, 0.019603456497192383, 0.01944780731201172, 0.01958502388000488, 0.01949286460876465, 0.019631103515625, 0.019580928802490235, 0.019559423446655275, 0.019532800674438477, 0.019587072372436523, 0.019602432250976562, 0.019506175994873046, 0.019540992736816407, 0.019739648818969727, 0.020460544586181642, 0.020523008346557618, 0.020206592559814454, 0.02006937599182129, 0.019494911193847657, 0.019673088073730468, 0.019796991348266603, 0.019545087814331053, 0.019687423706054686, 0.01957683181762695, 0.01946316719055176, 0.020048896789550782, 0.019998720169067383, 0.019566591262817384, 0.019553279876708983, 0.019595264434814453, 0.019981311798095702, 0.019703807830810546, 0.019510271072387696, 0.019519487380981446, 0.019604480743408204, 0.020039680480957032, 0.01948569679260254, 0.0196177921295166, 0.019603456497192383, 0.01960038375854492, 0.019706880569458008, 0.01963417625427246, 0.019569664001464843, 0.01965772819519043, 0.019547136306762695, 0.02001203155517578, 0.04705791854858398, 0.019501056671142578, 0.019579904556274414, 0.019899391174316407, 0.02085273551940918, 0.019495935440063478, 0.019583999633789064, 0.019533824920654298, 0.019936256408691407, 0.019564544677734375, 0.019487808227539063, 0.01955833625793457, 0.019543039321899415, 0.019687423706054686, 0.019579904556274414, 0.019589120864868165, 0.01966592025756836, 0.0195020809173584, 0.019769344329833984, 0.019532800674438477, 0.019834911346435547, 0.01963209533691406, 0.019553279876708983, 0.019591167449951173, 0.019927040100097656, 0.019802112579345704, 0.019354623794555666, 0.019812351226806642, 0.020262975692749024, 0.020265920639038086, 0.019543039321899415, 0.01965056037902832, 0.019544095993041993, 0.020062175750732422, 0.02002022361755371, 0.019474431991577147, 0.019793920516967774, 0.02002739143371582, 0.02007859230041504, 0.019887104034423828, 0.019458047866821288, 0.019372032165527343, 0.01948467254638672, 0.019753984451293945, 0.019979263305664064, 0.019772415161132813, 0.019478527069091797, 0.01990656089782715, 0.019960832595825196, 0.01984102439880371, 0.020082687377929686, 0.01945907211303711, 0.0196177921295166, 0.019561471939086913, 0.019969024658203126, 0.019900480270385743, 0.01972115135192871, 0.01957683181762695, 0.020057088851928712, 0.019987455368041994, 0.019350528717041016, 0.01943756866455078, 0.019497983932495116, 0.04791398239135742, 0.019306495666503908, 0.019522560119628905, 0.019486783981323242, 0.01995871925354004, 0.020290559768676757, 0.019602432250976562, 0.019720191955566405, 0.019519487380981446, 0.019732479095458985, 0.02008576011657715, 0.01984716796875, 0.019508224487304687, 0.01968332862854004, 0.019586048126220702, 0.019945472717285157, 0.019948575973510744, 0.019479520797729494, 0.01984409523010254, 0.019537919998168944, 0.020123647689819335, 0.01984716796875, 0.019532800674438477, 0.019693567276000978, 0.01948876762390137, 0.019705856323242187, 0.01940787124633789, 0.01933523178100586, 0.019267520904541015, 0.019405824661254883, 0.019562496185302734, 0.019513343811035155, 0.01984819221496582, 0.01963417625427246, 0.02021683120727539, 0.0195020809173584, 0.019573759078979493, 0.019467264175415038, 0.019751935958862304, 0.019628032684326172, 0.01999564743041992, 0.02027008056640625, 0.0198154239654541, 0.019915775299072267, 0.019567615509033204, 0.019978239059448243, 0.02005401611328125, 0.019697664260864257, 0.019737600326538086, 0.02006118392944336, 0.020262912750244142, 0.020132863998413086, 0.01986764717102051, 0.01957683181762695, 0.020105215072631837, 0.020214784622192384, 0.020076543807983398, 0.020174848556518556, 0.019588096618652344, 0.01966182327270508, 0.019596288681030274, 0.019786752700805665, 0.019753984451293945, 0.04746649551391602, 0.020074495315551756, 0.02006630325317383, 0.01997209548950195, 0.01970278358459473, 0.019775487899780272, 0.019787776947021486, 0.020006912231445313, 0.019499008178710937, 0.02007142448425293, 0.02012057685852051, 0.019869695663452147, 0.01969152069091797, 0.019536895751953123, 0.019599359512329103, 0.019495935440063478, 0.019990528106689453, 0.019756032943725587, 0.02000486373901367, 0.019885055541992186, 0.020008960723876954, 0.02048204803466797, 0.020092927932739257, 0.01979084777832031, 0.01943756866455078, 0.019885055541992186, 0.019917823791503905, 0.01946419143676758, 0.01971609687805176, 0.01966182327270508, 0.020093952178955078, 0.01986867141723633, 0.019903488159179687, 0.01948057556152344, 0.019508224487304687, 0.01964134407043457, 0.019869695663452147, 0.02001408004760742, 0.01947750473022461, 0.019923967361450197, 0.020040704727172853, 0.02007142448425293, 0.02009702491760254, 0.019556352615356445, 0.019701759338378907, 0.019831840515136718, 0.020070367813110352, 0.019932159423828123, 0.019677183151245118, 0.0194334716796875, 0.019739648818969727, 0.02003558349609375, 0.019894271850585937, 0.019978239059448243, 0.019526655197143555, 0.019579904556274414, 0.01940377616882324, 0.01946316719055176, 0.019610624313354492, 0.019721216201782226, 0.019676191329956055, 0.019994592666625975, 0.019943424224853516, 0.04702822494506836, 0.019796991348266603, 0.019961856842041017, 0.020136959075927736, 0.02007244873046875, 0.019705856323242187, 0.019548160552978516, 0.020307968139648438, 0.019979263305664064, 0.01965875244140625, 0.01986867141723633, 0.019701759338378907, 0.020057151794433594, 0.01975904083251953, 0.020101119995117187, 0.020321279525756835, 0.02008064079284668, 0.020084768295288085, 0.0200980167388916, 0.020139007568359374, 0.01984819221496582, 0.01945088005065918, 0.01942527961730957, 0.01963932800292969, 0.01988400077819824, 0.021565439224243164, 0.020529151916503906, 0.019919872283935547, 0.019610624313354492, 0.019523584365844726, 0.019543039321899415, 0.01961881637573242, 0.01945497512817383, 0.020152320861816408, 0.019743743896484374, 0.019430400848388672, 0.019594240188598632, 0.019571712493896484, 0.019615743637084963, 0.019588096618652344, 0.019580928802490235, 0.019532800674438477, 0.019619903564453124, 0.019587007522583008, 0.01967411231994629, 0.019563520431518554, 0.01967513656616211, 0.019580928802490235, 0.019591167449951173, 0.019664896011352538, 0.019513343811035155, 0.019558399200439454, 0.019499008178710937, 0.019779584884643556, 0.019554304122924804, 0.019771392822265626, 0.019704832077026366, 0.020051967620849608, 0.019916799545288084, 0.019560447692871095, 0.01957379150390625, 0.01950716781616211, 0.019520511627197267, 0.04714393615722656, 0.019533824920654298, 0.01955027198791504, 0.01954400062561035, 0.01960140800476074, 0.01944576072692871, 0.019504127502441407, 0.019555328369140625, 0.01958502388000488, 0.019537919998168944, 0.019573759078979493, 0.01966592025756836, 0.0194467830657959, 0.019559423446655275, 0.019527679443359376, 0.019587072372436523, 0.01967616081237793, 0.019588096618652344, 0.019497983932495116, 0.01963724708557129, 0.01984511947631836, 0.019544063568115236, 0.019521535873413084, 0.01960960006713867, 0.019751935958862304, 0.0196628475189209, 0.019539968490600586, 0.01947238349914551, 0.019570688247680663, 0.019517440795898438, 0.019538944244384765, 0.019566591262817384, 0.019479551315307618, 0.01943961524963379, 0.019491840362548828, 0.01957683181762695, 0.01946112060546875, 0.019505151748657225, 0.0194969596862793, 0.019350528717041016, 0.01939455986022949, 0.01940377616882324, 0.019507200241088866, 0.019475456237792968, 0.019910655975341796, 0.019482624053955077, 0.019573759078979493, 0.019516416549682617, 0.019495935440063478, 0.019611648559570313, 0.019534847259521485, 0.019558399200439454, 0.019574783325195313, 0.019513343811035155, 0.01945702362060547, 0.019781631469726564, 0.019611648559570313, 0.019568639755249022, 0.019534847259521485, 0.01946828842163086, 0.019672063827514647, 0.019598335266113282, 0.019536895751953123, 0.047026176452636716, 0.01962188720703125, 0.019534847259521485, 0.019543039321899415, 0.019545087814331053, 0.019581951141357423, 0.019606592178344727, 0.019630016326904295, 0.019583999633789064, 0.01945292854309082, 0.01960550308227539, 0.01944371223449707, 0.019479551315307618, 0.019487743377685548, 0.01960857582092285, 0.019479551315307618, 0.019523584365844726, 0.019555328369140625, 0.019311616897583008, 0.019518463134765626, 0.01929113578796387, 0.01948876762390137, 0.019575807571411134, 0.019664896011352538, 0.019534847259521485, 0.019474431991577147, 0.019438592910766602, 0.01958502388000488, 0.019550207138061524, 0.01960044860839844, 0.01957267189025879, 0.019570688247680663, 0.02003046417236328, 0.01971609687805176, 0.0195645751953125, 0.019586015701293945, 0.019544095993041993, 0.019584991455078124, 0.019524608612060547, 0.01963827133178711, 0.0194652156829834, 0.01962188720703125, 0.019876863479614256, 0.019526655197143555, 0.019575807571411134, 0.019375104904174805, 0.019583999633789064, 0.01969152069091797, 0.019614719390869142, 0.01960038375854492, 0.019529727935791014, 0.019521535873413084, 0.019422208786010742, 0.01945088005065918, 0.019358720779418945, 0.019322879791259767, 0.020797439575195312, 0.01995676803588867, 0.01929417610168457, 0.019313663482666017, 0.019559423446655275, 0.019591167449951173, 0.01986662483215332]",tokens/s,49.729485773371316,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-1.4b,EleutherAI/pythia-1.4b,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1283.33824,3848.798208,0.0,3202.351104,3105.82784,s,10,0.2737464962005615,0.027374649620056157,0.0006798933589168986,0.027108223915100097,0.02760761890411377,0.028481137561798094,0.029179952487945558,"[0.02935465621948242, 0.02739606475830078, 0.027413503646850586, 0.027041791915893554, 0.027088895797729492, 0.026967872619628908, 0.027058176040649414, 0.027127552032470702, 0.026950016021728515, 0.027347967147827147]",tokens/s,9351.717868653212,kWh,3.225466720883932e-07,1.767401729368767e-07,1.3880180646975452e-06,1.8873049097228154e-06,tokens/kWh,135643159.02595633,MB,1283.784704,3848.798208,0.0,3202.351104,3188.152832,s,10,12.537179443359374,1.2537179443359376,0.028535391056794507,1.2414683227539063,1.3036857299804687,1.3105398254394531,1.3160231018066406,"[1.3173939208984375, 1.30216259765625, 1.2359407958984374, 1.2357900390625, 1.2407454833984375, 1.242191162109375, 1.2385311279296876, 1.2330062255859375, 1.2484185791015625, 1.24299951171875]",tokens/s,50.250537040346416,kWh,1.468646836486543e-05,8.047551752407129e-06,3.957804248590371e-05,6.231206260317627e-05,tokens/kWh,1011040.1962009304,,s,629,12.760228870391852,0.020286532385360646,0.0033330012219880433,0.01966796875,0.02081628189086914,0.020942233657836914,0.04715091812133789,"[0.020752384185791017, 0.02205388832092285, 0.021751808166503905, 0.021104639053344726, 0.02088140869140625, 0.020815872192382814, 0.021105663299560547, 0.02084147262573242, 0.020876287460327148, 0.020999168395996092, 0.020888576507568358, 0.02066022491455078, 0.0215285758972168, 0.02082918357849121, 0.020912128448486327, 0.020779008865356444, 0.02101759910583496, 0.02084556770324707, 0.020922367095947265, 0.020817920684814452, 0.020831232070922853, 0.020884479522705078, 0.020876287460327148, 0.02083635139465332, 0.020773887634277344, 0.02068992042541504, 0.02126335906982422, 0.02141900825500488, 0.021533695220947266, 0.020939775466918945, 0.021082111358642578, 0.02090598487854004, 0.020802560806274413, 0.020731903076171874, 0.020817983627319337, 0.0207523193359375, 0.020578304290771485, 0.02048409652709961, 0.020519935607910156, 0.02046463966369629, 0.020503583908081054, 0.020443103790283204, 0.020504575729370117, 0.020761600494384767, 0.021131263732910157, 0.021990400314331054, 0.02130636787414551, 0.020955135345458984, 0.02090291213989258, 0.020781055450439453, 0.020626432418823244, 0.020625408172607423, 0.02084147262573242, 0.02083020782470703, 0.02079539108276367, 0.020908031463623047, 0.020839424133300782, 0.020782079696655274, 0.020839424133300782, 0.020747264862060546, 0.020585472106933594, 0.020488191604614257, 0.04849971389770508, 0.02047488021850586, 0.020789247512817383, 0.020855808258056642, 0.02059775924682617, 0.020713504791259767, 0.020808671951293944, 0.020616191864013672, 0.020479999542236327, 0.02043801689147949, 0.02045952033996582, 0.020447231292724608, 0.02040729522705078, 0.020478975296020507, 0.02042470359802246, 0.020477951049804686, 0.020493375778198243, 0.020526016235351562, 0.020404224395751954, 0.020701183319091796, 0.020930559158325195, 0.020867071151733398, 0.020847616195678712, 0.02082713508605957, 0.020824064254760744, 0.02077903938293457, 0.020808671951293944, 0.02087120056152344, 0.02074415969848633, 0.020904960632324218, 0.020616191864013672, 0.020970495223999023, 0.021130239486694336, 0.022185983657836913, 0.021149696350097655, 0.020943872451782225, 0.020711423873901368, 0.020899839401245117, 0.020682752609252928, 0.020558847427368163, 0.02069196891784668, 0.020488191604614257, 0.02049228858947754, 0.02042572784423828, 0.020428800582885744, 0.020504575729370117, 0.020473920822143554, 0.020490175247192384, 0.02045337677001953, 0.020406272888183592, 0.020462591171264647, 0.020576255798339844, 0.021213184356689452, 0.020418560028076172, 0.0204769287109375, 0.020534271240234374, 0.02045747184753418, 0.020449312210083007, 0.02049020767211914, 0.020470783233642577, 0.02051584053039551, 0.020504575729370117, 0.020794368743896483, 0.04849868774414062, 0.020488256454467772, 0.020458463668823243, 0.020754400253295897, 0.019559423446655275, 0.01964236831665039, 0.019742719650268553, 0.019551231384277345, 0.01947750473022461, 0.019491840362548828, 0.01946931266784668, 0.01944063949584961, 0.01965465545654297, 0.019701759338378907, 0.01965465545654297, 0.019706880569458008, 0.019604480743408204, 0.019567615509033204, 0.019451904296875, 0.019538944244384765, 0.01968639945983887, 0.01966796875, 0.019517440795898438, 0.01960860824584961, 0.019422176361083985, 0.019505151748657225, 0.01945497512817383, 0.019438592910766602, 0.019536895751953123, 0.019517440795898438, 0.01945907211303711, 0.019554304122924804, 0.019515392303466796, 0.019530752182006835, 0.019533824920654298, 0.019494911193847657, 0.01944883155822754, 0.01949286460876465, 0.019449855804443358, 0.01965363121032715, 0.019553279876708983, 0.019501056671142578, 0.01943654441833496, 0.01953286361694336, 0.019500991821289063, 0.019475456237792968, 0.019497983932495116, 0.019574783325195313, 0.019405824661254883, 0.019474431991577147, 0.01945395278930664, 0.019558399200439454, 0.01965260887145996, 0.0196495361328125, 0.01962598419189453, 0.01983078384399414, 0.019604480743408204, 0.019545087814331053, 0.019599359512329103, 0.019589120864868165, 0.01946316719055176, 0.01952672004699707, 0.019445695877075196, 0.04718694305419922, 0.01945395278930664, 0.01942527961730957, 0.019422208786010742, 0.01946419143676758, 0.019504127502441407, 0.01946828842163086, 0.019526655197143555, 0.019430400848388672, 0.019450944900512697, 0.019450815200805664, 0.0194467830657959, 0.019529727935791014, 0.019574783325195313, 0.019586048126220702, 0.019567615509033204, 0.019538944244384765, 0.01963417625427246, 0.01960550308227539, 0.019566591262817384, 0.019533824920654298, 0.019602432250976562, 0.019544063568115236, 0.019714048385620117, 0.019720191955566405, 0.01967820739746094, 0.019733503341674806, 0.01963520050048828, 0.019768320083618163, 0.01968435287475586, 0.019483648300170898, 0.019528703689575197, 0.019909631729125975, 0.019811328887939454, 0.019688447952270507, 0.019710975646972655, 0.01967513656616211, 0.01964851188659668, 0.01945702362060547, 0.0196945915222168, 0.019695615768432616, 0.019787776947021486, 0.019969024658203126, 0.019981311798095702, 0.019934207916259765, 0.019843072891235353, 0.019647487640380858, 0.019550207138061524, 0.019612672805786133, 0.01961676788330078, 0.019655679702758787, 0.019732479095458985, 0.01966592025756836, 0.019636224746704102, 0.0196628475189209, 0.019529727935791014, 0.01947750473022461, 0.019555328369140625, 0.01944063949584961, 0.01943654441833496, 0.019497983932495116, 0.019384319305419923, 0.019497983932495116, 0.04707123184204102, 0.019400703430175782, 0.019500032424926757, 0.019604480743408204, 0.019777536392211914, 0.01968230438232422, 0.02022604751586914, 0.0203284797668457, 0.02008265686035156, 0.019982336044311523, 0.02007756805419922, 0.020009983062744142, 0.020129791259765627, 0.02000588798522949, 0.020110368728637695, 0.02028848075866699, 0.02028646469116211, 0.020344831466674804, 0.02036332893371582, 0.020086719512939454, 0.019926015853881835, 0.019555328369140625, 0.019610624313354492, 0.01980521583557129, 0.01953481674194336, 0.0194334716796875, 0.019401727676391603, 0.01967001533508301, 0.019693567276000978, 0.0196997127532959, 0.019679231643676756, 0.019777536392211914, 0.01957683181762695, 0.01942630386352539, 0.01943449592590332, 0.019518463134765626, 0.019487743377685548, 0.019489791870117186, 0.019532800674438477, 0.019737600326538086, 0.019737600326538086, 0.01967513656616211, 0.01944371223449707, 0.0195020809173584, 0.01945599937438965, 0.019483648300170898, 0.019441728591918946, 0.01955731201171875, 0.01948569679260254, 0.01942937660217285, 0.019521535873413084, 0.019481599807739256, 0.01949388885498047, 0.019466239929199217, 0.019471359252929688, 0.019553279876708983, 0.019528703689575197, 0.01943552017211914, 0.01949286460876465, 0.01942630386352539, 0.019571712493896484, 0.019414016723632813, 0.019443775177001955, 0.04715206527709961, 0.019668991088867188, 0.01965056037902832, 0.019501056671142578, 0.019495935440063478, 0.019539968490600586, 0.019520511627197267, 0.019569664001464843, 0.019499008178710937, 0.019777536392211914, 0.019778560638427735, 0.01960960006713867, 0.019680255889892577, 0.01966182327270508, 0.01965670394897461, 0.019793920516967774, 0.019670047760009766, 0.01966998481750488, 0.019680255889892577, 0.01967001533508301, 0.019736576080322265, 0.01969152069091797, 0.01967411231994629, 0.019770368576049805, 0.019693567276000978, 0.019664896011352538, 0.019728384017944335, 0.019677183151245118, 0.019722240447998047, 0.019548160552978516, 0.01960140800476074, 0.019787776947021486, 0.019668991088867188, 0.01967616081237793, 0.01964646339416504, 0.019688447952270507, 0.020065280914306642, 0.019489791870117186, 0.019467264175415038, 0.019552255630493166, 0.01947648048400879, 0.019486719131469727, 0.019481599807739256, 0.019504127502441407, 0.019474431991577147, 0.019408895492553712, 0.0195020809173584, 0.01959321594238281, 0.01967417526245117, 0.019500991821289063, 0.019548160552978516, 0.019575807571411134, 0.01943449592590332, 0.019531776428222656, 0.019563520431518554, 0.019560447692871095, 0.01966182327270508, 0.020564992904663085, 0.02141900825500488, 0.02045644760131836, 0.020355104446411133, 0.020384735107421875, 0.02023219108581543, 0.04725350570678711, 0.019685375213623048, 0.01947340774536133, 0.019580928802490235, 0.01969152069091797, 0.019482624053955077, 0.019521600723266603, 0.019620800018310548, 0.01948057556152344, 0.01944780731201172, 0.019596288681030274, 0.019889184951782227, 0.01975600051879883, 0.019781631469726564, 0.019741695404052736, 0.019673088073730468, 0.019537984848022463, 0.019412927627563477, 0.01944063949584961, 0.02012057685852051, 0.020361215591430663, 0.020368383407592772, 0.020388864517211915, 0.02042572784423828, 0.020107263565063475, 0.01997209548950195, 0.019771392822265626, 0.01947648048400879, 0.01948467254638672, 0.0195850887298584, 0.019675071716308595, 0.019555328369140625, 0.01947340774536133, 0.01947648048400879, 0.019509248733520508, 0.01961884880065918, 0.019649503707885742, 0.019587072372436523, 0.01963315200805664, 0.019745792388916016, 0.019568639755249022, 0.019499008178710937, 0.01944371223449707, 0.01941913604736328, 0.019390464782714844, 0.019449855804443358, 0.019570688247680663, 0.019672063827514647, 0.01944576072692871, 0.019438592910766602, 0.019435552597045897, 0.019430368423461915, 0.01943961524963379, 0.01945497512817383, 0.01942630386352539, 0.0195020809173584, 0.0194969596862793, 0.01963212776184082, 0.019677248001098633, 0.019666879653930665, 0.01963417625427246, 0.01959321594238281, 0.020166656494140626, 0.04744192123413086, 0.019589120864868165, 0.019693567276000978, 0.01966796875, 0.019631103515625, 0.019614784240722657, 0.019687360763549804, 0.019723264694213868, 0.0196495361328125, 0.019731456756591798, 0.019483648300170898, 0.01944063949584961, 0.01944780731201172, 0.01945292854309082, 0.0194652156829834, 0.01949286460876465, 0.01941094398498535, 0.019555328369140625, 0.01966592025756836, 0.019629056930541993, 0.019606527328491212, 0.019615743637084963, 0.019681312561035155, 0.019786720275878907, 0.019518463134765626, 0.01947238349914551, 0.019386367797851564, 0.01943756866455078, 0.01957891273498535, 0.01964134407043457, 0.01968636894226074, 0.019731456756591798, 0.01960857582092285, 0.01946316719055176, 0.019636224746704102, 0.01960550308227539, 0.01967616081237793, 0.01945702362060547, 0.01946009635925293, 0.01965465545654297, 0.01965260887145996, 0.019599359512329103, 0.019449855804443358, 0.019458047866821288, 0.019520511627197267, 0.0194150390625, 0.019491840362548828, 0.019504127502441407, 0.019414016723632813, 0.01945599937438965, 0.0194334716796875, 0.0194703369140625, 0.019489791870117186, 0.019528703689575197, 0.01965056037902832, 0.019746816635131836, 0.01962393569946289, 0.01965772819519043, 0.01966694450378418, 0.01945702362060547, 0.01946419143676758, 0.019512319564819337, 0.019456064224243164, 0.04714796829223633, 0.01969049644470215, 0.019733503341674806, 0.019688447952270507, 0.019715072631835938, 0.01964543914794922, 0.01943961524963379, 0.019769344329833984, 0.01945292854309082, 0.01942835235595703, 0.019588096618652344, 0.01963417625427246, 0.019449855804443358, 0.01949286460876465, 0.01943961524963379, 0.019531776428222656, 0.019449855804443358, 0.01943961524963379, 0.019733503341674806, 0.019647487640380858, 0.01960857582092285, 0.019594240188598632, 0.01965056037902832, 0.019706880569458008, 0.019439647674560547, 0.01968943977355957, 0.019710975646972655, 0.01967103958129883, 0.01966182327270508, 0.01960960006713867, 0.019671104431152345, 0.019703744888305664, 0.01960550308227539, 0.01963212776184082, 0.01961881637573242, 0.019914751052856446, 0.021032960891723632, 0.020563968658447264, 0.020339712142944336, 0.02016972732543945, 0.02042367935180664, 0.020288511276245116, 0.0202106876373291, 0.02022707176208496, 0.019929088592529298, 0.019595264434814453, 0.019697664260864257, 0.01965670394897461, 0.01963827133178711, 0.01968639945983887, 0.01943552017211914, 0.019685375213623048, 0.0204769287109375, 0.020320255279541014, 0.02024448013305664, 0.02036844825744629, 0.02019424057006836, 0.019942399978637695, 0.01964543914794922, 0.01964646339416504, 0.02026188850402832, 0.01982464027404785, 0.02025881576538086, 0.04818329620361328, 0.02022809600830078, 0.02007347106933594, 0.01964646339416504, 0.01970790481567383, 0.01967103958129883, 0.01965875244140625, 0.019698688507080078, 0.019573759078979493, 0.01948569679260254, 0.019404800415039062, 0.01965158462524414, 0.019677183151245118, 0.01963417625427246, 0.01966694450378418, 0.019719167709350584, 0.019574783325195313, 0.01946112060546875, 0.019560447692871095, 0.019681280136108398, 0.019495935440063478, 0.01942732810974121, 0.019368959426879884, 0.019764223098754884, 0.020344831466674804, 0.020753408432006838, 0.01984511947631836, 0.019703807830810546, 0.019574783325195313, 0.01965465545654297, 0.019673088073730468, 0.019714048385620117, 0.019685375213623048, 0.01965772819519043, 0.019679231643676756, 0.019723264694213868, 0.01965363121032715, 0.01966694450378418, 0.01968435287475586, 0.01979903984069824, 0.019654687881469728, 0.01968841552734375, 0.01963417625427246, 0.019604480743408204, 0.019697664260864257, 0.019701759338378907, 0.01964031982421875, 0.019741695404052736, 0.019713056564331054, 0.01970479965209961, 0.01967001533508301, 0.019679231643676756, 0.01962700843811035, 0.01967820739746094, 0.01963212776184082, 0.019720191955566405, 0.019705856323242187, 0.01970790481567383, 0.01962291145324707, 0.01964851188659668, 0.01964646339416504, 0.020050943374633787, 0.020362239837646484]",tokens/s,49.293786685871915,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,x,x,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/x/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6693245a-3f6747ed47b21a340b7efe5c;25fbf99c-1f26-419f-be0b-cc77e40fcc27) Repository Not Found for url: https://huggingface.co/x/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: x is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,01-ai/Yi-34B,01-ai/Yi-34B,cuda,0,42,,,True,,,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",llama,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/xglm-4.5B,,cuda,0,42,,,,,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,d35829e539df8480b726c647eeabf91e41eae047,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,,MB,1399.795712,11335.630848,0.0,10689.183744,9357.549568,s,10,0.6931103744506837,0.06931103744506836,0.0007273844990548417,0.06909251022338866,0.06949516677856446,0.0704782543182373,0.07126472434997559,"[0.07146134185791016, 0.06901350402832031, 0.06909916687011719, 0.06926525115966797, 0.06927670288085938, 0.06911849975585938, 0.06899801635742188, 0.06908585357666015, 0.06888761901855468, 0.0689044189453125]",tokens/s,3693.495429250929,kWh,8.190749831126568e-07,4.487266654466331e-07,3.7329838292961606e-06,5.00078547785545e-06,tokens/kWh,51191957.97012746,MB,1400.123392,11335.630848,0.0,10689.183744,9490.407424,s,10,19.49065673828125,1.9490656738281253,0.00564540932897227,1.94903955078125,1.9541781249999999,1.9582277954101563,1.9614675317382813,"[1.9622774658203126, 1.951441650390625, 1.9532781982421874, 1.9501212158203125, 1.944335693359375, 1.9479578857421875, 1.9436546630859375, 1.950752197265625, 1.94369677734375, 1.9431409912109374]",tokens/s,32.323179688585256,kWh,2.291057484873867e-05,1.2555659928028286e-05,8.698272763210706e-05,0.000122448962408874,tokens/kWh,514500.07219852385,,s,629,20.080927755355795,0.03192516336304584,0.008289819758241598,0.030857215881347655,0.03127746543884277,0.03142776336669922,0.10010099670410157,"[0.03142451286315918, 0.03177471923828125, 0.03153203201293946, 0.03144806480407715, 0.03142451286315918, 0.03144704055786133, 0.03201228713989258, 0.0314768009185791, 0.03138041687011719, 0.031473663330078124, 0.031492095947265625, 0.031318016052246093, 0.0313118724822998, 0.03120639991760254, 0.03201433563232422, 0.03167948722839355, 0.03183923149108887, 0.03119206428527832, 0.031884288787841795, 0.031527936935424807, 0.031236095428466795, 0.031080448150634765, 0.03135487937927246, 0.030707712173461913, 0.03096780776977539, 0.031120384216308594, 0.03114291191101074, 0.03096575927734375, 0.030935039520263673, 0.03062579154968262, 0.03098624038696289, 0.0310435848236084, 0.031341567993164066, 0.031072256088256835, 0.03112652778625488, 0.030695423126220703, 0.030818304061889647, 0.0307957763671875, 0.03099750328063965, 0.031162368774414063, 0.03115519905090332, 0.03075379180908203, 0.030819328308105468, 0.03080089569091797, 0.030860288619995117, 0.030834688186645507, 0.03096883201599121, 0.030793727874755858, 0.030851072311401367, 0.030863359451293947, 0.03103443145751953, 0.030746559143066406, 0.03082444763183594, 0.030847999572753908, 0.031389696121215824, 0.03080601692199707, 0.030882816314697265, 0.030833663940429686, 0.030908416748046875, 0.030926847457885744, 0.030856191635131838, 0.03080601692199707, 0.10063871765136718, 0.030942207336425782, 0.030659584045410155, 0.030818304061889647, 0.03100262451171875, 0.030891008377075195, 0.031024127960205077, 0.03138457679748535, 0.031122432708740235, 0.030822399139404297, 0.030999551773071288, 0.031110143661499022, 0.03129958343505859, 0.030757888793945313, 0.031055871963500976, 0.031303680419921875, 0.03147878456115723, 0.030891008377075195, 0.030730272293090822, 0.030776287078857423, 0.030827520370483398, 0.030877695083618165, 0.030817279815673827, 0.03098521614074707, 0.031037439346313478, 0.03082854461669922, 0.03073843193054199, 0.030769151687622072, 0.03102207946777344, 0.03102720069885254, 0.030864383697509764, 0.03103436851501465, 0.031080448150634765, 0.0311910400390625, 0.031140863418579103, 0.03077939224243164, 0.030757888793945313, 0.030833663940429686, 0.0307906551361084, 0.030947328567504883, 0.031071231842041015, 0.030728191375732423, 0.031138816833496095, 0.030909503936767578, 0.03090937614440918, 0.030819328308105468, 0.030851072311401367, 0.03101900863647461, 0.03127910423278808, 0.03097292709350586, 0.03080703926086426, 0.030825471878051756, 0.030826496124267577, 0.030914560317993164, 0.030866432189941406, 0.030948352813720704, 0.030819328308105468, 0.03103027153015137, 0.031172607421875, 0.031127552032470703, 0.030859264373779297, 0.03138559913635254, 0.031144960403442383, 0.10035302734375, 0.030691328048706053, 0.032535552978515625, 0.03151769638061523, 0.031337472915649416, 0.031088640213012695, 0.031152128219604492, 0.0313118724822998, 0.030978080749511718, 0.030718944549560548, 0.03077324867248535, 0.030856191635131838, 0.030856191635131838, 0.03080089569091797, 0.030745599746704103, 0.031239200592041015, 0.03086332893371582, 0.03098931121826172, 0.03079475212097168, 0.030793727874755858, 0.031107072830200196, 0.03125760078430176, 0.03127603149414063, 0.03138355255126953, 0.031081472396850586, 0.030714879989624022, 0.030870527267456056, 0.03097702407836914, 0.030809087753295897, 0.03078758430480957, 0.031124479293823244, 0.03127705574035645, 0.030869504928588868, 0.030964736938476563, 0.031160320281982422, 0.03080601692199707, 0.031014911651611327, 0.03091254425048828, 0.030758880615234376, 0.03098624038696289, 0.031235071182250978, 0.031185920715332032, 0.030673919677734376, 0.031074304580688477, 0.0309749755859375, 0.030904319763183592, 0.030930944442749023, 0.031108095169067384, 0.031130624771118165, 0.03082956886291504, 0.030907392501831055, 0.030872575759887694, 0.03082035255432129, 0.03094528007507324, 0.030863359451293947, 0.03136000061035156, 0.030840831756591795, 0.03079475212097168, 0.030852096557617188, 0.030861312866210938, 0.031080448150634765, 0.030814207077026368, 0.03082035255432129, 0.10102067565917969, 0.030846975326538087, 0.03077017593383789, 0.03081318473815918, 0.03134668731689453, 0.031242271423339844, 0.030676959991455078, 0.03096883201599121, 0.030697471618652345, 0.03081216049194336, 0.03098111915588379, 0.03080601692199707, 0.030895103454589845, 0.03079680061340332, 0.031094783782958983, 0.030895103454589845, 0.030680063247680665, 0.03079270362854004, 0.030719999313354493, 0.030810111999511718, 0.030871551513671876, 0.030729215621948244, 0.03057766342163086, 0.03103539276123047, 0.030717952728271485, 0.030846975326538087, 0.0307906551361084, 0.031718399047851564, 0.03120947265625, 0.03092787170410156, 0.03075584030151367, 0.03133235168457031, 0.030833663940429686, 0.03118284797668457, 0.030871551513671876, 0.03080089569091797, 0.03102207946777344, 0.03081216049194336, 0.030719999313354493, 0.031025152206420898, 0.030670848846435547, 0.030855167388916017, 0.03081523132324219, 0.030912511825561522, 0.03082444763183594, 0.03081020736694336, 0.030791584014892577, 0.03127705574035645, 0.031265792846679685, 0.03129343986511231, 0.031065088272094726, 0.03081625556945801, 0.03119206428527832, 0.03097395133972168, 0.030873600006103515, 0.03156582450866699, 0.03138662338256836, 0.03082342338562012, 0.030707712173461913, 0.030988288879394532, 0.030742528915405274, 0.03152595138549805, 0.030772159576416016, 0.10050867462158203, 0.03074662399291992, 0.030819328308105468, 0.03076710319519043, 0.03059814453125, 0.03075993537902832, 0.03080499267578125, 0.03100262451171875, 0.030745599746704103, 0.03075379180908203, 0.030785535812377928, 0.030862335205078126, 0.030672895431518556, 0.030705696105957032, 0.030995424270629884, 0.03127193641662598, 0.03102822494506836, 0.0309616641998291, 0.030742528915405274, 0.030908416748046875, 0.030769151687622072, 0.030709760665893555, 0.03078963279724121, 0.030920703887939452, 0.03081523132324219, 0.030912511825561522, 0.031109119415283205, 0.03079270362854004, 0.03077939224243164, 0.030742528915405274, 0.030873600006103515, 0.030963712692260743, 0.03080396842956543, 0.030913536071777343, 0.030752767562866212, 0.030818304061889647, 0.030861312866210938, 0.03082041549682617, 0.030872512817382812, 0.03096063995361328, 0.030825471878051756, 0.03077836799621582, 0.03081318473815918, 0.030859264373779297, 0.030883840560913086, 0.030838783264160157, 0.031032320022583007, 0.03137228775024414, 0.030983167648315428, 0.030742528915405274, 0.030843904495239258, 0.030877695083618165, 0.030884864807128907, 0.030818304061889647, 0.030875648498535156, 0.030909439086914063, 0.030695423126220703, 0.030818304061889647, 0.030841856002807616, 0.030882879257202147, 0.030872543334960936, 0.030844959259033203, 0.030943168640136718, 0.10048102569580078, 0.030737407684326173, 0.030765056610107422, 0.030661632537841797, 0.030683135986328124, 0.030774272918701173, 0.03079884719848633, 0.03101900863647461, 0.030900224685668946, 0.030852096557617188, 0.030817279815673827, 0.03078144073486328, 0.03114905548095703, 0.030688255310058594, 0.030807071685791016, 0.03122172737121582, 0.03248953628540039, 0.031426496505737304, 0.030987295150756836, 0.03063804817199707, 0.03078041648864746, 0.030938112258911132, 0.03075481605529785, 0.030888959884643553, 0.03075379180908203, 0.030867456436157226, 0.03083263969421387, 0.030845951080322266, 0.030814207077026368, 0.03079270362854004, 0.03094528007507324, 0.03097804832458496, 0.03075584030151367, 0.03079884719848633, 0.030863359451293947, 0.030874624252319335, 0.03082035255432129, 0.030891008377075195, 0.030923776626586914, 0.031177728652954102, 0.03097395133972168, 0.030879743576049806, 0.030857215881347655, 0.030897151947021483, 0.03095961570739746, 0.03080089569091797, 0.030850048065185546, 0.03102822494506836, 0.03078348731994629, 0.03082444763183594, 0.030979103088378906, 0.03136508750915527, 0.031340543746948245, 0.030835712432861328, 0.030868480682373047, 0.03096985626220703, 0.030817279815673827, 0.030864383697509764, 0.030869504928588868, 0.030877695083618165, 0.030839807510375978, 0.03082342338562012, 0.030835712432861328, 0.1001164779663086, 0.030713855743408205, 0.030827520370483398, 0.03100160026550293, 0.030695423126220703, 0.030825471878051756, 0.030818304061889647, 0.0309586238861084, 0.030872543334960936, 0.03082035255432129, 0.030748672485351562, 0.03078451156616211, 0.030766080856323243, 0.030650367736816408, 0.030769151687622072, 0.030914560317993164, 0.03079270362854004, 0.03079167938232422, 0.0307957763671875, 0.030707712173461913, 0.03073843193054199, 0.030684160232543944, 0.030883840560913086, 0.031037439346313478, 0.03073843193054199, 0.03075071907043457, 0.030882816314697265, 0.030845951080322266, 0.030863359451293947, 0.03081523132324219, 0.030839807510375978, 0.03105075263977051, 0.0307640323638916, 0.03080703926086426, 0.030833663940429686, 0.030860288619995117, 0.030940160751342774, 0.030818304061889647, 0.030876672744750977, 0.03097907257080078, 0.03079270362854004, 0.03084492874145508, 0.03077120018005371, 0.030915584564208985, 0.03085318374633789, 0.030823360443115233, 0.030855167388916017, 0.03099750328063965, 0.03077836799621582, 0.030839807510375978, 0.03082444763183594, 0.030907392501831055, 0.030903295516967775, 0.03080806350708008, 0.030870527267456056, 0.03142860794067383, 0.030842880249023437, 0.03082444763183594, 0.03095347213745117, 0.030894079208374024, 0.030860288619995117, 0.030958591461181642, 0.030757888793945313, 0.10313728332519531, 0.03141734313964844, 0.031082496643066407, 0.031084543228149415, 0.030875648498535156, 0.031067136764526368, 0.030841856002807616, 0.03094937515258789, 0.030701568603515625, 0.031129600524902344, 0.03082137680053711, 0.03076710319519043, 0.03078963279724121, 0.030850048065185546, 0.0307589111328125, 0.031106048583984375, 0.03075379180908203, 0.030831615447998048, 0.030946304321289062, 0.030830591201782227, 0.030892032623291016, 0.0307906551361084, 0.03077631950378418, 0.03098624038696289, 0.03075071907043457, 0.030915584564208985, 0.030954496383666992, 0.030877695083618165, 0.03083263969421387, 0.030874624252319335, 0.030749696731567383, 0.0309749755859375, 0.030726144790649414, 0.030848031997680665, 0.030838752746582033, 0.030861312866210938, 0.0307957763671875, 0.030934015274047853, 0.030846975326538087, 0.031057920455932617, 0.03080089569091797, 0.030841856002807616, 0.030769151687622072, 0.031152128219604492, 0.03118182373046875, 0.031301631927490234, 0.030906368255615234, 0.030983167648315428, 0.0307957763671875, 0.030892032623291016, 0.03076300811767578, 0.031006719589233397, 0.03131699180603027, 0.030931968688964844, 0.0309749755859375, 0.03131391906738281, 0.030793727874755858, 0.030859264373779297, 0.03080089569091797, 0.030900224685668946, 0.03080089569091797, 0.03098624038696289, 0.03078348731994629, 0.09996697235107421, 0.030726144790649414, 0.030734399795532226, 0.030624704360961916, 0.03073023986816406, 0.030854143142700196, 0.030649343490600587, 0.03078860855102539, 0.030646272659301758, 0.030711807250976563, 0.03075993537902832, 0.03062272071838379, 0.03073945617675781, 0.030740480422973632, 0.03077120018005371, 0.03094118309020996, 0.03108768081665039, 0.030775232315063475, 0.030745599746704103, 0.03077529525756836, 0.03079680061340332, 0.03061043167114258, 0.030718975067138672, 0.030883840560913086, 0.03085312080383301, 0.030845951080322266, 0.0307906551361084, 0.030834688186645507, 0.030852096557617188, 0.03075481605529785, 0.030690303802490236, 0.03085312080383301, 0.030954496383666992, 0.03072204780578613, 0.030740480422973632, 0.030846975326538087, 0.030856191635131838, 0.03080396842956543, 0.03128217506408691, 0.0310435848236084, 0.03080703926086426, 0.030846975326538087, 0.030825471878051756, 0.030946304321289062, 0.030860288619995117, 0.030814207077026368, 0.030877695083618165, 0.03097088050842285, 0.03083263969421387, 0.03075174331665039, 0.030840831756591795, 0.030867456436157226, 0.030962688446044922, 0.03159347152709961, 0.03078246307373047, 0.030929920196533203, 0.03074764823913574, 0.030814207077026368, 0.03129241561889649, 0.031509504318237305, 0.030930944442749023, 0.03080601692199707, 0.030851072311401367, 0.10006118774414062, 0.030687231063842774, 0.030809087753295897, 0.030819328308105468, 0.030670848846435547, 0.030708736419677734, 0.03074355125427246, 0.030852096557617188, 0.030779487609863283, 0.030888864517211914, 0.030857215881347655, 0.03080089569091797, 0.030831615447998048, 0.030745599746704103, 0.030694400787353516, 0.030874624252319335, 0.03079680061340332, 0.030741504669189453, 0.030835712432861328, 0.030845951080322266, 0.031048704147338867, 0.030672895431518556, 0.03098624038696289, 0.03116543960571289, 0.03096985626220703, 0.030876672744750977, 0.030867456436157226, 0.030810111999511718, 0.030841856002807616, 0.03080703926086426, 0.030793727874755858, 0.03080396842956543, 0.030689279556274415, 0.03074051284790039, 0.030847967147827147, 0.030834688186645507, 0.03082342338562012, 0.03096780776977539, 0.030878719329833985, 0.031080448150634765, 0.03074355125427246, 0.03080294418334961, 0.03077734375, 0.030846975326538087, 0.030857215881347655, 0.030872575759887694, 0.030827520370483398, 0.03097088050842285, 0.030814207077026368, 0.030838783264160157, 0.03079782485961914, 0.030915647506713866, 0.03086329650878906, 0.03085312080383301, 0.030865407943725585, 0.030956575393676758, 0.03081007957458496, 0.03084492874145508, 0.030868480682373047, 0.030846975326538087, 0.030876672744750977, 0.030845951080322266, 0.03083776092529297]",tokens/s,31.323253968295255,,,main,False,False,,,,,,,,,,,,,,,,,,,,,,, float16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/xglm-7.5B,,cuda,0,42,,,,,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,d35829e539df8480b726c647eeabf91e41eae047,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,,MB,1396.400128,17796.95616,0.0,17150.509056,15280.037888,s,10,0.9786758422851562,0.09786758422851563,0.0014127563871552914,0.09739243316650391,0.0981432846069336,0.10010927429199219,0.10168206604003907,"[0.10207526397705079, 0.09754768371582032, 0.09725894165039063, 0.09719065856933594, 0.09762809753417968, 0.09734815979003907, 0.09721334075927734, 0.09727059173583984, 0.09770639801025391, 0.09743670654296875]",tokens/s,2615.779290129953,kWh,1.1484344977133997e-06,6.292862184919321e-07,5.927890394951905e-06,7.705611111157238e-06,tokens/kWh,33222543.456589468,MB,1396.400128,17796.95616,0.0,17150.509056,15412.539904,s,10,24.20992919921875,2.4209929199218747,0.0037899369998390686,2.4224346923828124,2.424226171875,2.4252472412109376,2.4260640966796876,"[2.420181396484375, 2.423999267578125, 2.42294921875, 2.426268310546875, 2.4236376953125, 2.423234375, 2.4151728515625, 2.421920166015625, 2.41392041015625, 2.4186455078125]",tokens/s,26.02238093370096,kWh,2.8460493765474012e-05,1.5597388115374024e-05,0.00012703610486504177,0.00017109398674588984,tokens/kWh,368218.6685705565,,s,629,25.047858192443837,0.0398217141374306,0.01164268431675323,0.038384639739990234,0.038637157440185546,0.03874039535522461,0.13619531677246094,"[0.03893254470825196, 0.03824736022949219, 0.03820646286010742, 0.03827711868286133, 0.03865087890625, 0.038286334991455076, 0.038338558197021484, 0.03827199935913086, 0.03840921783447265, 0.03823206329345703, 0.03818905639648437, 0.03827199935913086, 0.03863142395019531, 0.03845529556274414, 0.038284286499023434, 0.03842355346679688, 0.03847782516479492, 0.03833139038085937, 0.038204414367675785, 0.03832320022583008, 0.03834470367431641, 0.03827609634399414, 0.03832627105712891, 0.038348800659179685, 0.03834982299804687, 0.038577152252197267, 0.03825766372680664, 0.03834163284301758, 0.0385167350769043, 0.038269950866699216, 0.03840409469604492, 0.03838054275512695, 0.03830988693237305, 0.03832012939453125, 0.03834368133544922, 0.03849728012084961, 0.038316032409667966, 0.03838054275512695, 0.038348800659179685, 0.03860172653198242, 0.03825971221923828, 0.03833657455444336, 0.038287296295166015, 0.03851878356933594, 0.038370304107666016, 0.03862732696533203, 0.03843993759155274, 0.03854848098754883, 0.03835084915161133, 0.03846451187133789, 0.038555648803710936, 0.038588417053222655, 0.03837545776367188, 0.03841737747192383, 0.03846355056762695, 0.038685630798339844, 0.038312961578369144, 0.038383617401123046, 0.03840204620361328, 0.03859558486938477, 0.0382782096862793, 0.03880953598022461, 0.13671218872070312, 0.03837849426269531, 0.038438911437988284, 0.03823616027832031, 0.03827711868286133, 0.038214656829833986, 0.03836723327636719, 0.038847488403320314, 0.03859251022338867, 0.038269950866699216, 0.03851059341430664, 0.03835903930664063, 0.03850239944458008, 0.03839078521728516, 0.03855462265014648, 0.03834982299804687, 0.03848294448852539, 0.038386688232421876, 0.0384502067565918, 0.0382432975769043, 0.03829759979248047, 0.03832320022583008, 0.038468608856201174, 0.038474750518798825, 0.03904000091552735, 0.03836723327636719, 0.03851264190673828, 0.03847577667236328, 0.03836415863037109, 0.03829759979248047, 0.038299648284912106, 0.03844812774658203, 0.038468608856201174, 0.03847884750366211, 0.03846553421020508, 0.03845632171630859, 0.03833958435058594, 0.038432769775390625, 0.038414337158203124, 0.03848294448852539, 0.03872665786743164, 0.03843174362182617, 0.038757377624511716, 0.03828736114501953, 0.038351871490478515, 0.03846144104003906, 0.038438911437988284, 0.03846144104003906, 0.03867443084716797, 0.03851264190673828, 0.038670337677001954, 0.03863961410522461, 0.038545406341552735, 0.03846656036376953, 0.038474750518798825, 0.03860684967041016, 0.03854336166381836, 0.038421504974365236, 0.03863142395019531, 0.03862944030761719, 0.038526912689208985, 0.03851776123046875, 0.03847782516479492, 0.13622169494628905, 0.03817062377929688, 0.03835801696777344, 0.03837235260009766, 0.03830169677734375, 0.03899084854125977, 0.03868672180175781, 0.03850342559814453, 0.038551616668701175, 0.03845830535888672, 0.03848396682739258, 0.038609920501708986, 0.038545406341552735, 0.03860172653198242, 0.038558719635009765, 0.03851980972290039, 0.03853619384765625, 0.038470657348632815, 0.0385269775390625, 0.03852185440063476, 0.038348800659179685, 0.03853420639038086, 0.03867232131958008, 0.03851366424560547, 0.03864070510864258, 0.03834463882446289, 0.03861708831787109, 0.03820345687866211, 0.03828012847900391, 0.03844812774658203, 0.038537246704101566, 0.038314014434814456, 0.03831391906738281, 0.03830374526977539, 0.038508544921875, 0.03822079849243164, 0.038296577453613284, 0.038452224731445314, 0.03828326416015625, 0.03854336166381836, 0.03832831954956055, 0.038509567260742186, 0.03856588745117188, 0.03820851135253906, 0.03833651351928711, 0.03831193542480469, 0.03833651351928711, 0.038629375457763675, 0.038397953033447264, 0.038302719116210936, 0.038612991333007815, 0.038545406341552735, 0.03861503982543945, 0.038263809204101565, 0.038504447937011715, 0.03855257415771484, 0.03860275268554687, 0.03830169677734375, 0.0385167350769043, 0.03829452896118164, 0.03855974578857422, 0.03825766372680664, 0.038346752166748044, 0.13624729919433592, 0.038345729827880856, 0.038397953033447264, 0.038161407470703124, 0.0382105598449707, 0.03826483154296875, 0.03835801696777344, 0.038269950866699216, 0.03839590454101562, 0.038324222564697266, 0.038330368041992184, 0.03843484878540039, 0.03856278228759766, 0.03851059341430664, 0.03820032119750977, 0.03832320022583008, 0.038263809204101565, 0.03849523162841797, 0.038449153900146485, 0.038212608337402344, 0.038266880035400394, 0.03836723327636719, 0.03829555130004883, 0.0385873908996582, 0.038474750518798825, 0.038365184783935545, 0.03923766326904297, 0.038705120086669924, 0.038504447937011715, 0.03827302551269531, 0.03828531265258789, 0.038346782684326175, 0.03861603164672851, 0.03838259124755859, 0.03872665786743164, 0.03835903930664063, 0.03871443176269531, 0.03862521743774414, 0.0386324462890625, 0.038781951904296875, 0.038599681854248044, 0.038591487884521485, 0.0383375358581543, 0.03851366424560547, 0.03851468658447266, 0.038300670623779294, 0.03833651351928711, 0.03857305526733398, 0.038335487365722655, 0.0384983024597168, 0.038942718505859376, 0.038886398315429685, 0.038809600830078124, 0.038687744140625, 0.038708225250244144, 0.03878604888916016, 0.03877689743041992, 0.03858118438720703, 0.03868979263305664, 0.038626304626464845, 0.03872665786743164, 0.03863142395019531, 0.03872051239013672, 0.1369416046142578, 0.03824534225463867, 0.03838054275512695, 0.038553600311279294, 0.03829043197631836, 0.038182910919189454, 0.03817881774902344, 0.03837952041625976, 0.03847987365722656, 0.03826278305053711, 0.03836108779907227, 0.03853004837036133, 0.03836723327636719, 0.03820646286010742, 0.03847372817993164, 0.03853004837036133, 0.03828224182128906, 0.038299648284912106, 0.03843993759155274, 0.03836415863037109, 0.03828326416015625, 0.038163455963134765, 0.03856079864501953, 0.03867343902587891, 0.03868876647949219, 0.03853715133666992, 0.038556671142578124, 0.0385167350769043, 0.03866009521484375, 0.038604801177978515, 0.03854950332641602, 0.038438911437988284, 0.03830476760864258, 0.0383631362915039, 0.038365184783935545, 0.03823308944702149, 0.038354942321777344, 0.03850649642944336, 0.03863552093505859, 0.03858124923706055, 0.038362110137939456, 0.03848089599609375, 0.03842969512939453, 0.038269950866699216, 0.03854438400268555, 0.03856793594360351, 0.03841126251220703, 0.038431808471679686, 0.038393791198730466, 0.03833958435058594, 0.03873791885375977, 0.03841126251220703, 0.038523902893066404, 0.03831808090209961, 0.038343711853027346, 0.03840304183959961, 0.038529022216796875, 0.03836415863037109, 0.03869900894165039, 0.038458366394042966, 0.03893964767456055, 0.03917311859130859, 0.03882393646240234, 0.1364674530029297, 0.03832940673828125, 0.03857094573974609, 0.03832729721069336, 0.038383617401123046, 0.03868672180175781, 0.038470657348632815, 0.038661121368408206, 0.0383744010925293, 0.03821363067626953, 0.038204414367675785, 0.03854131317138672, 0.03839897537231445, 0.03838259124755859, 0.038623233795166016, 0.03852185440063476, 0.03859251022338867, 0.03834368133544922, 0.03849420928955078, 0.038247425079345705, 0.038383617401123046, 0.038424575805664066, 0.0383631362915039, 0.03845734405517578, 0.038474750518798825, 0.038593536376953126, 0.038491134643554685, 0.03833343887329101, 0.03861503982543945, 0.038365184783935545, 0.03848294448852539, 0.038509567260742186, 0.038419456481933595, 0.038422527313232424, 0.03854848098754883, 0.03837747192382813, 0.038458366394042966, 0.038540287017822264, 0.038626304626464845, 0.03873484802246094, 0.038719486236572266, 0.03863654327392578, 0.03865087890625, 0.0385873908996582, 0.03868467330932617, 0.03846963119506836, 0.03832831954956055, 0.038345729827880856, 0.03838566589355469, 0.03831193542480469, 0.038397953033447264, 0.038354942321777344, 0.03840409469604492, 0.03830169677734375, 0.03841126251220703, 0.03843379211425781, 0.038435840606689455, 0.038312961578369144, 0.03847577667236328, 0.03828121566772461, 0.03848806381225586, 0.03829452896118164, 0.03845017623901367, 0.13601181030273438, 0.038225887298583984, 0.03862527847290039, 0.03881369781494141, 0.03866726303100586, 0.0385269775390625, 0.03822489547729492, 0.038316032409667966, 0.03839078521728516, 0.03852288055419922, 0.038353919982910156, 0.03843993759155274, 0.03824639892578125, 0.03819417572021484, 0.03822079849243164, 0.03824639892578125, 0.03830886459350586, 0.03823616027832031, 0.03832627105712891, 0.038166526794433595, 0.038265857696533206, 0.03816447830200195, 0.03826278305053711, 0.03847270584106445, 0.03836006546020508, 0.038247425079345705, 0.03826073455810547, 0.03824639892578125, 0.038286334991455076, 0.038225921630859375, 0.03827302551269531, 0.038307838439941407, 0.03826176071166992, 0.03823820877075195, 0.03830988693237305, 0.038250495910644534, 0.03834163284301758, 0.038248481750488283, 0.03828118515014649, 0.03827916717529297, 0.038280193328857424, 0.038257728576660155, 0.038343616485595707, 0.038217727661132815, 0.038316032409667966, 0.03827510452270508, 0.03838665771484375, 0.038371326446533204, 0.038329345703125, 0.038430721282958984, 0.038474750518798825, 0.03827302551269531, 0.038365184783935545, 0.038351871490478515, 0.03836006546020508, 0.03837849426269531, 0.03841228866577148, 0.03827199935913086, 0.03838566589355469, 0.03829145431518555, 0.03829862213134766, 0.038319103240966795, 0.038403072357177735, 0.1361274871826172, 0.03830169677734375, 0.03861094284057617, 0.03823820877075195, 0.038416446685791014, 0.03823712158203125, 0.03841740798950195, 0.038373374938964845, 0.03835289764404297, 0.03845632171630859, 0.0383939208984375, 0.03887916946411133, 0.03893145751953125, 0.03858227157592774, 0.03844403076171875, 0.038563838958740236, 0.03828224182128906, 0.038373374938964845, 0.03830169677734375, 0.038286334991455076, 0.038348800659179685, 0.03816447830200195, 0.038370304107666016, 0.03863040161132812, 0.039180286407470705, 0.038695934295654294, 0.03828736114501953, 0.03828531265258789, 0.03829248046875, 0.0382105598449707, 0.03823206329345703, 0.038299648284912106, 0.038523902893066404, 0.03851059341430664, 0.03826073455810547, 0.03832831954956055, 0.038566913604736325, 0.03843686294555664, 0.038742046356201175, 0.03834774398803711, 0.03841843032836914, 0.038414337158203124, 0.03856793594360351, 0.03853619384765625, 0.038583297729492184, 0.03832118225097656, 0.03854844665527344, 0.03863040161132812, 0.03844403076171875, 0.038400001525878906, 0.03841843032836914, 0.038340606689453126, 0.0383109130859375, 0.038349857330322264, 0.038457313537597654, 0.03851980972290039, 0.03840204620361328, 0.03829248046875, 0.03842764663696289, 0.0383631362915039, 0.038384639739990234, 0.038370304107666016, 0.03841535949707031, 0.13680230712890626, 0.03813888168334961, 0.03815321731567383, 0.0381399040222168, 0.03819007873535156, 0.038201377868652346, 0.03819926452636719, 0.03819007873535156, 0.038161407470703124, 0.03814912033081055, 0.03834777450561523, 0.03815935897827148, 0.03819420623779297, 0.03817571258544922, 0.038225921630859375, 0.038265857696533206, 0.03823513412475586, 0.03832831954956055, 0.038204414367675785, 0.03819417572021484, 0.03828636932373047, 0.0381890869140625, 0.038222782135009764, 0.03834777450561523, 0.03825254440307617, 0.03824025726318359, 0.03823513412475586, 0.03819007873535156, 0.03829350280761719, 0.03822182464599609, 0.038266880035400394, 0.03835084915161133, 0.03824332809448242, 0.038312961578369144, 0.03832217788696289, 0.038245376586914064, 0.03826278305053711, 0.0382371826171875, 0.03831808090209961, 0.038263809204101565, 0.038286334991455076, 0.03823923110961914, 0.0383559684753418, 0.038234111785888675, 0.03832012939453125, 0.03830681610107422, 0.03831299209594727, 0.03840713500976563, 0.03827507019042969, 0.03827302551269531, 0.038391807556152346, 0.03825459289550781, 0.03834163284301758, 0.038286334991455076, 0.038351871490478515, 0.038392833709716793, 0.03839385604858398, 0.038286334991455076, 0.038469696044921876, 0.03923654556274414, 0.03904819107055664, 0.038653953552246094, 0.03860070419311523, 0.13646949768066408, 0.03844710540771484, 0.03849523162841797, 0.03846553421020508, 0.03853414535522461, 0.03852799987792969, 0.03862220764160156, 0.03857305526733398, 0.03847270584106445, 0.03856588745117188, 0.03855155181884766, 0.038474750518798825, 0.03844812774658203, 0.038470657348632815, 0.038612991333007815, 0.038537216186523435, 0.03831193542480469, 0.03824947357177735, 0.038266880035400394, 0.038191104888916014, 0.0383109130859375, 0.03824127960205078, 0.03831398391723633, 0.03830681610107422, 0.03828224182128906, 0.03820342254638672, 0.038328289031982425, 0.038162494659423826, 0.038295520782470706, 0.038233055114746096, 0.0382740478515625, 0.038300670623779294, 0.038302719116210936, 0.03834265518188477, 0.038305824279785156, 0.03828732681274414, 0.0383744010925293, 0.0382105598449707, 0.03840512084960938, 0.03842559814453125, 0.038384639739990234, 0.03830988693237305, 0.03837644958496094, 0.03830681610107422, 0.03844812774658203, 0.03829350280761719, 0.038419456481933595, 0.03840204620361328, 0.038452224731445314, 0.03833958435058594, 0.03840409469604492, 0.03827916717529297, 0.03838259124755859, 0.03830476760864258, 0.03842764663696289, 0.03841843032836914, 0.03843174362182617, 0.03828326416015625, 0.03844095993041992, 0.038302719116210936, 0.03849216079711914, 0.03859254455566406, 0.038434783935546876]",tokens/s,25.111927541563194,,,main,False,False,,,,,,,,,,,,,,,,,,,,,,, float16-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-14B,Qwen/Qwen1.5-14B,cuda,0,42,,,True,,,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",qwen2,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,meta-llama/Llama-2-70b-hf,meta-llama/Llama-2-70b-hf,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ self.load_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 250, in load_model_with_no_weights self.pretrained_model = self.automodel_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3710, in from_pretrained model = cls(config, *model_args, **model_kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 1101, in __init__ self.model = LlamaModel(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 884, in __init__ [LlamaDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 884, in [LlamaDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 679, in __init__ self.mlp = LlamaMLP(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 192, in __init__ self.gate_proj = nn.Linear(self.hidden_size, self.intermediate_size, bias=config.mlp_bias) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/linear.py"", line 98, in __init__ self.weight = Parameter(torch.empty((out_features, in_features), **factory_kwargs)) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_device.py"", line 78, in __torch_function__ return func(*args, **kwargs) torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 448.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Deci/DeciCoder-1b,Deci/DeciCoder-1b,cuda,0,42,,,True,,,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",llama,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,meta-llama/Meta-Llama-3-70B,meta-llama/Meta-Llama-3-70B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ self.load_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 250, in load_model_with_no_weights self.pretrained_model = self.automodel_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3710, in from_pretrained model = cls(config, *model_args, **model_kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 1101, in __init__ self.model = LlamaModel(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 884, in __init__ [LlamaDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 884, in [LlamaDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 679, in __init__ self.mlp = LlamaMLP(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 193, in __init__ self.up_proj = nn.Linear(self.hidden_size, self.intermediate_size, bias=config.mlp_bias) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/linear.py"", line 98, in __init__ self.weight = Parameter(torch.empty((out_features, in_features), **factory_kwargs)) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_device.py"", line 78, in __torch_function__ return func(*args, **kwargs) torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 448.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,r,r,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/r/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669473cb-28a2688f5f24ed381eeb55e5;a69867d9-f9ac-45b0-8f60-bb1458245830) Repository Not Found for url: https://huggingface.co/r/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: r is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,togethercomputer/RedPajama-INCITE-Base-7B-v0.1,togethercomputer/RedPajama-INCITE-Base-7B-v0.1,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1302.740992,15192.293376,0.0,14545.846272,14291.49696,s,10,0.9347154541015626,0.09347154541015626,0.0012546807259258953,0.09305742263793945,0.09355379028320313,0.0953935333251953,0.09686532775878906,"[0.0972332763671875, 0.09314495849609375, 0.09306623840332032, 0.09307231903076171, 0.0929599380493164, 0.09304860687255859, 0.09308080291748047, 0.09304192352294922, 0.09303504180908204, 0.0930323486328125]",tokens/s,2738.801406103467,kWh,1.101663208915373e-06,6.03658020195831e-07,6.038651950259344e-06,7.743973179370548e-06,tokens/kWh,33057965.73288344,MB,1302.740992,15192.293376,0.0,14545.846272,14489.229824,s,10,23.810830078125,2.3810830078125003,0.003005048913985965,2.381181396484375,2.3852510986328124,2.3853768920898437,2.385477526855469,"[2.38522314453125, 2.383134521484375, 2.376668212890625, 2.3812431640625, 2.385502685546875, 2.378052490234375, 2.3767705078125, 2.38064404296875, 2.38111962890625, 2.3824716796875]",tokens/s,26.458548397217815,kWh,2.8157483971535917e-05,1.5431305779633878e-05,0.0001252465364111398,0.0001688353261623096,tokens/kWh,373144.65776809666,,s,629,24.610084911346448,0.039125731178611185,0.011092873257697582,0.037768192291259765,0.03802439575195313,0.03818704681396485,0.13110745056152345,"[0.03818707275390625, 0.03814086532592773, 0.03785830307006836, 0.03823206329345703, 0.038284320831298825, 0.0382504653930664, 0.03815526580810547, 0.03837139129638672, 0.03820025634765625, 0.037969921112060545, 0.03824025726318359, 0.03848806381225586, 0.03822284698486328, 0.037967872619628903, 0.038010879516601564, 0.0377077751159668, 0.0375623664855957, 0.037622814178466794, 0.03762480163574219, 0.037599231719970705, 0.037644287109375, 0.03765350341796875, 0.037705726623535156, 0.037648384094238284, 0.037645313262939455, 0.03767193603515625, 0.037664768218994144, 0.03767500686645508, 0.03768012619018555, 0.03767603302001953, 0.0378081283569336, 0.03768320083618164, 0.03769548797607422, 0.037738494873046875, 0.037678081512451174, 0.037706817626953125, 0.03770566558837891, 0.03777024078369141, 0.03777228927612305, 0.03856486511230469, 0.037855232238769534, 0.03779174423217774, 0.03772927856445313, 0.03772108840942383, 0.03776921463012695, 0.0377446403503418, 0.03784089660644531, 0.03774566268920899, 0.03783987045288086, 0.037800960540771485, 0.037789695739746096, 0.0377968635559082, 0.037773311614990236, 0.0377446403503418, 0.037900318145751954, 0.03782345581054687, 0.03778662490844727, 0.03780307388305664, 0.03778144073486328, 0.037795841217041014, 0.03770982360839844, 0.037663745880126956, 0.1312788543701172, 0.03743334579467773, 0.037454849243164064, 0.03743145751953125, 0.03745775985717773, 0.03745487976074219, 0.03748451232910156, 0.03765862274169922, 0.03766886520385742, 0.037763072967529294, 0.038112255096435545, 0.037961727142333986, 0.037838848114013675, 0.03795558547973633, 0.03783987045288086, 0.03803955078125, 0.03782144165039063, 0.03760435104370117, 0.03789823913574219, 0.03762073516845703, 0.03776409530639648, 0.03793920135498047, 0.03775795364379883, 0.037718017578125, 0.03768217468261719, 0.037771263122558595, 0.038029312133789066, 0.03790643310546875, 0.03772825622558594, 0.03785318374633789, 0.03777740859985351, 0.0378419189453125, 0.03773952102661133, 0.03781119918823242, 0.037738494873046875, 0.037754878997802735, 0.037923839569091795, 0.03781836700439453, 0.037713920593261716, 0.03796889495849609, 0.03838566589355469, 0.03891302490234375, 0.038319103240966795, 0.03804159927368164, 0.03793817520141601, 0.03779072189331055, 0.037787647247314454, 0.037943294525146484, 0.037763072967529294, 0.03775795364379883, 0.03778559875488281, 0.03782553482055664, 0.03799961471557617, 0.037935104370117184, 0.03782147216796875, 0.037895137786865235, 0.03780710220336914, 0.03772723388671875, 0.037748737335205076, 0.0377077751159668, 0.037789695739746096, 0.037789695739746096, 0.037838848114013675, 0.13117543029785156, 0.037579776763916016, 0.037607425689697264, 0.03756032180786133, 0.03760844802856445, 0.03756851196289063, 0.037615615844726565, 0.03772108840942383, 0.03762483215332031, 0.03757056045532227, 0.037615615844726565, 0.037577728271484374, 0.03763097763061524, 0.03766681671142578, 0.037601280212402347, 0.03775283050537109, 0.03761459350585938, 0.037601280212402347, 0.03764019012451172, 0.03760639953613281, 0.03766886520385742, 0.03770982360839844, 0.037663745880126956, 0.03773952102661133, 0.03764633560180664, 0.03773235321044922, 0.03787676620483398, 0.037732318878173826, 0.037751808166503906, 0.03768627166748047, 0.03768217468261719, 0.03774259185791016, 0.037748737335205076, 0.03763097763061524, 0.03769753646850586, 0.037738494873046875, 0.037735424041748046, 0.037698558807373043, 0.037703678131103514, 0.037921791076660154, 0.03770982360839844, 0.03759718322753906, 0.037738494873046875, 0.03772723388671875, 0.037781505584716796, 0.03780915069580078, 0.037819393157958986, 0.03783987045288086, 0.0377968635559082, 0.037602302551269534, 0.03765862274169922, 0.037765121459960936, 0.037749759674072264, 0.03776204681396484, 0.037806079864501956, 0.03786240005493164, 0.037874687194824216, 0.03785728073120117, 0.037804031372070314, 0.03779891204833984, 0.03783168029785156, 0.038012928009033206, 0.03809894561767578, 0.13125120544433594, 0.03746201705932617, 0.03752140808105469, 0.037738494873046875, 0.037730304718017575, 0.037601280212402347, 0.037585918426513674, 0.03769241714477539, 0.037618686676025394, 0.037618686676025394, 0.03776716613769531, 0.03799347305297852, 0.03789516830444336, 0.03784601593017578, 0.037720062255859374, 0.03769753646850586, 0.03769139099121094, 0.03756032180786133, 0.037787647247314454, 0.03791360092163086, 0.03768729782104492, 0.0376258544921875, 0.03793203353881836, 0.037958656311035156, 0.03789823913574219, 0.03785625457763672, 0.03799961471557617, 0.0378388786315918, 0.037971935272216796, 0.037967872619628903, 0.03789926528930664, 0.03786240005493164, 0.037966846466064456, 0.037945343017578126, 0.03807027053833008, 0.03799552154541016, 0.03783065414428711, 0.03770675277709961, 0.03779180908203125, 0.037851070404052736, 0.03778559875488281, 0.03766067123413086, 0.037722110748291016, 0.03759820938110352, 0.03784703826904297, 0.03785932922363281, 0.03774156951904297, 0.03778867340087891, 0.037886112213134766, 0.037889888763427734, 0.037730304718017575, 0.03767603302001953, 0.037725215911865236, 0.03766284942626953, 0.03771478271484375, 0.037806079864501956, 0.03784294509887695, 0.037835777282714846, 0.03786956787109375, 0.03780505752563477, 0.03779993438720703, 0.03785830307006836, 0.03791257476806641, 0.13102490234375, 0.03763097763061524, 0.03765862274169922, 0.03769139099121094, 0.037596160888671876, 0.03759206390380859, 0.03764019012451172, 0.03767398452758789, 0.03765657424926758, 0.037585918426513674, 0.037736446380615234, 0.0376360969543457, 0.0376258544921875, 0.03783679962158203, 0.03790848159790039, 0.037967872619628903, 0.03809382247924804, 0.037937152862548826, 0.037894142150878905, 0.03790335845947266, 0.037773311614990236, 0.03763206481933594, 0.037829566955566406, 0.037833728790283204, 0.03773952102661133, 0.03764633560180664, 0.03779174423217774, 0.03763302230834961, 0.03772723388671875, 0.03767398452758789, 0.03778662490844727, 0.03803647994995117, 0.03799244689941406, 0.0378419189453125, 0.03772415924072266, 0.03786649703979492, 0.03791155242919922, 0.03771289443969727, 0.03781222534179687, 0.03784294509887695, 0.03779379272460937, 0.03784396743774414, 0.03790848159790039, 0.03860790252685547, 0.03831087875366211, 0.03799244689941406, 0.03801295852661133, 0.03807126235961914, 0.03806617736816406, 0.03802214431762695, 0.03801190567016602, 0.03835289764404297, 0.038100990295410156, 0.03806211090087891, 0.03809584045410156, 0.03803238296508789, 0.03789209747314453, 0.03789209747314453, 0.037920768737792966, 0.03783679962158203, 0.03784806442260742, 0.037771263122558595, 0.03793817520141601, 0.131230712890625, 0.03832524871826172, 0.03775897598266602, 0.037495807647705076, 0.037852161407470705, 0.03772825622558594, 0.03775897598266602, 0.0375654411315918, 0.03751935958862305, 0.03749792098999023, 0.0375530891418457, 0.037551136016845704, 0.037537761688232425, 0.0374917106628418, 0.03764646530151367, 0.037677951812744144, 0.037678081512451174, 0.03758796691894531, 0.03765555191040039, 0.03763814544677734, 0.03764326477050781, 0.03762278366088867, 0.037628929138183595, 0.03775795364379883, 0.03766886520385742, 0.03762688064575195, 0.037817344665527344, 0.03767910385131836, 0.037784576416015625, 0.03767500686645508, 0.037773311614990236, 0.037800960540771485, 0.037754878997802735, 0.037713920593261716, 0.037720062255859374, 0.03771804809570312, 0.03773641586303711, 0.03775392150878906, 0.037785537719726564, 0.03778355026245117, 0.03775795364379883, 0.03767193603515625, 0.037797889709472655, 0.03782860946655273, 0.0380211181640625, 0.03776310348510742, 0.037800926208496094, 0.03787571334838867, 0.03783065414428711, 0.03779072189331055, 0.037773311614990236, 0.037746688842773435, 0.03783679962158203, 0.037789695739746096, 0.03792998504638672, 0.03805388641357422, 0.037894142150878905, 0.037804031372070314, 0.0377968635559082, 0.03778662490844727, 0.03780710220336914, 0.037759998321533206, 0.03782860946655273, 0.13123277282714843, 0.037561344146728515, 0.037644287109375, 0.037510143280029294, 0.03755110549926758, 0.03757875061035156, 0.037618686676025394, 0.03765964889526367, 0.03758182525634766, 0.03765760040283203, 0.037651454925537106, 0.03758694458007812, 0.03763507080078125, 0.03761971282958984, 0.037663745880126956, 0.037789695739746096, 0.0376627197265625, 0.03759820938110352, 0.037766143798828124, 0.037651454925537106, 0.03763916778564453, 0.037615615844726565, 0.03769449615478516, 0.03772003173828125, 0.03767910385131836, 0.037719039916992186, 0.03768320083618164, 0.03767091369628906, 0.037720062255859374, 0.03780915069580078, 0.037751808166503906, 0.037874687194824216, 0.03773747253417969, 0.03777228927612305, 0.03769753646850586, 0.03773747253417969, 0.037776382446289065, 0.037784576416015625, 0.03777433776855469, 0.03782553482055664, 0.03781222534179687, 0.03768729782104492, 0.03784703826904297, 0.03770163345336914, 0.03780505752563477, 0.037768192291259765, 0.037678081512451174, 0.03792588806152344, 0.03769241714477539, 0.03770265579223633, 0.0377149429321289, 0.037759998321533206, 0.03779993438720703, 0.03770880126953125, 0.03783475112915039, 0.03786444854736328, 0.037760032653808596, 0.03775279998779297, 0.037795841217041014, 0.03774259185791016, 0.037748737335205076, 0.03797196960449219, 0.037868545532226565, 0.13109861755371094, 0.03753065490722656, 0.03755721664428711, 0.037548065185546875, 0.037574623107910154, 0.03753779220581055, 0.03761663818359375, 0.03762073516845703, 0.03763302230834961, 0.03756032180786133, 0.037612545013427735, 0.037599231719970705, 0.03772415924072266, 0.03770470428466797, 0.03829452896118164, 0.03815321731567383, 0.03803443145751953, 0.03770163345336914, 0.03767603302001953, 0.037610496520996094, 0.03769036865234375, 0.037679134368896486, 0.03772412872314453, 0.03770470428466797, 0.03760947036743164, 0.03759820938110352, 0.03765657424926758, 0.0377262077331543, 0.03776102447509765, 0.03781631851196289, 0.037771263122558595, 0.037972991943359374, 0.037719039916992186, 0.037602302551269534, 0.03767500686645508, 0.037636161804199215, 0.03778963088989258, 0.03775590515136719, 0.037739585876464844, 0.037854145050048825, 0.037792766571044925, 0.037961727142333986, 0.03781119918823242, 0.037803009033203126, 0.038019073486328124, 0.03795257568359375, 0.03774969482421875, 0.0377968635559082, 0.03780918502807617, 0.03810915374755859, 0.037806079864501956, 0.038198272705078126, 0.03790643310546875, 0.037917697906494144, 0.037889022827148434, 0.03790233612060547, 0.037928958892822266, 0.037806079864501956, 0.03809996795654297, 0.03807436752319336, 0.03799961471557617, 0.03778355026245117, 0.03778047943115234, 0.1312522277832031, 0.03781631851196289, 0.037822463989257815, 0.03773747253417969, 0.037594112396240234, 0.03795251083374023, 0.03797401428222656, 0.03784908676147461, 0.03750604629516602, 0.03756748962402344, 0.037571582794189456, 0.03760537719726562, 0.037718017578125, 0.03818700790405274, 0.037852161407470705, 0.03772518539428711, 0.03769347381591797, 0.037645278930664064, 0.03775904083251953, 0.037680065155029294, 0.037678081512451174, 0.03769651031494141, 0.037700607299804685, 0.03778662490844727, 0.03765964889526367, 0.03769036865234375, 0.037736446380615234, 0.03767910385131836, 0.037763072967529294, 0.03757888031005859, 0.03765235137939453, 0.03766988754272461, 0.03764633560180664, 0.037661697387695314, 0.03766579055786133, 0.03764441680908203, 0.03771686553955078, 0.03761459350585938, 0.03760537719726562, 0.03777536010742188, 0.03774054336547852, 0.037628929138183595, 0.03768832015991211, 0.03769651031494141, 0.03780505752563477, 0.037739551544189454, 0.037778400421142576, 0.03844710540771484, 0.03802316665649414, 0.037768192291259765, 0.03777433776855469, 0.03780710220336914, 0.037838848114013675, 0.0378081283569336, 0.038063102722167966, 0.038150142669677735, 0.037967872619628903, 0.03789926528930664, 0.037943294525146484, 0.03790950393676758, 0.03787776184082031, 0.03839385604858398, 0.03825872039794922, 0.1311108856201172, 0.03753583908081055, 0.03749372863769531, 0.038142913818359374, 0.038316032409667966, 0.037574657440185545, 0.037510143280029294, 0.037623809814453124, 0.03755929565429687, 0.037533695220947266, 0.03765657424926758, 0.037580799102783204, 0.037628929138183595, 0.03793612670898437, 0.03755212783813477, 0.03767705535888672, 0.03776211166381836, 0.03764118576049805, 0.037589984893798827, 0.037766143798828124, 0.03774272155761719, 0.03773632049560547, 0.037819393157958986, 0.0377262077331543, 0.0377723503112793, 0.03767903900146485, 0.03774259185791016, 0.0379576301574707, 0.0377149429321289, 0.038089729309082034, 0.03786444854736328, 0.03781324768066406, 0.03799347305297852, 0.03793305587768555, 0.03778559875488281, 0.037804031372070314, 0.03782451248168945, 0.03785420989990235, 0.03794944000244141, 0.03804876708984375, 0.03810815811157227, 0.037966846466064456, 0.037868545532226565, 0.03801497650146484, 0.037803009033203126, 0.03798220825195313, 0.03796889495849609, 0.03800371170043945, 0.03781631851196289, 0.037833728790283204, 0.037835777282714846, 0.03779481506347656, 0.0377784309387207, 0.037795841217041014, 0.03784396743774414, 0.03788185501098633, 0.03807027053833008, 0.03784396743774414, 0.03783782577514649, 0.03779481506347656, 0.037819393157958986, 0.038063102722167966, 0.03779072189331055]",tokens/s,25.558627784742047,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-72B,Qwen/Qwen1.5-72B,cuda,0,42,,,True,,,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",qwen2,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,internlm/internlm-20b,internlm/internlm-20b,cuda,0,42,,,True,,,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",internlm,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-110B,Qwen/Qwen1.5-110B,cuda,0,42,,,True,,,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",qwen2,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Salesforce/codegen-16B-nl,Salesforce/codegen-16B-nl,cuda,0,42,,,True,,,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",codegen,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,m,m,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/m/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669472cd-6b7deacc2cc3bb367942f8aa;caf3e1fe-dcba-46d1-8b33-c6e5baf854a7) Repository Not Found for url: https://huggingface.co/m/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: m is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,openai-community/gpt2-large,openai-community/gpt2-large,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1235.070976,2661.810176,0.0,2015.363072,1692.280832,s,10,0.25306150436401365,0.025306150436401366,0.0007421607027119801,0.025393808364868167,0.025771244621276855,0.026436005878448483,0.02696781488418579,"[0.027100767135620117, 0.02437424087524414, 0.02555308723449707, 0.024535104751586913, 0.025623519897460936, 0.02534947204589844, 0.024610048294067384, 0.0249649600982666, 0.025512159347534178, 0.025438144683837892]",tokens/s,10116.117844291302,kWh,2.9173254803211436e-07,1.5983169879392437e-07,1.0356414320789727e-06,1.4872056789050115e-06,tokens/kWh,172134899.4501458,MB,1235.070976,2663.907328,0.0,2017.460224,1740.086272,s,10,14.8480205078125,1.48480205078125,0.01303825607973958,1.4858297729492187,1.4948051391601562,1.4999372375488282,1.5040429162597657,"[1.5050693359375, 1.453038330078125, 1.487899658203125, 1.4837598876953124, 1.492065673828125, 1.49277001953125, 1.4786517333984375, 1.4936646728515626, 1.479913330078125, 1.4811878662109375]",tokens/s,42.429898293076604,kWh,1.7174111308302577e-05,9.411372716284756e-06,3.5714610190322394e-05,6.230009421490973e-05,tokens/kWh,1011234.4257887618,,s,629,15.048426496505733,0.023924366449134718,0.003080390206878851,0.023669759750366212,0.023932928085327147,0.02443222999572754,0.047998196411132815,"[0.024634368896484377, 0.02470911979675293, 0.024321056365966796, 0.02456777572631836, 0.02448486328125, 0.024464384078979492, 0.024379392623901368, 0.024425472259521484, 0.024772607803344726, 0.025038848876953124, 0.024633344650268556, 0.024436735153198243, 0.02448793601989746, 0.02451046371459961, 0.024803327560424804, 0.024585216522216798, 0.024349695205688478, 0.023204864501953124, 0.02330931282043457, 0.023357440948486328, 0.024214527130126954, 0.024996864318847657, 0.024658943176269533, 0.02384390449523926, 0.02384889602661133, 0.023579647064208984, 0.02373017692565918, 0.024518655776977538, 0.023702527999877928, 0.023584768295288085, 0.0233123836517334, 0.023410879135131835, 0.023336767196655273, 0.023241727828979493, 0.023284767150878908, 0.0233973445892334, 0.023519359588623046, 0.023440256118774414, 0.02328678321838379, 0.023360511779785157, 0.023558143615722657, 0.0235100154876709, 0.02332262420654297, 0.02331056022644043, 0.02330601692199707, 0.0232857608795166, 0.02332876777648926, 0.023536640167236327, 0.02351513671875, 0.02327756881713867, 0.023661567687988282, 0.02364313507080078, 0.02374963188171387, 0.023604223251342774, 0.023605247497558594, 0.023736320495605468, 0.02368819236755371, 0.023540735244750977, 0.023626752853393555, 0.023563264846801758, 0.023568384170532225, 0.0235100154876709, 0.04764672088623047, 0.022730752944946288, 0.022577152252197266, 0.022806528091430665, 0.02264473533630371, 0.02267955207824707, 0.022761568069458008, 0.02269379234313965, 0.022734848022460938, 0.022595584869384764, 0.022740991592407226, 0.023685247421264648, 0.023610240936279298, 0.02369033622741699, 0.023752607345581055, 0.02372198486328125, 0.023570432662963867, 0.022691839218139647, 0.02287615966796875, 0.02349260711669922, 0.023669759750366212, 0.023619583129882812, 0.022754304885864256, 0.022768640518188478, 0.02269900894165039, 0.02272870445251465, 0.022787071228027343, 0.022731775283813475, 0.022813695907592774, 0.022786048889160155, 0.02271027183532715, 0.02270515251159668, 0.023856128692626953, 0.024498176574707032, 0.02291916847229004, 0.022874111175537108, 0.022781951904296875, 0.022689792633056642, 0.022740991592407226, 0.022759424209594727, 0.022726655960083008, 0.02269491195678711, 0.022746112823486327, 0.022675455093383787, 0.02273689651489258, 0.022819839477539062, 0.022791168212890626, 0.022808576583862306, 0.022803520202636717, 0.02280441665649414, 0.022674432754516603, 0.022708223342895507, 0.022785024642944338, 0.02275225639343262, 0.023344127655029297, 0.023752704620361328, 0.023582719802856447, 0.02372812843322754, 0.0238786563873291, 0.02391244888305664, 0.023625728607177734, 0.023803903579711915, 0.02370047950744629, 0.05022822570800781, 0.023706623077392578, 0.02370355224609375, 0.0237127685546875, 0.023653375625610353, 0.023640064239501952, 0.023378944396972655, 0.02342911911010742, 0.02369126319885254, 0.02368716812133789, 0.023605247497558594, 0.02390118408203125, 0.023810047149658203, 0.023587839126586914, 0.02285977554321289, 0.02293350410461426, 0.022777856826782225, 0.022789119720458984, 0.02270207977294922, 0.02268876838684082, 0.023682048797607422, 0.023710720062255858, 0.023618560791015625, 0.023817216873168946, 0.02370355224609375, 0.023710720062255858, 0.02369638442993164, 0.02349977684020996, 0.023576576232910155, 0.02370150375366211, 0.02371174430847168, 0.023677951812744142, 0.023867391586303712, 0.023555072784423828, 0.02369136047363281, 0.023707551956176756, 0.023657472610473632, 0.02369331169128418, 0.023606271743774415, 0.023764991760253908, 0.023598079681396485, 0.023580671310424805, 0.023604223251342774, 0.023827455520629884, 0.023801855087280274, 0.023832576751708984, 0.02372096061706543, 0.02374143981933594, 0.02364825630187988, 0.023677951812744142, 0.023677951812744142, 0.02369945526123047, 0.023774208068847655, 0.023757823944091795, 0.023990272521972656, 0.0237127685546875, 0.02347110366821289, 0.023625728607177734, 0.02391142463684082, 0.023631872177124022, 0.023757823944091795, 0.023653375625610353, 0.023657472610473632, 0.04794879913330078, 0.022806528091430665, 0.022734848022460938, 0.02281062316894531, 0.02285977554321289, 0.02266828727722168, 0.022803455352783202, 0.023018495559692383, 0.02311577606201172, 0.022808576583862306, 0.022793216705322264, 0.023404544830322265, 0.02386534309387207, 0.02374553680419922, 0.023762943267822266, 0.023832576751708984, 0.023675039291381837, 0.02374844741821289, 0.023577600479125976, 0.02368819236755371, 0.023586816787719726, 0.023736320495605468, 0.023839744567871093, 0.023657472610473632, 0.023644159317016602, 0.023517183303833008, 0.023589887619018556, 0.02369740867614746, 0.023544832229614256, 0.023618560791015625, 0.02364825630187988, 0.023682048797607422, 0.023932928085327147, 0.023659584045410156, 0.023931840896606445, 0.023632896423339843, 0.023616512298583983, 0.023677951812744142, 0.02369126319885254, 0.023646207809448243, 0.023553056716918944, 0.02371785545349121, 0.023669759750366212, 0.023673856735229492, 0.02369945526123047, 0.023662591934204103, 0.023591936111450194, 0.023603200912475586, 0.023760896682739258, 0.023735296249389647, 0.023566335678100587, 0.02372096061706543, 0.023641088485717773, 0.024048639297485352, 0.02393907165527344, 0.023631872177124022, 0.023550975799560548, 0.02374963188171387, 0.023706623077392578, 0.02364313507080078, 0.023686208724975587, 0.023714752197265626, 0.023809024810791016, 0.04916940689086914, 0.02274508857727051, 0.023628799438476563, 0.0242872314453125, 0.024780799865722656, 0.023969791412353517, 0.023837696075439452, 0.02367283248901367, 0.023762943267822266, 0.023747583389282227, 0.023762943267822266, 0.02370560073852539, 0.02374963188171387, 0.024173568725585938, 0.02511257553100586, 0.02427801513671875, 0.023814144134521483, 0.023794687271118165, 0.023739391326904297, 0.02366464042663574, 0.023751680374145507, 0.02367487907409668, 0.023744512557983398, 0.023628799438476563, 0.023603200912475586, 0.022798336029052735, 0.022757375717163086, 0.022788095474243163, 0.022755327224731444, 0.02288435173034668, 0.023758848190307616, 0.0239237117767334, 0.023742464065551756, 0.023744512557983398, 0.023760896682739258, 0.023791648864746093, 0.02381923294067383, 0.023774208068847655, 0.023658496856689453, 0.023677951812744142, 0.023464960098266603, 0.023554048538208007, 0.023779327392578126, 0.023760896682739258, 0.023784448623657226, 0.02369945526123047, 0.023868415832519533, 0.023619583129882812, 0.02369843292236328, 0.023631872177124022, 0.023940095901489256, 0.023738367080688477, 0.02367897605895996, 0.023599103927612306, 0.02373017692565918, 0.023636991500854493, 0.023932928085327147, 0.02373017692565918, 0.02367897605895996, 0.02352639961242676, 0.023426048278808592, 0.023403520584106444, 0.023395328521728515, 0.04964863967895508, 0.02371174430847168, 0.023750656127929686, 0.02369024085998535, 0.02389708709716797, 0.023662591934204103, 0.023581695556640626, 0.023662591934204103, 0.02346700859069824, 0.023764991760253908, 0.023838720321655273, 0.023803903579711915, 0.02391551971435547, 0.0239237117767334, 0.023789567947387694, 0.023828479766845705, 0.02371174430847168, 0.02370150375366211, 0.023718912124633788, 0.023723007202148438, 0.02408345603942871, 0.02370047950744629, 0.023634944915771484, 0.0234833927154541, 0.02389504051208496, 0.02370867156982422, 0.0237260799407959, 0.02369843292236328, 0.02374143981933594, 0.02391347122192383, 0.02411520004272461, 0.023821311950683592, 0.023632896423339843, 0.023750656127929686, 0.023621631622314454, 0.02370047950744629, 0.02372403144836426, 0.023838720321655273, 0.0237127685546875, 0.023588863372802735, 0.022726655960083008, 0.022777856826782225, 0.022726655960083008, 0.022739967346191405, 0.02387046432495117, 0.023910432815551757, 0.023631839752197265, 0.02328678321838379, 0.02364316749572754, 0.023849952697753907, 0.023743488311767577, 0.023754751205444336, 0.023768064498901367, 0.023640127182006837, 0.02381100845336914, 0.02384486389160156, 0.02372403144836426, 0.02366771125793457, 0.023734272003173826, 0.023816192626953125, 0.023767040252685546, 0.02369024085998535, 0.02414182472229004, 0.04801740646362305, 0.022760448455810548, 0.022785055160522462, 0.022788063049316406, 0.022709247589111328, 0.022706239700317384, 0.022719423294067384, 0.022683647155761717, 0.022798336029052735, 0.023326719284057617, 0.023783424377441405, 0.023678016662597657, 0.023688127517700195, 0.023573535919189453, 0.023675872802734376, 0.02391756820678711, 0.02367487907409668, 0.023586816787719726, 0.02371283149719238, 0.023553983688354492, 0.023640064239501952, 0.023621631622314454, 0.023513151168823243, 0.023623615264892577, 0.02352025604248047, 0.02343337631225586, 0.023426015853881835, 0.023392127990722655, 0.023508991241455078, 0.023517183303833008, 0.023593984603881835, 0.023478271484375, 0.02332467269897461, 0.023370880126953125, 0.023438304901123048, 0.02328771209716797, 0.02334617614746094, 0.023504896163940428, 0.023305280685424805, 0.02277779197692871, 0.022666240692138673, 0.023719968795776366, 0.023714784622192384, 0.02366873550415039, 0.02372915267944336, 0.023855104446411132, 0.02369740867614746, 0.024025087356567384, 0.023621631622314454, 0.02330931282043457, 0.02373948860168457, 0.023903135299682618, 0.02408550453186035, 0.02371686363220215, 0.023639039993286134, 0.02388275146484375, 0.023724063873291016, 0.023593952178955078, 0.023836671829223634, 0.02368511962890625, 0.02364313507080078, 0.023635967254638672, 0.02348236846923828, 0.05030912017822266, 0.025421823501586914, 0.023992319107055664, 0.023975936889648438, 0.024814592361450196, 0.02410700798034668, 0.02395955276489258, 0.023770111083984375, 0.02410495948791504, 0.02368819236755371, 0.023805952072143553, 0.02371583938598633, 0.02372198486328125, 0.023819263458251954, 0.023706623077392578, 0.02393087959289551, 0.023795711517333985, 0.023686143875122072, 0.02367487907409668, 0.02394316864013672, 0.023628799438476563, 0.023670783996582033, 0.023632896423339843, 0.02369228744506836, 0.02366054344177246, 0.023747583389282227, 0.023573503494262696, 0.023564287185668945, 0.023608320236206053, 0.023820287704467775, 0.023793664932250977, 0.023771135330200196, 0.023626752853393555, 0.023695423126220704, 0.023719871520996094, 0.023642112731933593, 0.02390732765197754, 0.023631872177124022, 0.02368000030517578, 0.02364313507080078, 0.02366361618041992, 0.023601152420043944, 0.023809024810791016, 0.023670783996582033, 0.02305843162536621, 0.023029760360717775, 0.022742048263549804, 0.02270716857910156, 0.022815744400024415, 0.02262118339538574, 0.02249625587463379, 0.0236646728515625, 0.023679967880249023, 0.023638111114501953, 0.023803808212280272, 0.02374963188171387, 0.023834623336791993, 0.023763967514038087, 0.023875583648681642, 0.023627775192260742, 0.023619583129882812, 0.02369536018371582, 0.023833599090576172, 0.05006335830688476, 0.023819263458251954, 0.023705631256103515, 0.023655391693115233, 0.02372198486328125, 0.023753759384155273, 0.02370249557495117, 0.024025087356567384, 0.024844287872314453, 0.0243507194519043, 0.024945663452148437, 0.023983104705810547, 0.023778303146362305, 0.023908351898193358, 0.023848960876464844, 0.02386636734008789, 0.023830528259277343, 0.023810047149658203, 0.023627775192260742, 0.023786527633666992, 0.023635936737060548, 0.023602176666259765, 0.023802879333496094, 0.023812095642089845, 0.023665664672851562, 0.023797760009765623, 0.023677951812744142, 0.023723007202148438, 0.023835647583007814, 0.024023040771484375, 0.02365951919555664, 0.023411712646484374, 0.023428096771240234, 0.024057855606079103, 0.023422975540161133, 0.02346598434448242, 0.023052288055419923, 0.023332895278930663, 0.023443424224853515, 0.0235100154876709, 0.023407615661621094, 0.02366464042663574, 0.02345062446594238, 0.023544960021972657, 0.02350067138671875, 0.022533119201660155, 0.02245734405517578, 0.022458368301391602, 0.02247372817993164, 0.022622207641601562, 0.023047296524047852, 0.023242624282836914, 0.0232857608795166, 0.023377920150756838, 0.02332159996032715, 0.022417407989501953, 0.022571008682250978, 0.022560768127441407, 0.022412288665771486, 0.022609920501708985, 0.02263654327392578, 0.02288128089904785, 0.023601152420043944, 0.04965990447998047, 0.023601152420043944, 0.023687231063842774, 0.02376697540283203, 0.02371379280090332, 0.023586816787719726, 0.023609344482421874, 0.023677951812744142, 0.02387660789489746, 0.023738367080688477, 0.023646207809448243, 0.02363497543334961, 0.023629791259765626, 0.02369638442993164, 0.023751680374145507, 0.023940095901489256, 0.02372812843322754, 0.023669759750366212, 0.02370047950744629, 0.02370560073852539, 0.02367283248901367, 0.023605247497558594, 0.023568384170532225, 0.023658496856689453, 0.02349158477783203, 0.02369740867614746, 0.023811071395874024, 0.023699487686157227, 0.023707616806030275, 0.023825408935546875, 0.02325606346130371, 0.023560192108154295, 0.023557119369506836, 0.02364825630187988, 0.023589887619018556, 0.023793664932250977, 0.023825408935546875, 0.02353459167480469, 0.02370560073852539, 0.02269081687927246, 0.022814720153808594, 0.022656000137329102, 0.022683647155761717, 0.022633472442626954, 0.022675455093383787, 0.022740991592407226, 0.022794240951538085, 0.02263859176635742, 0.022790143966674805, 0.02277689552307129, 0.02301535987854004, 0.023788543701171876, 0.02364313507080078, 0.02373222351074219, 0.023677951812744142, 0.023756799697875978, 0.023770111083984375, 0.02373324775695801, 0.023537664413452147, 0.023608320236206053, 0.02369843292236328, 0.023793664932250977, 0.023780351638793946]",tokens/s,41.798390027426095,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,huggyllama/llama-65b,huggyllama/llama-65b,cuda,0,42,,,True,,,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",llama,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,l,l,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/l/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66947441-700c2fa25d9a47d060d6c891;67fb0dff-fbf7-4f10-b011-a8c17af2499c) Repository Not Found for url: https://huggingface.co/l/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: l is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,B,B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/B/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669324f5-0d0c31f15a68fef911124033;95c6c75c-3d5a-4703-84be-81cfff09131c) Repository Not Found for url: https://huggingface.co/B/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: B is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,huggyllama/llama-30b,huggyllama/llama-30b,cuda,0,42,,,True,,,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",llama,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,meta-llama/Llama-2-13b-hf,meta-llama/Llama-2-13b-hf,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ self.load_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 250, in load_model_with_no_weights self.pretrained_model = self.automodel_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3710, in from_pretrained model = cls(config, *model_args, **model_kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 1101, in __init__ self.model = LlamaModel(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 884, in __init__ [LlamaDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 884, in [LlamaDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 679, in __init__ self.mlp = LlamaMLP(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 192, in __init__ self.gate_proj = nn.Linear(self.hidden_size, self.intermediate_size, bias=config.mlp_bias) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/linear.py"", line 98, in __init__ self.weight = Parameter(torch.empty((out_features, in_features), **factory_kwargs)) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_device.py"", line 78, in __torch_function__ return func(*args, **kwargs) torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 136.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-7B,,cuda,0,42,,,,,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,217063f5c507ed7cc255df7e1f64c4333a0b4dfe,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,,MB,1318.805504,17102.798848,0.0,16456.351744,16359.853056,s,10,1.0681468887329102,0.10681468887329101,0.0002945769989060828,0.1067933464050293,0.10693519973754882,0.10728094291687011,0.10755753746032715,"[0.1076266860961914, 0.10685836791992187, 0.1065077133178711, 0.10670118713378907, 0.10657039642333985, 0.10680992126464844, 0.10684003448486327, 0.10677677154541015, 0.10662172698974609, 0.10683408355712891]",tokens/s,2396.6741157078136,kWh,1.2634387024046402e-06,6.922380217552114e-07,6.126260811170152e-06,8.081937535330004e-06,tokens/kWh,31675572.70529522,MB,1318.805504,17102.798848,0.0,16456.351744,16416.242176,s,10,23.64958349609375,2.364958349609375,0.003937901517589722,2.363560546875,2.3683912841796873,2.3719263549804688,2.3747544116210935,"[2.37546142578125, 2.362560546875, 2.36191064453125, 2.363364013671875, 2.367605712890625, 2.361707275390625, 2.363757080078125, 2.3654501953125, 2.365556884765625, 2.362209716796875]",tokens/s,26.638946943994107,kWh,2.784717529272357e-05,1.5261355628182618e-05,0.00011678417085042906,0.00015989270177133526,tokens/kWh,394014.23143188335,,s,629,24.57144319915771,0.039064297613923236,0.012748992410659883,0.03751424026489258,0.03768427658081054,0.03780403289794922,0.1447934802246094,"[0.038095870971679685, 0.03825155258178711, 0.03809481430053711, 0.0381736946105957, 0.038158336639404294, 0.03813683319091797, 0.03821670532226563, 0.038354942321777344, 0.037966846466064456, 0.03798732757568359, 0.03798015975952149, 0.038345729827880856, 0.038324222564697266, 0.03758899307250976, 0.037379070281982424, 0.037443584442138675, 0.037397502899169925, 0.03744153594970703, 0.03743648147583008, 0.037551040649414065, 0.03747225570678711, 0.03751935958862305, 0.03739852905273437, 0.03750502395629883, 0.037424129486083986, 0.03749990463256836, 0.03742313766479492, 0.0376923828125, 0.03776204681396484, 0.037766143798828124, 0.037585918426513674, 0.03765760040283203, 0.03743231964111328, 0.0375552978515625, 0.03743939208984375, 0.037541889190673826, 0.037485633850097656, 0.03756844711303711, 0.037634048461914066, 0.0376360969543457, 0.0374835205078125, 0.037571582794189456, 0.03747123336791992, 0.037561344146728515, 0.03755212783813477, 0.03765555191040039, 0.037612545013427735, 0.03768326568603515, 0.03752544021606445, 0.03762278366088867, 0.03750912094116211, 0.03759206390380859, 0.037556224822998044, 0.037664768218994144, 0.0379576301574707, 0.03766579055786133, 0.0375470085144043, 0.03761663818359375, 0.03750912094116211, 0.037703678131103514, 0.0375551986694336, 0.03763302230834961, 0.14494720458984375, 0.037305343627929685, 0.03734527969360352, 0.03726438522338867, 0.03735039901733399, 0.03739648056030274, 0.037375999450683595, 0.037367809295654295, 0.037424129486083986, 0.03736576080322265, 0.03743641662597656, 0.03754598236083984, 0.03752959823608398, 0.03738726425170898, 0.0374323501586914, 0.03739542388916016, 0.03747020721435547, 0.037384193420410154, 0.037410911560058595, 0.03741891098022461, 0.03746201705932617, 0.0373831672668457, 0.03747532653808594, 0.0374015998840332, 0.037466110229492186, 0.03746099090576172, 0.03749785614013672, 0.03743334579467773, 0.03754086303710937, 0.03743948745727539, 0.03751628875732422, 0.03746918487548828, 0.03750297546386719, 0.03745280075073242, 0.03750502395629883, 0.037424129486083986, 0.037577728271484374, 0.037479423522949216, 0.03751628875732422, 0.03751935958862305, 0.03753676986694336, 0.0374835205078125, 0.03752447891235351, 0.03750707244873047, 0.03759718322753906, 0.03747430419921875, 0.03759001541137695, 0.0375203857421875, 0.03757670211791992, 0.037541889190673826, 0.03758182525634766, 0.03755011367797852, 0.03762684631347656, 0.03752345657348633, 0.037634048461914066, 0.03758489608764649, 0.03765555191040039, 0.03760947036743164, 0.03763814544677734, 0.03760025787353516, 0.037769248962402344, 0.037604320526123045, 0.037644287109375, 0.14463282775878905, 0.037379070281982424, 0.037353473663330077, 0.03727155303955078, 0.03738214492797851, 0.03729715347290039, 0.03735039901733399, 0.037372928619384765, 0.037408767700195314, 0.037305343627929685, 0.037400577545166014, 0.03733196640014649, 0.037401630401611326, 0.037364734649658206, 0.0374425277709961, 0.03733401489257813, 0.0374200325012207, 0.03735756683349609, 0.037450782775878905, 0.03737091064453125, 0.03742822265625, 0.037382080078125, 0.03748761749267578, 0.03747225570678711, 0.037498878479003905, 0.03740979385375977, 0.03748454284667969, 0.03738726425170898, 0.03754291152954101, 0.03748044967651367, 0.03753676986694336, 0.037425151824951174, 0.0374835205078125, 0.03745075225830078, 0.037510143280029294, 0.03745792007446289, 0.0374835205078125, 0.03743436813354492, 0.037533695220947266, 0.03742924880981445, 0.0374835205078125, 0.03747430419921875, 0.03753267288208008, 0.037730304718017575, 0.037953536987304685, 0.03752140808105469, 0.03761356735229492, 0.03755929565429687, 0.037601280212402347, 0.03749785614013672, 0.03758694458007812, 0.037515262603759765, 0.03754291152954101, 0.037525505065917966, 0.03761459350585938, 0.03750502395629883, 0.03761459350585938, 0.03753676986694336, 0.03768832015991211, 0.03752447891235351, 0.03761356735229492, 0.0375623664855957, 0.03770880126953125, 0.1448079376220703, 0.037343231201171875, 0.03733708953857422, 0.037303295135498044, 0.037370880126953124, 0.037321727752685545, 0.037369857788085936, 0.03735551834106445, 0.03737702560424805, 0.037364734649658206, 0.03744255828857422, 0.03732992172241211, 0.03738828659057617, 0.037395454406738284, 0.03743027114868164, 0.037362686157226564, 0.037465087890625, 0.03734630584716797, 0.03745280075073242, 0.037354496002197264, 0.03742310333251953, 0.03736678314208984, 0.03743948745727539, 0.037375999450683595, 0.03750092697143555, 0.03741388702392578, 0.03745075225830078, 0.03741593551635742, 0.03750502395629883, 0.03741798400878906, 0.03754291152954101, 0.03745177459716797, 0.037485633850097656, 0.03755513763427734, 0.037797889709472655, 0.03751321411132812, 0.03752959823608398, 0.037477375030517575, 0.037539840698242184, 0.03748761749267578, 0.03757670211791992, 0.03751116943359375, 0.037566463470458986, 0.03752243041992188, 0.037610496520996094, 0.03751424026489258, 0.03761151885986328, 0.03751424026489258, 0.03762073516845703, 0.03749683380126953, 0.03765657424926758, 0.0378419189453125, 0.03779072189331055, 0.03754393768310547, 0.03767603302001953, 0.0376484489440918, 0.03766368103027344, 0.03761971282958984, 0.03770675277709961, 0.03770675277709961, 0.037703678131103514, 0.03761663818359375, 0.037754878997802735, 0.14479055786132813, 0.03738723373413086, 0.03736166381835938, 0.03741798400878906, 0.03748659133911133, 0.037438465118408204, 0.037370880126953124, 0.03737497711181641, 0.03752140808105469, 0.037365825653076175, 0.037401535034179687, 0.03757056045532227, 0.03783168029785156, 0.03757670211791992, 0.03743436813354492, 0.037473281860351565, 0.03784806442260742, 0.037663745880126956, 0.03759212875366211, 0.03739744186401367, 0.03747430419921875, 0.037599231719970705, 0.03748969650268555, 0.03757360076904297, 0.03764223861694336, 0.037623809814453124, 0.037612545013427735, 0.03752140808105469, 0.03767295837402344, 0.03762790298461914, 0.037666912078857424, 0.037580703735351564, 0.03761356735229492, 0.03750707244873047, 0.03750912094116211, 0.03747225570678711, 0.03760639953613281, 0.037574657440185545, 0.03763302230834961, 0.03752755355834961, 0.037610496520996094, 0.037550079345703126, 0.03766886520385742, 0.03763097763061524, 0.03761459350585938, 0.03766681671142578, 0.03771596908569336, 0.037570625305175784, 0.037676990509033205, 0.03757056045532227, 0.03759206390380859, 0.03754393768310547, 0.037598270416259766, 0.03759199905395508, 0.03758899307250976, 0.03756851196289063, 0.037644287109375, 0.03763916778564453, 0.037661697387695314, 0.03754598236083984, 0.03763814544677734, 0.03757363128662109, 0.03784499359130859, 0.14500352478027342, 0.037343231201171875, 0.037340160369873046, 0.03729715347290039, 0.037394432067871096, 0.037303359985351565, 0.03741484832763672, 0.03736678314208984, 0.03734630584716797, 0.037378047943115236, 0.03741593551635742, 0.03733196640014649, 0.03742105484008789, 0.037354496002197264, 0.0374128646850586, 0.037365791320800784, 0.037456863403320315, 0.03732992172241211, 0.037427200317382815, 0.037389312744140625, 0.03742211151123047, 0.03739337539672852, 0.037459968566894535, 0.037359615325927735, 0.0374917106628418, 0.03741183853149414, 0.0375203857421875, 0.03741491317749023, 0.037563392639160156, 0.037438465118408204, 0.03751839828491211, 0.03743020629882812, 0.03750502395629883, 0.03744255828857422, 0.03749273681640625, 0.037459968566894535, 0.037510143280029294, 0.037495807647705076, 0.03765555191040039, 0.03747635269165039, 0.03757056045532227, 0.037493759155273435, 0.037602302551269534, 0.03748863983154297, 0.03753881454467774, 0.03748966217041016, 0.03759308624267578, 0.03751424026489258, 0.03760332870483398, 0.03750707244873047, 0.03762176132202148, 0.03750604629516602, 0.03757875061035156, 0.037539840698242184, 0.037601280212402347, 0.03756854248046875, 0.037622753143310546, 0.03753676986694336, 0.03764121627807617, 0.037548030853271484, 0.03767295837402344, 0.03758694458007812, 0.03764019012451172, 0.14501785278320312, 0.03732992172241211, 0.0373309440612793, 0.037343231201171875, 0.037482494354248046, 0.037324798583984374, 0.03737702560424805, 0.037369857788085936, 0.037424129486083986, 0.03738828659057617, 0.03738623809814453, 0.0373831672668457, 0.03748147201538086, 0.03740262222290039, 0.037501953125, 0.0373831672668457, 0.03745894241333008, 0.03745280075073242, 0.03758182525634766, 0.03767295837402344, 0.03750912094116211, 0.03741593551635742, 0.03750400161743164, 0.037459968566894535, 0.03748044967651367, 0.0374835205078125, 0.0374917106628418, 0.03743129730224609, 0.03756851196289063, 0.03746918487548828, 0.03751935958862305, 0.03753267288208008, 0.03767295837402344, 0.03746406555175781, 0.03748659133911133, 0.03746099090576172, 0.03758899307250976, 0.03743948745727539, 0.037550079345703126, 0.03750809478759766, 0.03751424026489258, 0.03754707336425781, 0.03754387283325195, 0.037498878479003905, 0.037596160888671876, 0.03750400161743164, 0.03758796691894531, 0.037539840698242184, 0.037574657440185545, 0.0375470085144043, 0.03758694458007812, 0.03761459350585938, 0.0376360969543457, 0.037526527404785154, 0.03759513473510742, 0.03760537719726562, 0.03776716613769531, 0.03758694458007812, 0.037628929138183595, 0.03757875061035156, 0.0376627197265625, 0.037607425689697264, 0.03765350341796875, 0.14492160034179688, 0.03733401489257813, 0.037359615325927735, 0.037303359985351565, 0.03747116851806641, 0.037351425170898435, 0.03737395095825195, 0.03734527969360352, 0.03736783981323242, 0.03732988739013672, 0.037407745361328126, 0.03735968017578125, 0.037417919158935546, 0.03739347076416016, 0.03740972900390625, 0.03734527969360352, 0.03746713638305664, 0.03735551834106445, 0.03744768142700195, 0.03741491317749023, 0.037449726104736326, 0.0373831672668457, 0.03743641662597656, 0.03739136123657227, 0.03746918487548828, 0.03740979385375977, 0.03746201705932617, 0.03745180892944336, 0.037537761688232425, 0.03741388702392578, 0.03755417633056641, 0.03745587158203125, 0.037515262603759765, 0.037410816192626956, 0.03749273681640625, 0.03742524719238281, 0.03769232177734375, 0.03791360092163086, 0.03763097763061524, 0.037645313262939455, 0.03770265579223633, 0.03759206390380859, 0.03773235321044922, 0.03763916778564453, 0.037718017578125, 0.03768115234375, 0.03758694458007812, 0.03753574371337891, 0.03767193603515625, 0.038105152130126954, 0.03770566558837891, 0.03755215835571289, 0.037645278930664064, 0.03762483215332031, 0.03791872024536133, 0.037781505584716796, 0.03769139099121094, 0.037610496520996094, 0.037763072967529294, 0.037667839050292966, 0.03772723388671875, 0.03760639953613281, 0.03769036865234375, 0.14479461669921875, 0.03740979385375977, 0.037531646728515625, 0.03756032180786133, 0.03778355026245117, 0.037533695220947266, 0.037512191772460936, 0.037477375030517575, 0.0376258544921875, 0.03739136123657227, 0.03742105484008789, 0.03744153594970703, 0.03750707244873047, 0.037394432067871096, 0.03746918487548828, 0.03748761749267578, 0.037749759674072264, 0.03750912094116211, 0.03751628875732422, 0.0378081283569336, 0.037618686676025394, 0.03741593551635742, 0.03748966217041016, 0.037400577545166014, 0.037459968566894535, 0.03745587158203125, 0.03751935958862305, 0.03745177459716797, 0.03753779220581055, 0.03746406555175781, 0.03750912094116211, 0.03746815872192383, 0.03759308624267578, 0.03749068832397461, 0.03748966217041016, 0.03746713638305664, 0.03755427169799805, 0.03747113418579102, 0.037525505065917966, 0.03748767852783203, 0.03752544021606445, 0.03752447891235351, 0.037550079345703126, 0.037544960021972655, 0.037569534301757815, 0.03752447891235351, 0.03762688064575195, 0.037541889190673826, 0.037703678131103514, 0.03753779220581055, 0.03757567977905273, 0.03753676986694336, 0.037602302551269534, 0.037574657440185545, 0.03757567977905273, 0.037544960021972655, 0.03774566268920899, 0.037569534301757815, 0.03763916778564453, 0.037553150177001955, 0.03763302230834961, 0.0375880012512207, 0.03763504028320312, 0.14493907165527345, 0.037376960754394534, 0.03738521575927734, 0.03729100799560547, 0.037364734649658206, 0.0373309440612793, 0.037369857788085936, 0.03739648056030274, 0.03737395095825195, 0.037364734649658206, 0.03746815872192383, 0.03733299255371094, 0.0374466552734375, 0.03735756683349609, 0.03740979385375977, 0.03738009643554688, 0.03743948745727539, 0.03733708953857422, 0.03743641662597656, 0.03735859298706055, 0.037425151824951174, 0.037375999450683595, 0.03744768142700195, 0.037372928619384765, 0.03748454284667969, 0.03742822265625, 0.03752553558349609, 0.0374425277709961, 0.03751321411132812, 0.03744255828857422, 0.037531646728515625, 0.03743948745727539, 0.03752345657348633, 0.03743027114868164, 0.03746406555175781, 0.03750400161743164, 0.03753881454467774, 0.03749990463256836, 0.037526527404785154, 0.03747532653808594, 0.037563392639160156, 0.037550079345703126, 0.037571582794189456, 0.037533695220947266, 0.037580799102783204, 0.03751321411132812, 0.03763097763061524, 0.03750297546386719, 0.03764121627807617, 0.037493759155273435, 0.037596160888671876, 0.03752246475219727, 0.037606369018554686, 0.037526527404785154, 0.037594112396240234, 0.03760537719726562, 0.037644287109375, 0.037544960021972655, 0.0376545295715332, 0.037556224822998044, 0.03766681671142578, 0.037571582794189456, 0.03776921463012695]",tokens/s,25.598821969950933,,,main,False,False,,,,,,,,,,,,,,,,,,,,,,, float16-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,M,M,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/M/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66932288-7a097bbc73bff43d4af6c164;e291b65e-a4f7-4fd3-bd42-a03dcf8086f7) Repository Not Found for url: https://huggingface.co/M/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: M is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,stabilityai/stablelm-base-alpha-3b,stabilityai/stablelm-base-alpha-3b,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1293.131776,8550.612992,0.0,7904.165888,7767.63392,s,10,0.46857788467407224,0.04685778846740723,0.0011178978426431324,0.04650673484802246,0.04689347229003906,0.04855148773193359,0.04987790008544922,"[0.05020950317382813, 0.04650390243530273, 0.046509567260742186, 0.0465250244140625, 0.046521278381347654, 0.04651248168945313, 0.04648134231567383, 0.046456382751464846, 0.04639014434814453, 0.046468257904052734]",tokens/s,5463.339358793371,kWh,5.514102296311726e-07,3.021466363980137e-07,3.0148693369581254e-06,3.868426202987312e-06,tokens/kWh,66176782.6415583,MB,1293.131776,8550.612992,0.0,7904.165888,7864.441856,s,10,12.6907177734375,1.26907177734375,0.004631085100790426,1.2691265869140627,1.2744065673828124,1.2745195068359376,1.2746098583984375,"[1.2743814697265625, 1.2705679931640625, 1.2734173583984374, 1.2746324462890626, 1.27389208984375, 1.26605517578125, 1.263653564453125, 1.2622054443359374, 1.2676851806640625, 1.26422705078125]",tokens/s,49.64258218070463,kWh,1.4877533785209486e-05,8.152695692503562e-06,6.527865752004417e-05,8.830888699775721e-05,tokens/kWh,713404.9826899078,,s,629,13.088088054656987,0.020807771152077877,0.00556743031168173,0.020135936737060548,0.020280319213867186,0.0203321346282959,0.06691065887451171,"[0.021113855361938477, 0.020331520080566406, 0.020179967880249023, 0.02020351982116699, 0.020149248123168945, 0.02017791938781738, 0.02026393508911133, 0.0202106876373291, 0.020188159942626953, 0.020208639144897463, 0.020168703079223634, 0.02024550437927246, 0.020205568313598633, 0.020121631622314454, 0.020191200256347658, 0.02018611145019531, 0.020165632247924805, 0.020057088851928712, 0.02006937599182129, 0.020016128540039063, 0.019979263305664064, 0.02001203155517578, 0.02006220817565918, 0.02017791938781738, 0.02016972732543945, 0.0200949764251709, 0.020297727584838866, 0.020242431640625, 0.020175872802734376, 0.020205568313598633, 0.02024140739440918, 0.020172800064086914, 0.020222976684570314, 0.020206592559814454, 0.02005299186706543, 0.02010419273376465, 0.02008576011657715, 0.020183040618896485, 0.02022502326965332, 0.020075519561767577, 0.02021785545349121, 0.020343839645385744, 0.020415456771850585, 0.02027724838256836, 0.020247552871704103, 0.020115455627441405, 0.02023628807067871, 0.020171775817871093, 0.02020249557495117, 0.020280319213867186, 0.020178943634033202, 0.02019430351257324, 0.020230144500732423, 0.020238336563110353, 0.020353023529052734, 0.02024038314819336, 0.020328447341918944, 0.020393983840942383, 0.020287488937377928, 0.020207616806030275, 0.020230144500732423, 0.020215808868408205, 0.0669573745727539, 0.020020160675048828, 0.020155391693115234, 0.020147199630737304, 0.020082687377929686, 0.019975168228149414, 0.02004377555847168, 0.020174848556518556, 0.020171775817871093, 0.019971071243286134, 0.020063232421875, 0.020075519561767577, 0.02012057685852051, 0.02008678436279297, 0.020148223876953125, 0.02006425666809082, 0.020100095748901366, 0.020131839752197265, 0.020149248123168945, 0.020171775817871093, 0.020223039627075196, 0.02013587188720703, 0.02023526382446289, 0.020315135955810547, 0.020316160202026368, 0.02021990394592285, 0.020254720687866212, 0.02022604751586914, 0.02016979217529297, 0.020314048767089844, 0.020155391693115234, 0.02027212715148926, 0.020179967880249023, 0.020147199630737304, 0.02007859230041504, 0.02004172706604004, 0.020083711624145507, 0.020174848556518556, 0.02023423957824707, 0.020214784622192384, 0.020082687377929686, 0.0200949764251709, 0.020198400497436524, 0.020262912750244142, 0.02019327926635742, 0.020254720687866212, 0.02031718444824219, 0.020256767272949217, 0.02025574493408203, 0.020143104553222657, 0.020140031814575195, 0.020152320861816408, 0.02020351982116699, 0.020205568313598633, 0.020110336303710938, 0.02024345588684082, 0.02020351982116699, 0.020242431640625, 0.02022604751586914, 0.02021887969970703, 0.020135936737060548, 0.020083744049072264, 0.02008367919921875, 0.0669521942138672, 0.02003763198852539, 0.02011955261230469, 0.020157440185546875, 0.02011955261230469, 0.02011238479614258, 0.020171775817871093, 0.020212736129760742, 0.020170751571655272, 0.020151296615600587, 0.02023423957824707, 0.020175872802734376, 0.02026393508911133, 0.020222976684570314, 0.020117504119873047, 0.02021990394592285, 0.020182016372680665, 0.020183040618896485, 0.020149248123168945, 0.02017791938781738, 0.020174848556518556, 0.020147199630737304, 0.020182016372680665, 0.020060159683227538, 0.020001792907714845, 0.020009983062744142, 0.02004172706604004, 0.02003046417236328, 0.02002739143371582, 0.02006425666809082, 0.019955711364746095, 0.020125696182250977, 0.02012774467468262, 0.020192256927490236, 0.020212736129760742, 0.020195327758789062, 0.02021990394592285, 0.02018611145019531, 0.02025574493408203, 0.02021785545349121, 0.020239360809326173, 0.020222976684570314, 0.020212736129760742, 0.020370431900024414, 0.020759552001953126, 0.020329471588134765, 0.02024550437927246, 0.020346879959106445, 0.02023628807067871, 0.020250656127929687, 0.020222944259643556, 0.020239360809326173, 0.02027622413635254, 0.020280319213867186, 0.02079539108276367, 0.020271104812622072, 0.020388864517211915, 0.02025984001159668, 0.02028441619873047, 0.02030182456970215, 0.020289535522460937, 0.020289535522460937, 0.020280319213867186, 0.06716722869873047, 0.020185087203979494, 0.020155391693115234, 0.020197376251220703, 0.02020147132873535, 0.020141056060791016, 0.020106239318847655, 0.02021990394592285, 0.02022604751586914, 0.02016972732543945, 0.020152320861816408, 0.0202106876373291, 0.02024038314819336, 0.02022604751586914, 0.02014419174194336, 0.02020345687866211, 0.02022604751586914, 0.02020966339111328, 0.02017791938781738, 0.020157440185546875, 0.020132863998413086, 0.020159488677978517, 0.020238336563110353, 0.020220928192138672, 0.020121599197387697, 0.02021683120727539, 0.020180992126464844, 0.02016972732543945, 0.02027827262878418, 0.020164608001708984, 0.020172800064086914, 0.020327423095703127, 0.020246528625488282, 0.02027724838256836, 0.02020147132873535, 0.020385791778564453, 0.02043391990661621, 0.020330495834350586, 0.020289535522460937, 0.02021171188354492, 0.020237312316894532, 0.02025984001159668, 0.02022809600830078, 0.020397056579589845, 0.020246528625488282, 0.020246528625488282, 0.02021990394592285, 0.020319295883178712, 0.020238271713256838, 0.020257791519165038, 0.020197376251220703, 0.020281343460083007, 0.02021683120727539, 0.020214784622192384, 0.02010316848754883, 0.02021887969970703, 0.020206592559814454, 0.020304895401000975, 0.020220928192138672, 0.020256767272949217, 0.020299776077270508, 0.020273151397705077, 0.02023628807067871, 0.06724505615234375, 0.020100095748901366, 0.020090879440307616, 0.02008883285522461, 0.020164608001708984, 0.020205568313598633, 0.020182016372680665, 0.020231168746948244, 0.020185087203979494, 0.020166656494140626, 0.020237312316894532, 0.020183040618896485, 0.02011136054992676, 0.020039680480957032, 0.020165632247924805, 0.020220928192138672, 0.02016972732543945, 0.020187135696411132, 0.020167680740356447, 0.020068416595458983, 0.020182975769042967, 0.02023334312438965, 0.02014092826843262, 0.02023017692565918, 0.020222944259643556, 0.020189184188842774, 0.020257791519165038, 0.020190208435058594, 0.020170751571655272, 0.020332544326782227, 0.02021785545349121, 0.020230144500732423, 0.02025369644165039, 0.02016972732543945, 0.020155391693115234, 0.02020966339111328, 0.020157503128051757, 0.020080575942993163, 0.02030080032348633, 0.02027724838256836, 0.02020147132873535, 0.02022809600830078, 0.020242431640625, 0.020238336563110353, 0.020196352005004883, 0.020264959335327147, 0.020279327392578126, 0.020361183166503906, 0.02021171188354492, 0.020328447341918944, 0.02023321533203125, 0.020242431640625, 0.02019327926635742, 0.02023628807067871, 0.020344831466674804, 0.02025984001159668, 0.0202608642578125, 0.020222976684570314, 0.020316160202026368, 0.020238336563110353, 0.02029465675354004, 0.020350976943969725, 0.02030598449707031, 0.06709241485595703, 0.020178943634033202, 0.020142143249511718, 0.020107200622558594, 0.020221952438354493, 0.020242431640625, 0.02011955261230469, 0.020189184188842774, 0.020187135696411132, 0.02021683120727539, 0.020182016372680665, 0.020198400497436524, 0.020142080307006836, 0.02012774467468262, 0.02022604751586914, 0.020191232681274415, 0.02007347106933594, 0.02004991912841797, 0.020133888244628906, 0.020154367446899413, 0.02009702491760254, 0.0200949764251709, 0.020172800064086914, 0.020163583755493163, 0.020109312057495117, 0.02006220817565918, 0.02007040023803711, 0.02008780860900879, 0.019957759857177734, 0.01999667167663574, 0.019974143981933593, 0.019950592041015625, 0.02000592041015625, 0.01997001647949219, 0.020067327499389647, 0.02004377555847168, 0.02004172706604004, 0.020048896789550782, 0.02012774467468262, 0.020125696182250977, 0.020091903686523437, 0.020048896789550782, 0.020074495315551756, 0.02006937599182129, 0.02008780860900879, 0.020063232421875, 0.020122623443603514, 0.020100095748901366, 0.02003046417236328, 0.020018239974975587, 0.020023231506347657, 0.02003049659729004, 0.020056032180786134, 0.02004991912841797, 0.02003865623474121, 0.02011238479614258, 0.020048896789550782, 0.02004172706604004, 0.020076543807983398, 0.02003865623474121, 0.02003455924987793, 0.020015104293823242, 0.02003046417236328, 0.06711296081542968, 0.020015104293823242, 0.019978239059448243, 0.019964927673339843, 0.019964927673339843, 0.019970048904418947, 0.019940351486206053, 0.020000768661499024, 0.020011007308959963, 0.019965951919555663, 0.020000768661499024, 0.019983360290527344, 0.019999744415283204, 0.019992576599121094, 0.02000486373901367, 0.02004172706604004, 0.020002815246582033, 0.020032512664794923, 0.0200263671875, 0.020016128540039063, 0.02003558349609375, 0.020048896789550782, 0.019987455368041994, 0.020142080307006836, 0.020031488418579102, 0.020040704727172853, 0.02001408004760742, 0.019999744415283204, 0.020056095123291016, 0.020070367813110352, 0.020025344848632814, 0.02008576011657715, 0.02004377555847168, 0.019998720169067383, 0.019985408782958985, 0.020051967620849608, 0.02002841567993164, 0.02004991912841797, 0.02007961654663086, 0.02009702491760254, 0.020057088851928712, 0.020063232421875, 0.020032512664794923, 0.02001919937133789, 0.02004684829711914, 0.02003763198852539, 0.02003660774230957, 0.020129791259765627, 0.02012057685852051, 0.0200263671875, 0.02004787254333496, 0.020518911361694335, 0.020199424743652345, 0.020122623443603514, 0.020152320861816408, 0.020117504119873047, 0.02007244873046875, 0.02006937599182129, 0.020050943374633787, 0.020075519561767577, 0.020074495315551756, 0.020092927932739257, 0.020092927932739257, 0.06687232208251953, 0.02022707176208496, 0.0202608642578125, 0.019979263305664064, 0.02001408004760742, 0.01997113609313965, 0.01998739242553711, 0.02003865623474121, 0.02004377555847168, 0.01998847961425781, 0.019956735610961913, 0.01998847961425781, 0.020001792907714845, 0.019983360290527344, 0.020011007308959963, 0.02003865623474121, 0.020002815246582033, 0.019968000411987305, 0.019971071243286134, 0.019966976165771484, 0.02001513671875, 0.019995616912841796, 0.02002022361755371, 0.02007040023803711, 0.020011007308959963, 0.02002124786376953, 0.02001408004760742, 0.019987455368041994, 0.020008960723876954, 0.0200447998046875, 0.020025344848632814, 0.020060159683227538, 0.02002124786376953, 0.019999744415283204, 0.02002022361755371, 0.02001408004760742, 0.02002943992614746, 0.020057088851928712, 0.02007244873046875, 0.020113407135009767, 0.02004377555847168, 0.0200581111907959, 0.02004377555847168, 0.02005299186706543, 0.02003660774230957, 0.02005504035949707, 0.020048896789550782, 0.020075519561767577, 0.020032512664794923, 0.020016128540039063, 0.02000486373901367, 0.020009983062744142, 0.02001203155517578, 0.02001203155517578, 0.019994623184204103, 0.020067327499389647, 0.02001817512512207, 0.02004582405090332, 0.02003558349609375, 0.02006118392944336, 0.020025407791137696, 0.020036544799804688, 0.0200263671875, 0.06672895812988282, 0.019876863479614256, 0.019902463912963866, 0.019902463912963866, 0.019942399978637695, 0.019954687118530275, 0.01998847961425781, 0.019999744415283204, 0.019982336044311523, 0.019947519302368166, 0.019983360290527344, 0.019975168228149414, 0.019991552352905274, 0.019982336044311523, 0.02008780860900879, 0.020528127670288086, 0.020303871154785155, 0.020297727584838866, 0.020160512924194338, 0.020199424743652345, 0.020157440185546875, 0.02003046417236328, 0.02003865623474121, 0.020083711624145507, 0.02000383949279785, 0.020023296356201172, 0.0200898551940918, 0.02003763198852539, 0.020083711624145507, 0.020098047256469728, 0.0200581111907959, 0.02008166313171387, 0.02004172706604004, 0.02002943992614746, 0.02007961654663086, 0.02004275131225586, 0.02006630325317383, 0.02003763198852539, 0.02008780860900879, 0.020113407135009767, 0.02005606460571289, 0.02002227210998535, 0.020172800064086914, 0.020247552871704103, 0.020117504119873047, 0.020463615417480468, 0.020438079833984376, 0.020274112701416016, 0.020161535263061522, 0.02021887969970703, 0.020215808868408205, 0.020230144500732423, 0.020306943893432617, 0.020205568313598633, 0.020338687896728515, 0.02025574493408203, 0.02029465675354004, 0.020196352005004883, 0.020237312316894532, 0.02023526382446289, 0.0202076473236084, 0.02014204788208008, 0.0200447998046875, 0.06692556762695312, 0.019962879180908204, 0.020015104293823242, 0.02008576011657715, 0.02009702491760254, 0.020106239318847655, 0.020115455627441405, 0.020153343200683595, 0.020118528366088868, 0.02007142448425293, 0.0200949764251709, 0.020144128799438478, 0.020192256927490236, 0.020105215072631837, 0.020140031814575195, 0.020176895141601564, 0.020148223876953125, 0.020167680740356447, 0.019959808349609375, 0.019975168228149414, 0.020068351745605468, 0.020154367446899413, 0.020206592559814454, 0.02011955261230469, 0.019958784103393554, 0.019957759857177734, 0.020007936477661133, 0.019971071243286134, 0.02000383949279785, 0.020008960723876954, 0.020007936477661133, 0.02002124786376953, 0.019987455368041994, 0.019997695922851562, 0.019997695922851562, 0.019992576599121094, 0.020002815246582033, 0.02001408004760742, 0.019992576599121094, 0.02004172706604004, 0.019992576599121094, 0.020001792907714845, 0.02002124786376953, 0.02000383949279785, 0.02004684829711914, 0.02006220817565918, 0.0200581111907959, 0.020106239318847655, 0.02004377555847168, 0.020075519561767577, 0.020040704727172853, 0.02004582405090332, 0.02005401611328125, 0.020036640167236327, 0.020010976791381835, 0.0200949764251709, 0.02002739143371582, 0.02003865623474121, 0.02005606460571289, 0.020051967620849608, 0.02011955261230469, 0.020079679489135742, 0.020412351608276365]",tokens/s,48.05896761797764,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,google/recurrentgemma-9b,google/recurrentgemma-9b,cuda,0,42,,,True,,,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",recurrent_gemma,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,mistralai/Mixtral-8x22B-v0.1,mistralai/Mixtral-8x22B-v0.1,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/base.py"", line 43, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 54, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1021, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 403 Client Error: Forbidden for url: https://huggingface.co/mistralai/Mixtral-8x22B-v0.1/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 398, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 321, in hf_raise_for_status raise GatedRepoError(message, response) from e huggingface_hub.utils._errors.GatedRepoError: 403 Client Error. (Request ID: Root=1-6647506f-219d52e36118313616e8198c;cf04188b-5e28-4957-9db5-3049676e2072) Cannot access gated repo for url https://huggingface.co/mistralai/Mixtral-8x22B-v0.1/resolve/main/config.json. Access to model mistralai/Mixtral-8x22B-v0.1 is restricted and you are not in the authorized list. Visit https://huggingface.co/mistralai/Mixtral-8x22B-v0.1 to ask for access. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 87, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/base.py"", line 56, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 37, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 62, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 928, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 631, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 686, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 416, in cached_file raise EnvironmentError( OSError: You are trying to access a gated repo. Make sure to have access to it at https://huggingface.co/mistralai/Mixtral-8x22B-v0.1. 403 Client Error. (Request ID: Root=1-6647506f-219d52e36118313616e8198c;cf04188b-5e28-4957-9db5-3049676e2072) Cannot access gated repo for url https://huggingface.co/mistralai/Mixtral-8x22B-v0.1/resolve/main/config.json. Access to model mistralai/Mixtral-8x22B-v0.1 is restricted and you are not in the authorized list. Visit https://huggingface.co/mistralai/Mixtral-8x22B-v0.1 to ask for access. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,01-ai/Yi-6B,01-ai/Yi-6B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1266.307072,13032.226816,0.0,12385.779712,12315.692032,s,10,0.8885439453125001,0.08885439453124999,0.0006202610278086367,0.08866265487670899,0.0889212989807129,0.08981546592712403,0.09053079948425294,"[0.09070963287353516, 0.08858099365234375, 0.08872259521484376, 0.08863533020019532, 0.08864991760253907, 0.08860150146484375, 0.0886941146850586, 0.08857638549804688, 0.08869808197021484, 0.08867539215087891]",tokens/s,2881.118051060098,kWh,1.0489974929168162e-06,5.748036887305602e-07,5.251574997716825e-06,6.875376179364201e-06,tokens/kWh,37234326.28579074,MB,1266.307072,13032.226816,0.0,12385.779712,12315.694592,s,10,22.21704541015625,2.2217045410156246,0.043435932916983826,2.207030517578125,2.232852392578125,2.2919032470703122,2.3391439306640627,"[2.3509541015625, 2.20105224609375, 2.206557861328125, 2.21972998046875, 2.212437744140625, 2.207503173828125, 2.199526123046875, 2.210241943359375, 2.204247314453125, 2.204794921875]",tokens/s,28.356605856870747,kWh,2.5993747196600738e-05,1.4242780702781795e-05,0.00010122828796388322,0.00014146481586326575,tokens/kWh,445340.4163823554,,s,629,22.978625537872283,0.03653199608564756,0.010585250860440178,0.034925567626953126,0.036967627716064454,0.03716300735473633,0.12401004455566406,"[0.038537216186523435, 0.03852185440063476, 0.03830176162719726, 0.03805184173583984, 0.038576065063476564, 0.038263809204101565, 0.03832320022583008, 0.03794432067871094, 0.03926220703125, 0.03826892852783203, 0.037713920593261716, 0.037956607818603515, 0.03810508728027344, 0.03768217468261719, 0.037407745361328126, 0.036992000579833983, 0.036944896697998046, 0.03740364837646484, 0.03707187271118164, 0.037190654754638675, 0.03701964950561523, 0.037054462432861326, 0.03714252853393555, 0.03696844863891602, 0.0369541130065918, 0.03696230316162109, 0.03702067184448242, 0.03700940704345703, 0.03697971343994141, 0.03704627227783203, 0.037212158203125, 0.037082111358642575, 0.037000190734863284, 0.036972545623779295, 0.0370063362121582, 0.037269504547119144, 0.0371671028137207, 0.03705036926269531, 0.03715686416625977, 0.03701657485961914, 0.036994049072265625, 0.036967422485351564, 0.03684044647216797, 0.03697049713134765, 0.03706777572631836, 0.037048320770263675, 0.037085182189941404, 0.03711897659301758, 0.03705241775512695, 0.037526527404785154, 0.03720191955566406, 0.036923393249511716, 0.03695718383789062, 0.03696640014648438, 0.03709235382080078, 0.03706367874145508, 0.037087230682373046, 0.03700940704345703, 0.037013504028320314, 0.036907009124755856, 0.03695513534545898, 0.036907009124755856, 0.12403097534179687, 0.03483340835571289, 0.034825214385986326, 0.034859073638916015, 0.0349552001953125, 0.03472895812988281, 0.03492454528808594, 0.03486412811279297, 0.034767871856689454, 0.03483135986328125, 0.03478015899658203, 0.034893856048583985, 0.034881504058837894, 0.03489894485473633, 0.03480678558349609, 0.03504947280883789, 0.034988033294677735, 0.03506687927246094, 0.03484672164916992, 0.03481804656982422, 0.03479654312133789, 0.035800064086914066, 0.035046398162841795, 0.03488665771484375, 0.03490816116333008, 0.034947071075439456, 0.03486624145507813, 0.03490399932861328, 0.03486003112792969, 0.034869312286376956, 0.03496646499633789, 0.034993152618408206, 0.03488870239257812, 0.034885631561279294, 0.034904064178466795, 0.03516211318969727, 0.034994174957275394, 0.034950145721435545, 0.03495219039916992, 0.035110912322998046, 0.035019775390625, 0.03486310577392578, 0.03488153457641602, 0.03482422256469726, 0.03481391906738281, 0.034971710205078124, 0.03482419204711914, 0.03493881607055664, 0.03492966461181641, 0.0347770881652832, 0.0351907844543457, 0.03469311904907227, 0.03477401733398437, 0.034723838806152346, 0.03479449462890625, 0.03497574234008789, 0.03479654312133789, 0.03486003112792969, 0.03527679824829102, 0.03514572906494141, 0.03554816055297851, 0.034869247436523435, 0.03484774398803711, 0.12433203125, 0.03490508651733398, 0.03528806304931641, 0.035710975646972655, 0.03588915252685547, 0.03530547332763672, 0.035125247955322264, 0.03522252655029297, 0.035185695648193356, 0.03487740707397461, 0.03479654312133789, 0.03490508651733398, 0.03483443069458008, 0.03487648010253906, 0.03483027267456055, 0.03491020965576172, 0.03492659378051758, 0.03487948989868164, 0.034887680053710936, 0.03488051223754883, 0.03476684951782227, 0.03505561447143555, 0.03652096176147461, 0.03535769653320313, 0.034974720001220705, 0.03492147064208984, 0.034920448303222655, 0.035639297485351565, 0.0349409294128418, 0.03491328048706055, 0.03489382553100586, 0.03502182388305664, 0.03499622344970703, 0.035020801544189455, 0.0350013427734375, 0.03504230499267578, 0.03498291015625, 0.03489894485473633, 0.03488358306884766, 0.03502182388305664, 0.03496857452392578, 0.03483241653442383, 0.034871295928955076, 0.035012577056884764, 0.03492659378051758, 0.03478732681274414, 0.03473408126831055, 0.03541401672363281, 0.034819072723388675, 0.03524505615234375, 0.034874366760253905, 0.034844703674316406, 0.03490409469604492, 0.03474016189575195, 0.0347955207824707, 0.03508019256591797, 0.035113983154296875, 0.03466239929199219, 0.03486822509765625, 0.034767871856689454, 0.03488358306884766, 0.03491839981079101, 0.03489894485473633, 0.12437510681152343, 0.034840511322021483, 0.03483750534057617, 0.03485696029663086, 0.03711283111572266, 0.035773441314697264, 0.03504844665527344, 0.036634624481201174, 0.03502592086791992, 0.03493580627441406, 0.0350904312133789, 0.03487948989868164, 0.03483750534057617, 0.035318782806396484, 0.03501875305175781, 0.03474943923950195, 0.03470950317382813, 0.03487948989868164, 0.035773441314697264, 0.03496857452392578, 0.03506175994873047, 0.03487027359008789, 0.03518054580688477, 0.03505049514770508, 0.03561164855957031, 0.034871295928955076, 0.034902015686035154, 0.03506687927246094, 0.03588710403442383, 0.034988033294677735, 0.03496550369262695, 0.03502284622192383, 0.035955711364746096, 0.035113983154296875, 0.03499622344970703, 0.03491635131835937, 0.03537612915039062, 0.03544985580444336, 0.03501055908203125, 0.034977790832519534, 0.035487743377685545, 0.035062782287597655, 0.03489382553100586, 0.03486617660522461, 0.03491020965576172, 0.03524915313720703, 0.035471359252929685, 0.03544678497314453, 0.034887680053710936, 0.03492147064208984, 0.035520511627197264, 0.035827713012695314, 0.03569356918334961, 0.035683326721191407, 0.03628851318359375, 0.03545804977416992, 0.035318782806396484, 0.03535564804077149, 0.034825214385986326, 0.03538227081298828, 0.035361793518066405, 0.03488153457641602, 0.03467468643188477, 0.12391731262207031, 0.03482112121582031, 0.03484467315673828, 0.035975166320800785, 0.0347955207824707, 0.035209217071533204, 0.03498905563354492, 0.03540377426147461, 0.034977790832519534, 0.03475046539306641, 0.03483340835571289, 0.034797569274902344, 0.035691520690917966, 0.03514572906494141, 0.03497062301635742, 0.035299327850341795, 0.03518873596191406, 0.03479654312133789, 0.03476684951782227, 0.03530854415893555, 0.0351836166381836, 0.03502489471435547, 0.03480575942993164, 0.03489279937744141, 0.03486515045166016, 0.034885631561279294, 0.03555430221557617, 0.034969600677490234, 0.0347852783203125, 0.0348221435546875, 0.035454975128173825, 0.03591372680664062, 0.0359024658203125, 0.03485184097290039, 0.03496243286132812, 0.03588608169555664, 0.03508531188964844, 0.034872318267822264, 0.034920448303222655, 0.035329025268554685, 0.03598950576782227, 0.0348590087890625, 0.03484774398803711, 0.03480473709106445, 0.035490814208984374, 0.03474739074707031, 0.034772991180419925, 0.034890750885009765, 0.03521843338012695, 0.03557580947875977, 0.034974720001220705, 0.03476070404052734, 0.03478732681274414, 0.03540991973876953, 0.03529420852661133, 0.034776065826416014, 0.03472281646728516, 0.034784255981445314, 0.0357212142944336, 0.03487955093383789, 0.035621822357177736, 0.035471359252929685, 0.03523993682861328, 0.12413337707519531, 0.03501363372802734, 0.034812992095947265, 0.034807743072509764, 0.0346879997253418, 0.03476172637939453, 0.03485081481933594, 0.03578060913085938, 0.03499212646484375, 0.03487744140625, 0.03483852767944336, 0.035945472717285154, 0.03496038436889649, 0.034993152618408206, 0.03475971221923828, 0.03540278244018555, 0.03521120071411133, 0.03486822509765625, 0.03480473709106445, 0.03476172637939453, 0.03579289627075195, 0.03489177703857422, 0.03482419204711914, 0.03549593734741211, 0.035323902130126955, 0.035402751922607424, 0.03498495864868164, 0.034966529846191405, 0.03481702423095703, 0.035527679443359376, 0.034871295928955076, 0.035004417419433595, 0.03482112121582031, 0.035194881439208986, 0.036037631988525394, 0.03590655899047852, 0.03587996673583985, 0.035229663848876956, 0.035019840240478516, 0.03542214584350586, 0.035280895233154294, 0.034754558563232424, 0.03484262466430664, 0.03482419204711914, 0.03520307159423828, 0.034825214385986326, 0.03483340835571289, 0.034802688598632815, 0.034945022583007815, 0.03488256072998047, 0.03474431991577148, 0.034799617767333986, 0.034816001892089846, 0.03486617660522461, 0.03484979248046875, 0.03487539291381836, 0.03473612976074219, 0.03475763320922851, 0.034694145202636716, 0.03458355331420898, 0.034904064178466795, 0.03481497573852539, 0.034840576171875, 0.1242060775756836, 0.034862144470214844, 0.035068862915039065, 0.03485184097290039, 0.03521331024169922, 0.03486723327636719, 0.034835422515869144, 0.0348221435546875, 0.034769920349121096, 0.034850879669189455, 0.03492755126953125, 0.03484364700317383, 0.03483647918701172, 0.034786304473876956, 0.0348671989440918, 0.03498188781738281, 0.03482419204711914, 0.03485388946533203, 0.034830337524414064, 0.03485388946533203, 0.0347935676574707, 0.03487017440795898, 0.03487334442138672, 0.03499827194213867, 0.03483347320556641, 0.034893760681152346, 0.03491839981079101, 0.03500543975830078, 0.034890750885009765, 0.034866207122802736, 0.03489174270629883, 0.03518771362304687, 0.03516108703613281, 0.034936832427978515, 0.03492966461181641, 0.03487744140625, 0.03478015899658203, 0.03505152130126953, 0.034854911804199216, 0.034871295928955076, 0.03465523147583008, 0.03480678558349609, 0.03486412811279297, 0.03478937530517578, 0.03544271850585937, 0.03552252960205078, 0.03511603164672852, 0.034825214385986326, 0.03483135986328125, 0.03483955383300781, 0.034961406707763674, 0.034920448303222655, 0.03473408126831055, 0.03476790237426758, 0.03478524780273438, 0.03496550369262695, 0.03481087875366211, 0.034767871856689454, 0.034825214385986326, 0.034776065826416014, 0.035085342407226563, 0.034896865844726566, 0.03483647918701172, 0.12395622253417969, 0.0347474250793457, 0.03472995376586914, 0.03483852767944336, 0.03473100662231445, 0.03475049591064453, 0.034796512603759766, 0.034931713104248044, 0.0348590087890625, 0.03481497573852539, 0.034813953399658204, 0.03480575942993164, 0.034740222930908206, 0.034837535858154293, 0.034761695861816405, 0.035026943206787106, 0.03483340835571289, 0.03488870239257812, 0.03491328048706055, 0.034813953399658204, 0.0349306869506836, 0.03486617660522461, 0.03484467315673828, 0.03511296081542969, 0.03495423889160156, 0.034909183502197266, 0.03487948989868164, 0.03558911895751953, 0.03877785491943359, 0.036631553649902344, 0.03653529739379883, 0.03526348876953125, 0.03504127883911133, 0.03501055908203125, 0.035004417419433595, 0.03488153457641602, 0.03491839981079101, 0.03491328048706055, 0.0349224967956543, 0.03508019256591797, 0.03498086547851562, 0.03501772689819336, 0.03482419204711914, 0.034945022583007815, 0.0358922233581543, 0.03568025588989258, 0.035141632080078124, 0.035081214904785156, 0.03486006546020508, 0.03507913589477539, 0.03491020965576172, 0.036506622314453126, 0.03487334442138672, 0.03476889419555664, 0.034871295928955076, 0.034841598510742186, 0.034781185150146485, 0.0347852783203125, 0.035369983673095705, 0.03473920059204102, 0.034772991180419925, 0.03469823837280273, 0.034781185150146485, 0.12419692993164062, 0.034896831512451175, 0.034699264526367186, 0.03480166244506836, 0.03482726287841797, 0.034852863311767575, 0.034746368408203124, 0.034974720001220705, 0.03474431991577148, 0.03488153457641602, 0.034840576171875, 0.03567308807373047, 0.036550655364990234, 0.03508428955078125, 0.03498291015625, 0.0348671989440918, 0.03477612686157226, 0.03497055816650391, 0.0347770881652832, 0.0348487663269043, 0.03482316970825195, 0.0347955207824707, 0.035004478454589844, 0.035009471893310544, 0.03481497573852539, 0.034816001892089846, 0.03490508651733398, 0.034854911804199216, 0.03480780792236328, 0.03488460922241211, 0.03482726287841797, 0.03491430282592774, 0.03484364700317383, 0.0349409294128418, 0.03496352005004883, 0.035684288024902346, 0.03547443389892578, 0.034829376220703125, 0.03502995300292969, 0.03500543975830078, 0.034898975372314456, 0.03452617645263672, 0.03474534225463867, 0.0348487663269043, 0.03480575942993164, 0.03510784149169922, 0.03633663940429688, 0.035748863220214845, 0.034928638458251955, 0.03475558471679688, 0.03484467315673828, 0.034754558563232424, 0.03485184097290039, 0.034797569274902344, 0.034993152618408206, 0.03545600128173828, 0.03511296081542969, 0.03489279937744141, 0.03493580627441406, 0.03481292724609375, 0.03490412902832031, 0.03480160140991211, 0.034885631561279294, 0.12413542175292969, 0.03501055908203125, 0.03504127883911133, 0.034871295928955076, 0.03487641525268555, 0.03481497573852539, 0.035667999267578125, 0.035222496032714844, 0.034920448303222655, 0.0356577262878418, 0.0353546257019043, 0.03512115097045899, 0.03505459213256836, 0.03483340835571289, 0.034925567626953126, 0.035361793518066405, 0.03491123199462891, 0.035046398162841795, 0.034907135009765625, 0.034928638458251955, 0.034933761596679686, 0.03483238220214844, 0.034816001892089846, 0.034971649169921876, 0.03496243286132812, 0.034941951751708986, 0.03489484786987305, 0.034915328979492184, 0.03482828903198242, 0.03521638488769531, 0.0352174072265625, 0.03698995208740234, 0.0348671989440918, 0.035023872375488284, 0.03494297790527344, 0.03516928100585937, 0.03506995010375977, 0.03494508743286133, 0.03486918258666992, 0.0349409294128418, 0.03482112121582031, 0.0347852783203125, 0.03488256072998047, 0.03482009506225586, 0.03486822509765625, 0.034786304473876956, 0.03500543975830078, 0.035074047088623043, 0.034699264526367186, 0.03512115097045899, 0.03485388946533203, 0.03484672164916992, 0.03477401733398437, 0.03481087875366211, 0.03477913665771484, 0.034925567626953126, 0.03479859161376953, 0.03465625762939453, 0.034931743621826175, 0.034884574890136716, 0.03488665771484375, 0.03479244613647461, 0.034699264526367186]",tokens/s,27.373264730882664,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,/,/,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 373, in cached_file raise EnvironmentError( OSError: / does not appear to have a file named config.json. Checkout 'https://huggingface.co///tree/None' for available files. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,TencentARC/Mistral_Pro_8B_v0.1,,cuda,0,42,,,,,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,d35829e539df8480b726c647eeabf91e41eae047,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,,MB,1279.22176,19665.518592,0.0,19019.071488,18747.268096,s,10,1.0421181182861328,0.10421181182861328,0.0005720682456904527,0.10404446411132813,0.10449944915771484,0.105170157623291,0.10570672439575195,"[0.10584086608886718, 0.10381343841552734, 0.10395977783203125, 0.1042738265991211, 0.10411148834228516, 0.10435040283203124, 0.10380352020263672, 0.1039774398803711, 0.10414643096923829, 0.10384092712402344]",tokens/s,2456.53535341097,kWh,1.232232272418009e-06,6.749982484279827e-07,7.832465756708236e-06,9.739696277554228e-06,tokens/kWh,26284187.17634644,MB,1279.22176,19667.615744,0.0,19021.16864,18774.94784,s,10,29.178298095703123,2.9178298095703123,0.0009953691693975557,2.9175145263671873,2.9189455078125,2.9194814208984377,2.9199101513671875,"[2.920017333984375, 2.91865087890625, 2.917649658203125, 2.91737939453125, 2.9170830078125, 2.918826416015625, 2.917940673828125, 2.9172119140625, 2.91686669921875, 2.916672119140625]",tokens/s,21.591389529767522,kWh,3.439379668587612e-05,1.8849631595207943e-05,0.00014520374000549156,0.0001984471682865756,tokens/kWh,317464.8474148158,,s,629,30.067661838531546,0.047802324067617634,0.0124230431858849,0.04630527877807617,0.04641484832763672,0.046491441345214844,0.1506380194091797,"[0.046459903717041014, 0.04619878387451172, 0.046238719940185545, 0.04616396713256836, 0.0462295036315918, 0.04618239974975586, 0.04621516799926758, 0.04618239974975586, 0.046219264984130856, 0.046311424255371096, 0.04660940933227539, 0.04642201614379883, 0.0462479362487793, 0.046222335815429685, 0.04622643280029297, 0.04620492935180664, 0.04625920104980469, 0.04626124954223633, 0.0463125114440918, 0.046222335815429685, 0.04620588684082031, 0.04623974227905273, 0.04621625518798828, 0.04623660659790039, 0.046235649108886716, 0.04627558517456055, 0.04625612640380859, 0.04628684616088867, 0.04630220794677734, 0.04639231872558594, 0.0463185920715332, 0.04626739120483398, 0.046318622589111326, 0.04637488174438477, 0.04634931182861328, 0.04634726333618164, 0.046292991638183595, 0.04630732727050781, 0.046309375762939455, 0.046360576629638675, 0.04629401779174805, 0.04644966506958008, 0.046620670318603515, 0.046344192504882815, 0.04636876678466797, 0.04640256118774414, 0.04637388610839844, 0.04650086212158203, 0.046699520111083984, 0.04658892822265625, 0.046413822174072264, 0.046357505798339846, 0.04636159896850586, 0.0463902702331543, 0.04638003158569336, 0.046443519592285154, 0.04641177749633789, 0.046445568084716796, 0.0463636474609375, 0.046389278411865235, 0.046414817810058594, 0.04647731018066406, 0.15066009521484375, 0.04614451217651367, 0.046235649108886716, 0.046203903198242184, 0.04612812805175781, 0.04623155212402344, 0.04616499328613281, 0.046222335815429685, 0.046205951690673826, 0.046273536682128906, 0.04621209716796875, 0.04619366455078125, 0.046220287322998044, 0.04632883071899414, 0.046238719940185545, 0.046203903198242184, 0.04629401779174805, 0.04627046585083008, 0.046238719940185545, 0.046258174896240234, 0.04619366455078125, 0.04630527877807617, 0.046317569732666014, 0.04636262512207031, 0.046317569732666014, 0.04627763366699219, 0.046235649108886716, 0.04624076843261719, 0.046306304931640625, 0.04627046585083008, 0.046271488189697264, 0.046296062469482424, 0.046271488189697264, 0.0463185920715332, 0.04635647964477539, 0.04642303848266602, 0.046296062469482424, 0.04631449508666992, 0.04631961441040039, 0.046358528137207033, 0.04639542388916015, 0.04632163238525391, 0.04639539337158203, 0.04635136032104492, 0.04651212692260742, 0.04677017593383789, 0.046461952209472655, 0.04636671829223633, 0.04635238265991211, 0.046322689056396485, 0.04635033416748047, 0.046325759887695314, 0.046461952209472655, 0.04633497619628906, 0.046394367218017575, 0.04651318359375, 0.04638819122314453, 0.04642105484008789, 0.04637177658081055, 0.046355457305908204, 0.046511104583740234, 0.04642303848266602, 0.046440448760986325, 0.1507379150390625, 0.046167041778564455, 0.046192638397216795, 0.04616806411743164, 0.04619980621337891, 0.046183425903320315, 0.04616396713256836, 0.04619878387451172, 0.04620185470581055, 0.04621619033813477, 0.04616396713256836, 0.04625305557250976, 0.046186496734619144, 0.04631654357910156, 0.04616908645629883, 0.04621209716796875, 0.0461844482421875, 0.046279678344726564, 0.04631760025024414, 0.04623465728759766, 0.04623660659790039, 0.046263294219970705, 0.04622438430786133, 0.04628275299072265, 0.046222335815429685, 0.04623769760131836, 0.04626124954223633, 0.04628582382202148, 0.046295040130615236, 0.046415870666503906, 0.04628070449829102, 0.046281726837158206, 0.046241790771484374, 0.04629401779174805, 0.04633599853515625, 0.046360576629638675, 0.04633497619628906, 0.046357505798339846, 0.046292991638183595, 0.04636262512207031, 0.046317569732666014, 0.04629094314575195, 0.0463001594543457, 0.046312446594238284, 0.04632985687255859, 0.04634726333618164, 0.04637497711181641, 0.04631135940551758, 0.046399486541748046, 0.046344192504882815, 0.04647423934936523, 0.046860286712646484, 0.04655411148071289, 0.046322689056396485, 0.04635955047607422, 0.04637696075439453, 0.04641484832763672, 0.046355457305908204, 0.04645273590087891, 0.0463370246887207, 0.04640153503417969, 0.04641177749633789, 0.046355457305908204, 0.15130726623535157, 0.0463001594543457, 0.04615475082397461, 0.04631039810180664, 0.046238719940185545, 0.046311424255371096, 0.04613324737548828, 0.04624998474121094, 0.046325759887695314, 0.04617830276489258, 0.04619366455078125, 0.046170112609863284, 0.046192638397216795, 0.0462213134765625, 0.04620800018310547, 0.046203968048095706, 0.04623660659790039, 0.04632064056396484, 0.04618854522705078, 0.04625100708007813, 0.046186496734619144, 0.04624895858764649, 0.04628684616088867, 0.04632166290283203, 0.046266368865966793, 0.04626739120483398, 0.04626124954223633, 0.04622643280029297, 0.04628275299072265, 0.04627046585083008, 0.04625920104980469, 0.04629708862304688, 0.04623155212402344, 0.046301185607910154, 0.046296062469482424, 0.04636159896850586, 0.04630220794677734, 0.04631039810180664, 0.046325759887695314, 0.04635238265991211, 0.04629401779174805, 0.04630220794677734, 0.046317569732666014, 0.046276607513427735, 0.04639846420288086, 0.04636569595336914, 0.0463370246887207, 0.04637081527709961, 0.046415870666503906, 0.04638515090942383, 0.04652236938476562, 0.046358528137207033, 0.04639744186401367, 0.04629913711547851, 0.04634316635131836, 0.04638617706298828, 0.0464005126953125, 0.046357505798339846, 0.046399486541748046, 0.04632473754882813, 0.04637696075439453, 0.04635340881347656, 0.04636979293823242, 0.15098265075683595, 0.046153728485107424, 0.04617728042602539, 0.04625408172607422, 0.04656639862060547, 0.04630940628051758, 0.04617116928100586, 0.046201793670654294, 0.04617116928100586, 0.04622230529785156, 0.04611174392700195, 0.0462366714477539, 0.04619468688964844, 0.046203903198242184, 0.04619366455078125, 0.046225406646728515, 0.04624281692504883, 0.046296062469482424, 0.04619468688964844, 0.046225406646728515, 0.04623462295532227, 0.04624895858764649, 0.046203903198242184, 0.04621619033813477, 0.046246910095214845, 0.04627046585083008, 0.046238719940185545, 0.046263294219970705, 0.046289920806884766, 0.04636262512207031, 0.046257152557373046, 0.046260223388671876, 0.046279678344726564, 0.04628377532958984, 0.046357505798339846, 0.04629302215576172, 0.046289886474609375, 0.0462213134765625, 0.04627251052856445, 0.046371841430664064, 0.04626432037353516, 0.04634527969360352, 0.04634719848632812, 0.046322689056396485, 0.04631760025024414, 0.046332897186279295, 0.046306304931640625, 0.04633087921142578, 0.046339073181152345, 0.0463001594543457, 0.0464005126953125, 0.0463636474609375, 0.0463636474609375, 0.046317569732666014, 0.046342144012451174, 0.04665651321411133, 0.046399486541748046, 0.046377983093261715, 0.04641484832763672, 0.04632064056396484, 0.04652646255493164, 0.04635340881347656, 0.04637900924682617, 0.1516584930419922, 0.046227455139160156, 0.04627763366699219, 0.04619571304321289, 0.046150657653808595, 0.04620601654052734, 0.04617619323730469, 0.04628684616088867, 0.046220287322998044, 0.04617830276489258, 0.04616806411743164, 0.04623052978515625, 0.04624998474121094, 0.04620806503295898, 0.04621612930297852, 0.046189567565917966, 0.04620291137695313, 0.04632675170898438, 0.046289920806884766, 0.04622438430786133, 0.04620185470581055, 0.04624486541748047, 0.0462366714477539, 0.04637286376953125, 0.046615550994873044, 0.04629094314575195, 0.04627046585083008, 0.04623257446289063, 0.04634624099731445, 0.046246944427490236, 0.04631343841552735, 0.046300224304199215, 0.046261184692382815, 0.046698497772216796, 0.04632166290283203, 0.04636671829223633, 0.04633913421630859, 0.04634003067016602, 0.046532608032226565, 0.046322689056396485, 0.04633599853515625, 0.04634931182861328, 0.04629913711547851, 0.04630732727050781, 0.04635136032104492, 0.04630732727050781, 0.04634316635131836, 0.04634828948974609, 0.046365760803222654, 0.0463256950378418, 0.04641484832763672, 0.04630220794677734, 0.04635446548461914, 0.04632271957397461, 0.04637177658081055, 0.04640972900390625, 0.0464087028503418, 0.046403648376464844, 0.0464700813293457, 0.046377983093261715, 0.04652236938476562, 0.04637081527709961, 0.04641484832763672, 0.1505628204345703, 0.046166015625, 0.046167041778564455, 0.04616499328613281, 0.04619776153564453, 0.04631654357910156, 0.046459903717041014, 0.046427135467529294, 0.04626124954223633, 0.04623257446289063, 0.04618137741088867, 0.0464087028503418, 0.04626227188110352, 0.04619059371948242, 0.04619673538208008, 0.04623052978515625, 0.04620185470581055, 0.046258174896240234, 0.046344192504882815, 0.046246910095214845, 0.04628073501586914, 0.04622639846801758, 0.04628684616088867, 0.04623257446289063, 0.046159870147705076, 0.04628275299072265, 0.046246910095214845, 0.04633804702758789, 0.04642508697509766, 0.046295040130615236, 0.046265342712402346, 0.04629817581176758, 0.04637996673583984, 0.04641996765136719, 0.04636467361450195, 0.04631039810180664, 0.04631964874267578, 0.0463595199584961, 0.046295040130615236, 0.04629913711547851, 0.04627763366699219, 0.04631145477294922, 0.046322654724121094, 0.046309375762939455, 0.0463524169921875, 0.046289886474609375, 0.04636262512207031, 0.04636876678466797, 0.046341121673583986, 0.046298110961914066, 0.04642099380493164, 0.04647225570678711, 0.04631238555908203, 0.046312446594238284, 0.0464005126953125, 0.04640768051147461, 0.04637900924682617, 0.04634009552001953, 0.04644147109985351, 0.04635033416748047, 0.046410751342773435, 0.04636774444580078, 0.04634726333618164, 0.15076864624023437, 0.04616294479370117, 0.046139392852783206, 0.04619059371948242, 0.04616089630126953, 0.04633087921142578, 0.04616499328613281, 0.046241790771484374, 0.0461844482421875, 0.04619571304321289, 0.04618239974975586, 0.04619878387451172, 0.04614963150024414, 0.04621004867553711, 0.04619059371948242, 0.046196800231933594, 0.04624582290649414, 0.04620185470581055, 0.04619776153564453, 0.04638003158569336, 0.046342144012451174, 0.046258174896240234, 0.04619059371948242, 0.04634009552001953, 0.04659097671508789, 0.04654694366455078, 0.04630835342407227, 0.04624998474121094, 0.04629094314575195, 0.0462295036315918, 0.046333953857421874, 0.046312446594238284, 0.046266368865966793, 0.046312446594238284, 0.04626124954223633, 0.04633292770385742, 0.04627251052856445, 0.046342144012451174, 0.046279678344726564, 0.046306304931640625, 0.04628582382202148, 0.046309375762939455, 0.04632166290283203, 0.04627558517456055, 0.04633292770385742, 0.04628793716430664, 0.04633695983886719, 0.04633190536499023, 0.046371841430664064, 0.04636262512207031, 0.04635238265991211, 0.04632064056396484, 0.04637286376953125, 0.046360671997070314, 0.04646390533447266, 0.04637900924682617, 0.04636675262451172, 0.04636463928222656, 0.04640358352661133, 0.046413822174072264, 0.046445568084716796, 0.046322689056396485, 0.046399486541748046, 0.15092326354980468, 0.04618751907348633, 0.04616191864013672, 0.04616396713256836, 0.04613017654418945, 0.04618239974975586, 0.04614451217651367, 0.04613945770263672, 0.046157760620117186, 0.04619059371948242, 0.046238719940185545, 0.04633804702758789, 0.04617420959472656, 0.04623052978515625, 0.04622643280029297, 0.046209022521972655, 0.04625100708007813, 0.04628070449829102, 0.04620800018310547, 0.04639846420288086, 0.04655615997314453, 0.04637593460083008, 0.04623052978515625, 0.04621619033813477, 0.04628582382202148, 0.04624486541748047, 0.046322689056396485, 0.046289920806884766, 0.04627763366699219, 0.04633497619628906, 0.04625408172607422, 0.04635647964477539, 0.04631552124023437, 0.04632371139526367, 0.0462479362487793, 0.04628377532958984, 0.04628582382202148, 0.046279678344726564, 0.046284801483154295, 0.04628684616088867, 0.0463001594543457, 0.04625920104980469, 0.046355457305908204, 0.04640972900390625, 0.046311424255371096, 0.046333953857421874, 0.046339073181152345, 0.046312446594238284, 0.04635955047607422, 0.046309375762939455, 0.04639129638671875, 0.04630220794677734, 0.04638719940185547, 0.04630633544921875, 0.046375904083251956, 0.046372894287109376, 0.046390239715576174, 0.04636262512207031, 0.04639129638671875, 0.04639129638671875, 0.046410751342773435, 0.046301185607910154, 0.04639334487915039, 0.15058125305175782, 0.04616294479370117, 0.04622848129272461, 0.046148609161376954, 0.046183425903320315, 0.04619059371948242, 0.04620800018310547, 0.04618035125732422, 0.04615475082397461, 0.04622441482543945, 0.0462059211730957, 0.046220287322998044, 0.046150657653808595, 0.04624588775634766, 0.04619366455078125, 0.046241790771484374, 0.04619366455078125, 0.04628377532958984, 0.04618547058105469, 0.04625920104980469, 0.04619776153564453, 0.04623155212402344, 0.04621311950683594, 0.046202880859375, 0.046265342712402346, 0.0462479362487793, 0.04643328094482422, 0.04653363037109375, 0.04628380966186523, 0.046298080444335935, 0.04627046585083008, 0.04636467361450195, 0.04634521484375, 0.04633804702758789, 0.04632064056396484, 0.046322689056396485, 0.0463370246887207, 0.04630220794677734, 0.04635955047607422, 0.04629708862304688, 0.04632169723510742, 0.04632060623168945, 0.04627251052856445, 0.046298110961914066, 0.046333953857421874, 0.04627865600585938, 0.046304256439208984, 0.046287872314453124, 0.04635955047607422, 0.04634931182861328, 0.046358528137207033, 0.04627558517456055, 0.046371841430664064, 0.046306304931640625, 0.04632883071899414, 0.046357505798339846, 0.04640972900390625, 0.04635446548461914, 0.046436321258544924, 0.04636671829223633, 0.04644147109985351, 0.04636569595336914, 0.046430206298828124]",tokens/s,20.919484972853493,,,main,False,False,,,,,,,,,,,,,,,,,,,,,,, float16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-160m,EleutherAI/pythia-160m,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1287.389184,1147.666432,0.0,501.219328,454.728704,s,15,0.17006012821197508,0.01133734188079834,0.0002849923468754685,0.011215904235839843,0.011597791862487793,0.011808687877655029,0.01215660451889038,"[0.012243583679199218, 0.011485183715820312, 0.011278592109680175, 0.011136768341064453, 0.011193152427673339, 0.011155455589294434, 0.011228096008300782, 0.011205951690673828, 0.011115008354187012, 0.011201855659484863, 0.0111810884475708, 0.011215904235839843, 0.011236160278320313, 0.011561023712158202, 0.01162230396270752]",tokens/s,22580.24876479894,kWh,1.3697373102695706e-07,7.505512290149027e-08,3.167383762178472e-07,5.287672301462945e-07,tokens/kWh,484144979.8792793,MB,1287.389184,1147.666432,0.0,501.219328,469.3632,s,15,9.863984619140625,0.657598974609375,0.010718256338232646,0.654005126953125,0.6688111083984375,0.6752843383789062,0.6845559887695313,"[0.6868739013671875, 0.6665516967773437, 0.6462833862304688, 0.654005126953125, 0.650184814453125, 0.6459300537109375, 0.6535787963867188, 0.6566679077148437, 0.6592887573242188, 0.6465830688476563, 0.65210205078125, 0.6498375244140625, 0.6613882446289062, 0.6703173828125, 0.6643919067382813]",tokens/s,95.8030690930184,kWh,7.723144461076218e-06,4.231973346107364e-06,1.295462640753183e-05,2.490974421471541e-05,tokens/kWh,2529130.7472672807,,s,944,10.009720834732054,0.010603517833402602,0.00143755452198337,0.010309632301330567,0.010773299026489259,0.010956032419204712,0.021927895050048826,"[0.01194700813293457, 0.011645952224731445, 0.011081727981567382, 0.01080832004547119, 0.010802176475524903, 0.010753024101257324, 0.01071718406677246, 0.01082470417022705, 0.010895359992980956, 0.01094758415222168, 0.010881024360656738, 0.010767359733581543, 0.010873855590820313, 0.010915840148925781, 0.010890239715576172, 0.010799103736877442, 0.010862591743469239, 0.010829824447631836, 0.01076633644104004, 0.010703871726989746, 0.010760191917419434, 0.010791935920715333, 0.010861568450927735, 0.010959872245788574, 0.010884096145629883, 0.010803199768066407, 0.01083289623260498, 0.010819583892822266, 0.010736639976501466, 0.010721280097961425, 0.01084620761871338, 0.011355135917663574, 0.011068415641784669, 0.011109375953674316, 0.010770432472229004, 0.010706944465637207, 0.01095680046081543, 0.010993663787841796, 0.0110632963180542, 0.010889216423034668, 0.010774527549743652, 0.010705920219421386, 0.010920960426330567, 0.010943488121032715, 0.010982399940490722, 0.01084620761871338, 0.01095680046081543, 0.010974207878112792, 0.011106304168701172, 0.011149312019348144, 0.011324416160583496, 0.010958847999572753, 0.010589183807373047, 0.010644479751586914, 0.010688511848449708, 0.010650624275207519, 0.010621952056884766, 0.010655743598937988, 0.010573823928833008, 0.010369024276733398, 0.010475520133972169, 0.010765312194824218, 0.02255564880371094, 0.010587136268615722, 0.010561535835266114, 0.01064243221282959, 0.010549247741699219, 0.01071513557434082, 0.010584063529968261, 0.0106496000289917, 0.010617856025695802, 0.010644479751586914, 0.010575872421264648, 0.010639360427856445, 0.011003904342651367, 0.01073459243774414, 0.01062604808807373, 0.010622976303100586, 0.010677248001098634, 0.010599424362182617, 0.010601471900939942, 0.010552319526672363, 0.010951680183410644, 0.01064038372039795, 0.010584063529968261, 0.010647551536560058, 0.010590208053588868, 0.010389504432678222, 0.010317824363708495, 0.01031276798248291, 0.010306495666503906, 0.010671104431152344, 0.010445823669433594, 0.010536959648132324, 0.010599424362182617, 0.010580991744995117, 0.010620927810668946, 0.010584063529968261, 0.01042124843597412, 0.01044480037689209, 0.010571776390075683, 0.010652671813964844, 0.010593279838562012, 0.010483712196350097, 0.010996735572814942, 0.011268095970153809, 0.011736063957214356, 0.011225088119506836, 0.010821632385253906, 0.01062399959564209, 0.010629119873046875, 0.010387455940246582, 0.010259455680847167, 0.010267647743225097, 0.010379263877868652, 0.010339327812194824, 0.010353695869445801, 0.010347488403320313, 0.0103372802734375, 0.01032192039489746, 0.010243071556091308, 0.010202112197875977, 0.010213376045227051, 0.010209280014038086, 0.010227711677551269, 0.022036544799804686, 0.010308544158935546, 0.010287103652954101, 0.010337311744689942, 0.010313695907592773, 0.010298368453979492, 0.01033011245727539, 0.010275839805603027, 0.010322943687438964, 0.010191871643066406, 0.010225664138793946, 0.010217472076416016, 0.01019596767425537, 0.010225664138793946, 0.010250240325927735, 0.010236991882324218, 0.010216383934020997, 0.010218496322631837, 0.010177536010742188, 0.010224639892578125, 0.010226688385009765, 0.010233856201171876, 0.010298368453979492, 0.01031987190246582, 0.010334207534790038, 0.01029430389404297, 0.010328031539916993, 0.01024614429473877, 0.010199040412902831, 0.010208255767822265, 0.010169343948364258, 0.010239999771118164, 0.01019596767425537, 0.010265600204467774, 0.010166272163391114, 0.010213376045227051, 0.010261504173278809, 0.010387455940246582, 0.010198016166687012, 0.01019596767425537, 0.010203136444091796, 0.010177536010742188, 0.010231807708740234, 0.010216447830200195, 0.01023795223236084, 0.010203136444091796, 0.010178560256958008, 0.01023795223236084, 0.010193920135498047, 0.010210304260253907, 0.010212384223937988, 0.010221535682678223, 0.010203136444091796, 0.010270719528198241, 0.010297344207763673, 0.010293248176574708, 0.010299391746520996, 0.010293248176574708, 0.010353664398193359, 0.0103372802734375, 0.010301440238952637, 0.010278911590576171, 0.010275872230529785, 0.02186851119995117, 0.010468352317810058, 0.010290176391601562, 0.010293248176574708, 0.010314751625061035, 0.010268671989440918, 0.01030246353149414, 0.01031167984008789, 0.010342399597167968, 0.01031987190246582, 0.01043558406829834, 0.010267711639404297, 0.010176447868347168, 0.010200063705444335, 0.010208255767822265, 0.010244095802307129, 0.010306559562683105, 0.010270719528198241, 0.010331199645996094, 0.010189760208129882, 0.010205183982849121, 0.010254336357116698, 0.010284031867980957, 0.01030348777770996, 0.010298368453979492, 0.01027996826171875, 0.010292192459106446, 0.011000831604003907, 0.010777600288391113, 0.01061683177947998, 0.01063321590423584, 0.010637311935424805, 0.010658816337585449, 0.010312704086303711, 0.010177536010742188, 0.01016320037841797, 0.010166272163391114, 0.010174495697021484, 0.010244064331054688, 0.010262528419494628, 0.010276864051818848, 0.010265600204467774, 0.010238975524902343, 0.010249216079711914, 0.010244095802307129, 0.010151935577392577, 0.010243071556091308, 0.010275839805603027, 0.010287103652954101, 0.010151935577392577, 0.0101396484375, 0.010174464225769043, 0.010274815559387206, 0.010285056114196778, 0.01131827163696289, 0.010774527549743652, 0.010656767845153809, 0.010694656372070312, 0.010979328155517578, 0.010674176216125488, 0.010673151969909669, 0.010482687950134278, 0.010471424102783204, 0.021975040435791016, 0.010297344207763673, 0.01032806396484375, 0.010210304260253907, 0.01018777561187744, 0.010201087951660156, 0.010608639717102051, 0.010395648002624512, 0.010272768020629883, 0.010268671989440918, 0.010295295715332031, 0.010264575958251953, 0.010284031867980957, 0.010275839805603027, 0.01032192039489746, 0.010306559562683105, 0.010488832473754883, 0.01083903980255127, 0.010654720306396484, 0.010474495887756348, 0.010475520133972169, 0.010592255592346191, 0.010750975608825684, 0.010567680358886718, 0.010202112197875977, 0.010169343948364258, 0.010166272163391114, 0.01023692798614502, 0.010280960083007813, 0.010254336357116698, 0.010299391746520996, 0.010265600204467774, 0.010290176391601562, 0.010264575958251953, 0.010314751625061035, 0.010272768020629883, 0.010259455680847167, 0.010262528419494628, 0.010153984069824219, 0.010157055854797363, 0.010275839805603027, 0.010334207534790038, 0.010290176391601562, 0.010304512023925782, 0.010308608055114746, 0.010251263618469238, 0.010309632301330567, 0.010263551712036132, 0.010294272422790527, 0.010250240325927735, 0.010306559562683105, 0.010184767723083497, 0.010179519653320312, 0.010217472076416016, 0.010214400291442872, 0.010300415992736817, 0.010281984329223632, 0.010292223930358887, 0.010326016426086425, 0.010250240325927735, 0.010309632301330567, 0.010151935577392577, 0.010359807968139649, 0.02191155242919922, 0.010306559562683105, 0.010279935836791992, 0.010296319961547852, 0.010251263618469238, 0.010264575958251953, 0.010312735557556152, 0.010259424209594727, 0.010267647743225097, 0.010160127639770507, 0.010208288192749023, 0.010183648109436035, 0.010235903739929199, 0.010347519874572754, 0.010285056114196778, 0.01031372833251953, 0.010258432388305663, 0.010293248176574708, 0.010490880012512208, 0.010339327812194824, 0.010278911590576171, 0.01030246353149414, 0.010297344207763673, 0.010233856201171876, 0.01025331211090088, 0.010541055679321289, 0.01043455982208252, 0.010317824363708495, 0.010264575958251953, 0.010278911590576171, 0.010261504173278809, 0.010257408142089844, 0.010258432388305663, 0.010186752319335938, 0.010193920135498047, 0.010189824104309082, 0.01019596767425537, 0.01016320037841797, 0.010214400291442872, 0.010158080101013184, 0.010163264274597168, 0.010183615684509277, 0.010205183982849121, 0.0102041597366333, 0.0101396484375, 0.010216447830200195, 0.010184703826904297, 0.010190848350524903, 0.010177536010742188, 0.01022156810760498, 0.010169407844543458, 0.010184639930725098, 0.010181632041931152, 0.01018777561187744, 0.010258432388305663, 0.010277888298034669, 0.010280960083007813, 0.010238975524902343, 0.01018777561187744, 0.010168319702148437, 0.01019596767425537, 0.010202112197875977, 0.01015500831604004, 0.0220579833984375, 0.010280960083007813, 0.010298368453979492, 0.010285056114196778, 0.010217472076416016, 0.010249216079711914, 0.010196991920471191, 0.010307583808898926, 0.010273792266845704, 0.010445823669433594, 0.010683391571044922, 0.010716159820556641, 0.010476544380187988, 0.010293248176574708, 0.01017958354949951, 0.010192928314208984, 0.01019388771057129, 0.010136575698852538, 0.010152959823608398, 0.01017958354949951, 0.010156031608581542, 0.010186752319335938, 0.010159104347229005, 0.010142720222473145, 0.010170368194580079, 0.010199040412902831, 0.01015500831604004, 0.010256383895874023, 0.010232831954956055, 0.010314751625061035, 0.01023078441619873, 0.010257439613342285, 0.010316767692565919, 0.010272768020629883, 0.010293248176574708, 0.010257408142089844, 0.010208255767822265, 0.01022054386138916, 0.010145792007446289, 0.0102041597366333, 0.010164223670959472, 0.010160127639770507, 0.010156031608581542, 0.0102041597366333, 0.01017961597442627, 0.01031164836883545, 0.010260479927062988, 0.010312704086303711, 0.010188799858093261, 0.01041919994354248, 0.010858495712280274, 0.012529664039611817, 0.011480064392089843, 0.010736639976501466, 0.010696767807006836, 0.010734527587890626, 0.010716192245483399, 0.010587136268615722, 0.010934240341186523, 0.010275839805603027, 0.010255423545837403, 0.010290111541748047, 0.010184703826904297, 0.021957632064819335, 0.010241056442260742, 0.0102706880569458, 0.010264575958251953, 0.010320896148681641, 0.010305536270141602, 0.010185728073120117, 0.010222592353820802, 0.010174464225769043, 0.010271743774414062, 0.010280960083007813, 0.010280960083007813, 0.010294272422790527, 0.010294272422790527, 0.010288127899169922, 0.010306559562683105, 0.01025331211090088, 0.010144767761230468, 0.01020524787902832, 0.01015494441986084, 0.010217472076416016, 0.010183679580688476, 0.010176544189453124, 0.01028707218170166, 0.010576895713806152, 0.010655743598937988, 0.010687487602233887, 0.010651647567749023, 0.010686464309692383, 0.010638336181640624, 0.010664959907531739, 0.010647551536560058, 0.010656767845153809, 0.010644479751586914, 0.010688511848449708, 0.010670080184936523, 0.010645503997802735, 0.010639360427856445, 0.01064857578277588, 0.01064038372039795, 0.010693632125854492, 0.010692607879638672, 0.010713088035583495, 0.010701824188232421, 0.01070899200439453, 0.010678272247314453, 0.010705920219421386, 0.010661888122558593, 0.010300415992736817, 0.010323967933654785, 0.010256383895874023, 0.010288127899169922, 0.01030246353149414, 0.010262528419494628, 0.010279935836791992, 0.010297344207763673, 0.010345472335815429, 0.010243103981018067, 0.010313695907592773, 0.010279935836791992, 0.010249216079711914, 0.010329088211059571, 0.010255359649658203, 0.02187468719482422, 0.010267647743225097, 0.010304512023925782, 0.010250240325927735, 0.010275839805603027, 0.010268671989440918, 0.010268671989440918, 0.010263551712036132, 0.010241024017333985, 0.010198016166687012, 0.010193920135498047, 0.01018777561187744, 0.010140671730041503, 0.010186752319335938, 0.01021235179901123, 0.010167296409606934, 0.010232831954956055, 0.010191871643066406, 0.010161151885986328, 0.010140671730041503, 0.010180607795715332, 0.01015708827972412, 0.010220512390136719, 0.010576895713806152, 0.01062399959564209, 0.01144422435760498, 0.010851327896118163, 0.010688511848449708, 0.010724351882934571, 0.01073459243774414, 0.010651647567749023, 0.010652671813964844, 0.010655743598937988, 0.010675200462341309, 0.010667008399963379, 0.010665984153747558, 0.010680319786071778, 0.010651647567749023, 0.010667008399963379, 0.010685440063476562, 0.010667008399963379, 0.010674240112304688, 0.010661824226379395, 0.010697728157043456, 0.010670111656188965, 0.010633184432983398, 0.010674176216125488, 0.01081446361541748, 0.010710016250610351, 0.010686464309692383, 0.010672127723693848, 0.010667008399963379, 0.01032192039489746, 0.01030246353149414, 0.010315775871276855, 0.010324992179870606, 0.010285056114196778, 0.010316800117492676, 0.010261504173278809, 0.01032806396484375, 0.010306559562683105, 0.010272768020629883, 0.010292223930358887, 0.021940223693847655, 0.010307583808898926, 0.01031987190246582, 0.010296319961547852, 0.010299391746520996, 0.01025331211090088, 0.010295295715332031, 0.010281984329223632, 0.010285056114196778, 0.01031884765625, 0.010339327812194824, 0.010277888298034669, 0.010284031867980957, 0.01021132755279541, 0.010207263946533204, 0.010226655960083007, 0.010167296409606934, 0.010248191833496094, 0.010262528419494628, 0.010303520202636718, 0.010313695907592773, 0.010264575958251953, 0.010269696235656739, 0.010273792266845704, 0.010332159996032715, 0.010261504173278809, 0.010287103652954101, 0.01030348777770996, 0.01031884765625, 0.01032806396484375, 0.010293248176574708, 0.010314751625061035, 0.01033011245727539, 0.010308608055114746, 0.010300415992736817, 0.01024614429473877, 0.010281984329223632, 0.010264575958251953, 0.010262528419494628, 0.010222592353820802, 0.010156031608581542, 0.010162176132202149, 0.010166272163391114, 0.010165247917175293, 0.010182656288146973, 0.010231807708740234, 0.010276864051818848, 0.010273792266845704, 0.010262528419494628, 0.01025331211090088, 0.010275839805603027, 0.010249216079711914, 0.01031884765625, 0.01025331211090088, 0.010213376045227051, 0.010196991920471191, 0.010196991920471191, 0.010171392440795898, 0.010202112197875977, 0.01018777561187744, 0.010160127639770507, 0.01019596767425537, 0.010164223670959472, 0.021906431198120118, 0.010268671989440918, 0.010267647743225097, 0.010199040412902831, 0.01022873592376709, 0.01016320037841797, 0.010165247917175293, 0.010173439979553223, 0.010182656288146973, 0.010189824104309082, 0.010134528160095215, 0.010269696235656739, 0.010144767761230468, 0.010199040412902831, 0.01023692798614502, 0.010194944381713868, 0.010164223670959472, 0.010156031608581542, 0.010175488471984863, 0.010174464225769043, 0.010193920135498047, 0.010142720222473145, 0.01018777561187744, 0.011408384323120118, 0.01152511978149414, 0.010867712020874023, 0.01071513557434082, 0.010645503997802735, 0.010655743598937988, 0.010339327812194824, 0.010275872230529785, 0.010327008247375489, 0.01043455982208252, 0.010655743598937988, 0.0107325439453125, 0.01057596778869629, 0.010319775581359863, 0.010169343948364258, 0.010164223670959472, 0.010286080360412597, 0.010270719528198241, 0.01032703971862793, 0.010315775871276855, 0.010320896148681641, 0.010329088211059571, 0.010283007621765136, 0.01032806396484375, 0.010273792266845704, 0.010299391746520996, 0.010289152145385743, 0.010296319961547852, 0.010334207534790038, 0.010317824363708495, 0.010309632301330567, 0.010309632301330567, 0.010376192092895508, 0.01030348777770996, 0.010317824363708495, 0.01032703971862793, 0.010315775871276855, 0.010303520202636718, 0.010255328178405762, 0.010294272422790527, 0.021961727142333985, 0.01031372833251953, 0.01031987190246582, 0.010274911880493164, 0.010286016464233399, 0.010308575630187989, 0.010306559562683105, 0.010285056114196778, 0.010300415992736817, 0.010299391746520996, 0.01031167984008789, 0.01032703971862793, 0.010305567741394043, 0.010339296340942383, 0.010339327812194824, 0.010295295715332031, 0.010380288124084473, 0.010292223930358887, 0.010304512023925782, 0.010299391746520996, 0.010262528419494628, 0.010307583808898926, 0.010183679580688476, 0.010209280014038086, 0.010180607795715332, 0.010184703826904297, 0.010185728073120117, 0.010200063705444335, 0.010214431762695313, 0.010220512390136719, 0.010198016166687012, 0.010175488471984863, 0.010205183982849121, 0.010226688385009765, 0.010393600463867187, 0.01042636775970459, 0.010290176391601562, 0.010762240409851074, 0.010694656372070312, 0.010365951538085937, 0.010274815559387206, 0.010277888298034669, 0.010297344207763673, 0.01031884765625, 0.010329088211059571, 0.01033523178100586, 0.010312704086303711, 0.010396672248840331, 0.010289152145385743, 0.010331135749816894, 0.010275839805603027, 0.010304512023925782, 0.010297344207763673, 0.010296319961547852, 0.010340352058410645, 0.010300415992736817, 0.010386431694030761, 0.010370047569274902, 0.010296319961547852, 0.010333184242248536, 0.010369024276733398, 0.010305536270141602, 0.010284031867980957, 0.02208870315551758, 0.010304512023925782, 0.010288127899169922, 0.010329088211059571, 0.01031884765625, 0.010287103652954101, 0.010281984329223632, 0.010389504432678222, 0.010284031867980957, 0.01032192039489746, 0.01034444808959961, 0.010285056114196778, 0.010370047569274902, 0.010346495628356933, 0.010315775871276855, 0.010315775871276855, 0.010291199684143066, 0.010294272422790527, 0.010258432388305663, 0.01030348777770996, 0.010267647743225097, 0.01032192039489746, 0.010324992179870606, 0.010297344207763673, 0.010295295715332031, 0.010589183807373047, 0.010703871726989746, 0.010667008399963379, 0.01064243221282959, 0.01071513557434082, 0.010683391571044922, 0.010308608055114746, 0.010843135833740235, 0.011199487686157226, 0.011066368103027344, 0.010689536094665527, 0.010691583633422852, 0.010676223754882813, 0.010706944465637207, 0.010656767845153809, 0.010664959907531739, 0.01064243221282959, 0.01076633644104004, 0.010700799942016602, 0.010628095626831055, 0.010685440063476562, 0.0106496000289917, 0.010761216163635253, 0.010743807792663575, 0.010702848434448242, 0.010631168365478515, 0.010722304344177246, 0.010698752403259277, 0.010295295715332031, 0.010296319961547852, 0.010294272422790527, 0.010279935836791992, 0.010307583808898926, 0.010261504173278809, 0.010309632301330567, 0.010263551712036132, 0.010306559562683105, 0.010688511848449708, 0.022708223342895507, 0.010661888122558593, 0.010804224014282226, 0.010703871726989746, 0.010769408226013183, 0.010691583633422852, 0.010657792091369628, 0.01073459243774414, 0.010665984153747558, 0.010653696060180663, 0.010786879539489745, 0.010708928108215332, 0.010802176475524903, 0.01074176025390625, 0.010681344032287597, 0.01032703971862793, 0.01033523178100586, 0.010322943687438964, 0.010272768020629883, 0.010276864051818848, 0.010537983894348145, 0.010799103736877442, 0.010680319786071778, 0.010911744117736816, 0.010748928070068359, 0.010788864135742187, 0.010856448173522949, 0.01073971176147461, 0.010681344032287597, 0.010654720306396484, 0.010671104431152344, 0.010704895973205567, 0.010690560340881347, 0.0106496000289917, 0.01071513557434082, 0.010683391571044922, 0.01061068820953369, 0.010320896148681641, 0.010285056114196778, 0.010329088211059571, 0.010474495887756348, 0.010500096321105956, 0.01044377613067627, 0.010645503997802735, 0.010710016250610351, 0.010693632125854492, 0.010660863876342774, 0.010684415817260743, 0.010710016250610351, 0.01063526439666748, 0.010662912368774414, 0.010674176216125488, 0.010653696060180663, 0.010580991744995117, 0.010507264137268067, 0.010716159820556641, 0.010702848434448242, 0.010722304344177246, 0.010639360427856445, 0.010659839630126953, 0.01063424015045166, 0.010573856353759766, 0.010669024467468261, 0.022772768020629882, 0.010669024467468261, 0.010667008399963379, 0.010680319786071778, 0.010488896369934083, 0.01080620765686035, 0.010677248001098634, 0.010247167587280273, 0.010291263580322266, 0.01024505615234375, 0.010251263618469238, 0.010210304260253907, 0.010201087951660156, 0.010189824104309082, 0.010317824363708495, 0.010384384155273438, 0.010718208312988281, 0.010664959907531739, 0.010671104431152344, 0.010796031951904296, 0.010652735710144043, 0.010653632164001465, 0.01062707233428955, 0.010628095626831055, 0.01064345645904541, 0.010701824188232421, 0.01063424015045166, 0.010675200462341309, 0.010650624275207519, 0.01063526439666748, 0.01063526439666748, 0.010647551536560058, 0.010659839630126953, 0.010791935920715333, 0.010728447914123536, 0.010333184242248536, 0.010177536010742188, 0.010183679580688476, 0.010259455680847167, 0.010264575958251953, 0.010300415992736817, 0.010296319961547852, 0.010273792266845704, 0.010314751625061035, 0.010350591659545898, 0.010274815559387206, 0.010261504173278809, 0.01030348777770996, 0.010591232299804687, 0.010631168365478515, 0.010671104431152344, 0.011006976127624512, 0.010704895973205567, 0.010665984153747558, 0.010972160339355469, 0.010657792091369628, 0.010670080184936523, 0.010680319786071778, 0.01064243221282959, 0.01064857578277588, 0.010398719787597656, 0.010783743858337403, 0.010810367584228516]",tokens/s,94.30832443642963,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-6.7b,,cuda,0,42,,,,,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,217063f5c507ed7cc255df7e1f64c4333a0b4dfe,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,,MB,1298.464768,15125.184512,0.0,14478.737408,14241.165312,s,10,0.9019541702270508,0.09019541702270509,0.0014070067081702311,0.08971971130371094,0.09028701171875,0.09235055999755859,0.09400139862060547,"[0.09441410827636719, 0.08973532867431641, 0.08973795318603515, 0.08967823791503907, 0.08969872283935547, 0.08982844543457032, 0.08970409393310547, 0.08967440032958984, 0.08969414520263672, 0.08978873443603516]",tokens/s,2838.2816827107363,kWh,1.0628362437562336e-06,5.823814746512263e-07,5.464292565874963e-06,7.109510284282423e-06,tokens/kWh,36008106.01061513,MB,1298.792448,15125.184512,0.0,14478.737408,14438.898176,s,10,23.840818603515626,2.3840818603515626,0.0038407329163161565,2.3829688720703124,2.3896716552734376,2.3911252563476566,2.3922881372070313,"[2.392578857421875, 2.38539697265625, 2.3893486328125, 2.38275048828125, 2.38209912109375, 2.383369384765625, 2.38180029296875, 2.381286376953125, 2.379001220703125, 2.383187255859375]",tokens/s,26.425267121788288,kWh,2.8074540981934186e-05,1.5385748123880732e-05,0.00011904586259772357,0.0001625061517035385,tokens/kWh,387677.6315208762,,s,629,24.610037715911872,0.03912565614612379,0.01069945614568616,0.03781631851196289,0.03808378982543945,0.03821977615356445,0.12785565490722656,"[0.038160385131835936, 0.038356990814208985, 0.03793612670898437, 0.03827302551269531, 0.03815321731567383, 0.0381736946105957, 0.038370304107666016, 0.038329345703125, 0.03819007873535156, 0.03809280014038086, 0.038166526794433595, 0.03850137710571289, 0.03811123275756836, 0.03811840057373047, 0.038191104888916014, 0.037935104370117184, 0.03791462326049805, 0.03787366485595703, 0.037988353729248046, 0.03803340911865234, 0.03801804733276367, 0.037940223693847655, 0.03791155242919922, 0.03766988754272461, 0.0376995849609375, 0.03779072189331055, 0.0377457275390625, 0.037734336853027343, 0.03803852844238281, 0.038040576934814455, 0.037884929656982425, 0.03802521514892578, 0.037958656311035156, 0.037972991943359374, 0.037991424560546876, 0.0379422721862793, 0.037716991424560545, 0.03786342239379883, 0.03785318374633789, 0.03779993438720703, 0.037776382446289065, 0.0377446403503418, 0.037795841217041014, 0.03778662490844727, 0.03778662490844727, 0.038075393676757815, 0.03825356674194336, 0.03782451248168945, 0.037765121459960936, 0.037935104370117184, 0.03790643310546875, 0.03793612670898437, 0.03788800048828125, 0.03784806442260742, 0.037958656311035156, 0.037874687194824216, 0.03793920135498047, 0.037926910400390625, 0.03788700866699219, 0.03785827255249023, 0.0378419189453125, 0.03788185501098633, 0.12786483001708984, 0.03756748962402344, 0.037736446380615234, 0.03762278366088867, 0.03761151885986328, 0.03758694458007812, 0.03767097473144531, 0.038031295776367186, 0.03779072189331055, 0.03758899307250976, 0.03765862274169922, 0.0376360969543457, 0.03763814544677734, 0.03763814544677734, 0.03765964889526367, 0.037751808166503906, 0.037684223175048825, 0.037631999969482424, 0.03776409530639648, 0.03765760040283203, 0.0376995849609375, 0.03765657424926758, 0.037684223175048825, 0.037823486328125, 0.03770470428466797, 0.037718017578125, 0.03770880126953125, 0.03769651031494141, 0.03774771118164062, 0.03767295837402344, 0.03777536010742188, 0.03784499359130859, 0.03777228927612305, 0.03781836700439453, 0.03849420928955078, 0.03841846466064453, 0.03812662506103515, 0.03796985626220703, 0.038046718597412106, 0.03781836700439453, 0.037945343017578126, 0.03781017684936523, 0.03786137771606445, 0.03796889495849609, 0.03781631851196289, 0.03776204681396484, 0.03782758331298828, 0.03801804733276367, 0.03782963180541992, 0.0378419189453125, 0.037817344665527344, 0.03812454223632813, 0.03841843032836914, 0.03810508728027344, 0.03787161636352539, 0.038300670623779294, 0.03822387313842773, 0.03794841766357422, 0.03809382247924804, 0.0381399040222168, 0.038174720764160154, 0.038250495910644534, 0.03801190567016602, 0.12795187377929687, 0.03770675277709961, 0.037969921112060545, 0.03774054336547852, 0.037779457092285154, 0.0377262077331543, 0.037884929656982425, 0.03777433776855469, 0.03765353775024414, 0.03775689697265625, 0.03794739151000977, 0.0378419189453125, 0.037959678649902344, 0.03790857696533203, 0.037978015899658206, 0.03792793655395508, 0.03798015975952149, 0.03845529556274414, 0.03829452896118164, 0.03778662490844727, 0.037689342498779296, 0.038007808685302735, 0.03810508728027344, 0.0380313606262207, 0.03783782577514649, 0.037928958892822266, 0.037959678649902344, 0.0378419189453125, 0.03773952102661133, 0.037981182098388674, 0.03804876708984375, 0.03805491256713867, 0.03775590515136719, 0.03786444854736328, 0.03793203353881836, 0.038004737854003906, 0.037928958892822266, 0.037823486328125, 0.037943294525146484, 0.03804159927368164, 0.03782656097412109, 0.03800678253173828, 0.03809894561767578, 0.037994495391845705, 0.037823486328125, 0.03788496017456055, 0.037933025360107425, 0.03845529556274414, 0.03800166320800781, 0.037787647247314454, 0.0378869743347168, 0.03784908676147461, 0.03785728073120117, 0.03784601593017578, 0.038091777801513675, 0.038035457611083984, 0.037874687194824216, 0.03786751937866211, 0.037884929656982425, 0.03782656097412109, 0.03793817520141601, 0.03787059020996094, 0.037923839569091795, 0.12783206176757814, 0.037689342498779296, 0.037719039916992186, 0.03771187210083008, 0.03787161636352539, 0.03788185501098633, 0.03760332870483398, 0.03772518539428711, 0.03775590515136719, 0.037689342498779296, 0.03766067123413086, 0.03903180694580078, 0.03814092636108399, 0.037997566223144534, 0.03782451248168945, 0.037784576416015625, 0.03763711929321289, 0.03768115234375, 0.037703678131103514, 0.03761459350585938, 0.03776921463012695, 0.03770880126953125, 0.0377446403503418, 0.03784499359130859, 0.03771596908569336, 0.03769241714477539, 0.03778559875488281, 0.03771187210083008, 0.03762688064575195, 0.03779891204833984, 0.037751808166503906, 0.037819393157958986, 0.037850112915039064, 0.03774054336547852, 0.03775897598266602, 0.0378081283569336, 0.03788800048828125, 0.03771187210083008, 0.03784601593017578, 0.03792793655395508, 0.03776515197753906, 0.037797855377197265, 0.03776409530639648, 0.03770265579223633, 0.03786547088623047, 0.037748737335205076, 0.03784294509887695, 0.03806105422973633, 0.03782758331298828, 0.03784499359130859, 0.03787571334838867, 0.03779379272460937, 0.037814273834228515, 0.03784396743774414, 0.03781631851196289, 0.037920768737792966, 0.03781631851196289, 0.03782553482055664, 0.0378603515625, 0.03785318374633789, 0.037891071319580076, 0.03788185501098633, 0.03800166320800781, 0.1278730239868164, 0.03751833724975586, 0.037631999969482424, 0.03757567977905273, 0.03765862274169922, 0.037591041564941405, 0.03764326477050781, 0.037720062255859374, 0.03769036865234375, 0.03764633560180664, 0.03772927856445313, 0.03763820648193359, 0.037772224426269534, 0.03777024078369141, 0.03776921463012695, 0.037766143798828124, 0.037743614196777346, 0.03782553482055664, 0.03775392150878906, 0.037682113647460935, 0.03772518539428711, 0.03762588882446289, 0.0378900146484375, 0.03798015975952149, 0.03775897598266602, 0.037735424041748046, 0.03780710220336914, 0.0376627197265625, 0.037771263122558595, 0.03764326477050781, 0.03781222534179687, 0.03782048034667969, 0.0380568962097168, 0.037967872619628903, 0.03797919845581055, 0.03772716903686523, 0.037823486328125, 0.03773132705688476, 0.03784396743774414, 0.03785420989990235, 0.03782553482055664, 0.037784576416015625, 0.037820415496826174, 0.037722110748291016, 0.03783987045288086, 0.0377077751159668, 0.037852161407470705, 0.03786240005493164, 0.03783475112915039, 0.03798732757568359, 0.03809382247924804, 0.03799552154541016, 0.03807027053833008, 0.03791667175292969, 0.03812454223632813, 0.037959678649902344, 0.03790240097045899, 0.03793196868896485, 0.03789004898071289, 0.03775283050537109, 0.03789516830444336, 0.03788595199584961, 0.037966846466064456, 0.1280061492919922, 0.03753472137451172, 0.037720062255859374, 0.03759308624267578, 0.037781505584716796, 0.037773311614990236, 0.03770675277709961, 0.0377968635559082, 0.037713920593261716, 0.03781836700439453, 0.037904384613037106, 0.03783375930786133, 0.0377558708190918, 0.03770265579223633, 0.037713920593261716, 0.037771263122558595, 0.03776204681396484, 0.03763302230834961, 0.03772927856445313, 0.03766681671142578, 0.037667839050292966, 0.03761875152587891, 0.037690303802490235, 0.03776310348510742, 0.03773129653930664, 0.037689342498779296, 0.03782656097412109, 0.037700607299804685, 0.037700607299804685, 0.037733375549316404, 0.03783679962158203, 0.03781843185424805, 0.03779884719848633, 0.037749759674072264, 0.03779379272460937, 0.03770470428466797, 0.037852161407470705, 0.03827302551269531, 0.03829558563232422, 0.03823817443847656, 0.037989376068115234, 0.03789209747314453, 0.037874687194824216, 0.037978111267089845, 0.03787161636352539, 0.037754878997802735, 0.03783478546142578, 0.037881824493408205, 0.03783168029785156, 0.03778559875488281, 0.03778047943115234, 0.03778662490844727, 0.03785625457763672, 0.03782553482055664, 0.0384634895324707, 0.038247425079345705, 0.037884929656982425, 0.03792998504638672, 0.037959678649902344, 0.03781119918823242, 0.03784499359130859, 0.03784294509887695, 0.03790233612060547, 0.1278699493408203, 0.037564414978027344, 0.037628929138183595, 0.03758489608764649, 0.03766067123413086, 0.03759206390380859, 0.037651454925537106, 0.037776382446289065, 0.037655616760253904, 0.03765651321411133, 0.03769343948364258, 0.037641246795654296, 0.03767804718017578, 0.03780505752563477, 0.03793407821655274, 0.03801500701904297, 0.03773641586303711, 0.03768320083618164, 0.03769241714477539, 0.0378603515625, 0.03839900970458984, 0.03777942276000976, 0.037817344665527344, 0.0380948486328125, 0.03793305587768555, 0.03774771118164062, 0.037735424041748046, 0.03775283050537109, 0.03770880126953125, 0.03776409530639648, 0.037743614196777346, 0.03788083267211914, 0.03776204681396484, 0.03774054336547852, 0.03768320083618164, 0.03772415924072266, 0.03777228927612305, 0.03775692749023438, 0.03774566268920899, 0.037823486328125, 0.03782656097412109, 0.03774054336547852, 0.03779481506347656, 0.03779993438720703, 0.03775078582763672, 0.03787673568725586, 0.03788294219970703, 0.03790943908691406, 0.037776382446289065, 0.03781222534179687, 0.038184959411621096, 0.037953536987304685, 0.037889022827148434, 0.037838848114013675, 0.037789695739746096, 0.037940223693847655, 0.037814273834228515, 0.03780505752563477, 0.037803009033203126, 0.037835777282714846, 0.038009952545166016, 0.03791247940063477, 0.03784198379516602, 0.12780025482177734, 0.037574657440185545, 0.037689342498779296, 0.03759718322753906, 0.037647361755371096, 0.03764940643310547, 0.03764019012451172, 0.0377077751159668, 0.03766988754272461, 0.03762688064575195, 0.03762483215332031, 0.03768115234375, 0.03775692749023438, 0.03765760040283203, 0.037823486328125, 0.03818086242675781, 0.03785113525390625, 0.03764940643310547, 0.03772723388671875, 0.03767910385131836, 0.03775590515136719, 0.03782758331298828, 0.037664768218994144, 0.03784707260131836, 0.037832672119140626, 0.037689342498779296, 0.03774054336547852, 0.037733375549316404, 0.03787673568725586, 0.03807027053833008, 0.03794841766357422, 0.03805593490600586, 0.03790643310546875, 0.03779379272460937, 0.03786240005493164, 0.03777849578857422, 0.03777231979370117, 0.037769119262695314, 0.037959678649902344, 0.0381317138671875, 0.03788595199584961, 0.03778559875488281, 0.037773311614990236, 0.03769753646850586, 0.03779379272460937, 0.03773446273803711, 0.03772819137573242, 0.037920768737792966, 0.03785113525390625, 0.037810207366943356, 0.037814239501953124, 0.037781505584716796, 0.037781566619873044, 0.03779065704345703, 0.037814273834228515, 0.037904384613037106, 0.03784601593017578, 0.03783683013916016, 0.03785007858276367, 0.03782252883911133, 0.03784799957275391, 0.03778355026245117, 0.03790335845947266, 0.12795187377929687, 0.037571582794189456, 0.03760947036743164, 0.037561344146728515, 0.03763097763061524, 0.03761356735229492, 0.03763302230834961, 0.037719039916992186, 0.037694465637207034, 0.037550079345703126, 0.0377262077331543, 0.03759001541137695, 0.03772825622558594, 0.03763916778564453, 0.03769651031494141, 0.037823486328125, 0.03783065414428711, 0.03775692749023438, 0.03769548797607422, 0.03764223861694336, 0.037743614196777346, 0.03767910385131836, 0.03773235321044922, 0.037749759674072264, 0.03776313781738281, 0.037639137268066405, 0.03782447814941406, 0.03768832015991211, 0.03774771118164062, 0.0376360969543457, 0.03778559875488281, 0.03784806442260742, 0.037779457092285154, 0.037700607299804685, 0.037806079864501956, 0.03794636917114258, 0.03799244689941406, 0.03773132705688476, 0.037817344665527344, 0.037806079864501956, 0.03780710220336914, 0.03768524932861328, 0.0378092155456543, 0.037690303802490235, 0.037820415496826174, 0.03767603302001953, 0.03787571334838867, 0.03784601593017578, 0.03789004898071289, 0.037730304718017575, 0.03780198287963867, 0.037716991424560545, 0.03781836700439453, 0.03784703826904297, 0.03787366485595703, 0.037928958892822266, 0.03785318374633789, 0.037776382446289065, 0.03788390350341797, 0.03778047943115234, 0.0378869743347168, 0.03777228927612305, 0.03787980651855469, 0.12812083435058594, 0.03765657424926758, 0.03765350341796875, 0.03755724716186523, 0.037645313262939455, 0.03758489608764649, 0.03768217468261719, 0.03793817520141601, 0.03759206390380859, 0.03762688064575195, 0.03763097763061524, 0.03764940643310547, 0.037698558807373043, 0.03762688064575195, 0.0377784309387207, 0.03776409530639648, 0.03770265579223633, 0.03771084976196289, 0.037694465637207034, 0.03762688064575195, 0.03770470428466797, 0.037730304718017575, 0.037726272583007814, 0.03785004806518555, 0.03773747253417969, 0.037663745880126956, 0.037787647247314454, 0.0376627197265625, 0.03767705535888672, 0.037678081512451174, 0.03769548797607422, 0.0378152961730957, 0.037776382446289065, 0.03772415924072266, 0.0378081283569336, 0.03777024078369141, 0.03784294509887695, 0.037835777282714846, 0.0378152961730957, 0.037928958892822266, 0.03818086242675781, 0.03828940963745117, 0.03800166320800781, 0.03778867340087891, 0.03781740951538086, 0.03781830215454102, 0.03806924819946289, 0.038201343536376955, 0.03808256149291992, 0.03789926528930664, 0.0379422721862793, 0.03803859329223633, 0.03790943908691406, 0.037806079864501956, 0.037874687194824216, 0.037907455444335936, 0.03783168029785156, 0.03810406494140625, 0.03812044906616211, 0.037926910400390625, 0.03791155242919922, 0.03808870315551758, 0.03821363067626953]",tokens/s,25.55867679931729,,,main,False,False,,,,,,,,,,,,,,,,,,,,,,, float16-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-13b,facebook/opt-13b,cuda,0,42,,,True,,,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",opt,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-6.7b,,cuda,0,42,,,,,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,d35829e539df8480b726c647eeabf91e41eae047,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,,MB,1210.916864,14382.792704,0.0,13736.3456,13487.531008,s,10,0.9088285980224609,0.0908828598022461,0.0011188802829312761,0.0904892463684082,0.091262548828125,0.09272698516845702,0.09389853424072266,"[0.09419142150878906, 0.09067654418945313, 0.09067558288574219, 0.09093711853027343, 0.09034416198730469, 0.09030604553222656, 0.09044735717773437, 0.09036310577392578, 0.09035612487792968, 0.09053113555908203]",tokens/s,2816.812769283842,kWh,1.0696121238850738e-06,5.861005212913417e-07,5.323087341549429e-06,6.978799986725844e-06,tokens/kWh,36682524.2859703,MB,1210.916864,14382.792704,0.0,13736.3456,13661.259776,s,10,22.005575439453125,2.2005575439453127,0.0067534240568617905,2.197949462890625,2.2105153076171873,2.210736901855469,2.2109141772460936,"[2.2097392578125, 2.21095849609375, 2.210466064453125, 2.192959716796875, 2.19519482421875, 2.199596923828125, 2.1960966796875, 2.196302001953125, 2.194256591796875, 2.2000048828125]",tokens/s,28.629108188213618,kWh,2.591680861017606e-05,1.4203239006414692e-05,0.00011372264678385089,0.00015384269440044164,tokens/kWh,409509.2083866879,,s,629,22.785184753417955,0.036224459067437154,0.010793842260097662,0.034890750885009765,0.0351723518371582,0.035259391784667964,0.12553335479736327,"[0.034729984283447264, 0.03492147064208984, 0.03466239929199219, 0.034726913452148435, 0.03465427017211914, 0.03471558380126953, 0.034783233642578126, 0.034738208770751955, 0.03466134262084961, 0.034802688598632815, 0.034715648651123046, 0.03478227233886719, 0.03479750442504883, 0.03497983932495117, 0.035151870727539065, 0.0347955207824707, 0.034890750885009765, 0.03502592086791992, 0.034872318267822264, 0.03528396987915039, 0.0354334716796875, 0.03521331024169922, 0.03520000076293945, 0.03530342483520508, 0.03505254364013672, 0.0351907844543457, 0.03506892776489258, 0.035160064697265625, 0.035105792999267575, 0.0351016960144043, 0.035253246307373046, 0.03513139343261719, 0.03510476684570313, 0.03511500930786133, 0.03514572906494141, 0.03513651275634765, 0.035149822235107424, 0.035151870727539065, 0.035266559600830076, 0.03520512008666992, 0.035113983154296875, 0.03518463897705078, 0.03514470291137695, 0.03523379135131836, 0.035110912322998046, 0.03525222396850586, 0.03516211318969727, 0.03523075103759766, 0.035122142791748044, 0.03521129608154297, 0.03513238525390625, 0.035250175476074216, 0.03515903854370117, 0.035326976776123044, 0.035266559600830076, 0.035236862182617186, 0.03511603164672852, 0.03519385528564453, 0.035176448822021485, 0.035269630432128905, 0.035053569793701174, 0.035266559600830076, 0.1260953598022461, 0.0350013427734375, 0.034955265045166016, 0.03492454528808594, 0.0350013427734375, 0.034988033294677735, 0.034985984802246094, 0.03500236892700195, 0.03505254364013672, 0.03496345520019531, 0.03504435348510742, 0.03492454528808594, 0.0350832633972168, 0.035020801544189455, 0.03505459213256836, 0.03502489471435547, 0.03504540634155273, 0.035012577056884764, 0.03504230499267578, 0.03498393630981445, 0.035026943206787106, 0.03497574234008789, 0.03504230499267578, 0.035053569793701174, 0.03501670455932617, 0.03504947280883789, 0.0351099853515625, 0.03498179244995117, 0.03516108703613281, 0.0350013427734375, 0.03510988616943359, 0.03512627029418945, 0.03508736038208008, 0.03502489471435547, 0.03522364807128906, 0.035313568115234374, 0.03527065658569336, 0.03509968185424805, 0.03512931060791016, 0.03513753509521484, 0.0351016960144043, 0.035138561248779294, 0.03509145736694336, 0.03508633422851563, 0.03506175994873047, 0.03514572906494141, 0.035197952270507815, 0.03511296081542969, 0.035166206359863283, 0.035122177124023435, 0.0351723518371582, 0.03512934494018555, 0.03513446426391602, 0.03510067367553711, 0.035195903778076174, 0.03519385528564453, 0.03521023941040039, 0.03526348876953125, 0.035323936462402346, 0.03511088180541992, 0.035181568145751956, 0.03508531188964844, 0.03519180679321289, 0.12617318725585938, 0.0349378547668457, 0.035079166412353514, 0.03498700714111328, 0.03501772689819336, 0.03491027069091797, 0.035101631164550784, 0.035138561248779294, 0.035040256500244144, 0.034871295928955076, 0.03506995010375977, 0.03498092651367188, 0.035047359466552734, 0.03500032043457031, 0.035026943206787106, 0.03508531188964844, 0.035053569793701174, 0.03485184097290039, 0.03495935821533203, 0.03502182388305664, 0.03527167892456055, 0.03511500930786133, 0.03508633422851563, 0.03509862518310547, 0.0350382080078125, 0.034917377471923826, 0.035062782287597655, 0.0351723518371582, 0.03571814346313477, 0.03512831878662109, 0.035160064697265625, 0.035110912322998046, 0.034852897644042966, 0.034930656433105466, 0.03508736038208008, 0.035148799896240236, 0.035178497314453126, 0.035171329498291014, 0.03512115097045899, 0.035113983154296875, 0.03506687927246094, 0.03499212646484375, 0.03513241577148438, 0.03503615951538086, 0.035076095581054685, 0.034988033294677735, 0.0351201286315918, 0.035138561248779294, 0.0351836166381836, 0.03504844665527344, 0.03506790542602539, 0.03512627029418945, 0.03512319946289062, 0.035056640625, 0.035110912322998046, 0.0351539192199707, 0.03511603164672852, 0.0350750732421875, 0.03517542266845703, 0.0351016960144043, 0.03518054580688477, 0.035093505859375, 0.03516211318969727, 0.12640870666503906, 0.03479040145874023, 0.03474534225463867, 0.03468492889404297, 0.034678783416748044, 0.03475558471679688, 0.034835456848144535, 0.03487539291381836, 0.03471462249755859, 0.03468601608276367, 0.034755519866943356, 0.034654239654541015, 0.034728927612304686, 0.034715648651123046, 0.03477196884155274, 0.03472281646728516, 0.03476377487182617, 0.034738174438476564, 0.03479142379760742, 0.03471257781982422, 0.034767871856689454, 0.03470848083496094, 0.03476070404052734, 0.034732032775878906, 0.034702335357666016, 0.03470336151123047, 0.034902015686035154, 0.03489996719360351, 0.034746368408203124, 0.03481190490722656, 0.03477913665771484, 0.03483135986328125, 0.03479142379760742, 0.034801727294921876, 0.034834369659423825, 0.03482726287841797, 0.03479244613647461, 0.03474943923950195, 0.03474943923950195, 0.03482112121582031, 0.03481292724609375, 0.03482726287841797, 0.03480473709106445, 0.034723838806152346, 0.03484569549560547, 0.03482624053955078, 0.03486412811279297, 0.0348487663269043, 0.03485696029663086, 0.03477401733398437, 0.0348590087890625, 0.03476070404052734, 0.034816001892089846, 0.03485184097290039, 0.03486207962036133, 0.03488467025756836, 0.034990047454833986, 0.03485795211791992, 0.03491020965576172, 0.034831390380859376, 0.03486921691894531, 0.034841598510742186, 0.035127296447753906, 0.12554137420654296, 0.03465625762939453, 0.03468492889404297, 0.03463065719604492, 0.0346879997253418, 0.03465830230712891, 0.03465727996826172, 0.034764801025390625, 0.034718719482421875, 0.0346429443359375, 0.03470438385009766, 0.03467270278930664, 0.03479955291748047, 0.03465830230712891, 0.034729984283447264, 0.034802688598632815, 0.03467673492431641, 0.03465830230712891, 0.034753536224365236, 0.03491839981079101, 0.034797569274902344, 0.03468288040161133, 0.034767871856689454, 0.034812992095947265, 0.03476883316040039, 0.03477612686157226, 0.035131328582763674, 0.03489177703857422, 0.03481705474853516, 0.034819038391113284, 0.03482729721069336, 0.0349071044921875, 0.034783233642578126, 0.034710529327392575, 0.03479142379760742, 0.03476172637939453, 0.03481087875366211, 0.03478220748901367, 0.03484364700317383, 0.03486310577392578, 0.03484467315673828, 0.03479654312133789, 0.035165184020996096, 0.03510784149169922, 0.03555430221557617, 0.035160064697265625, 0.034917377471923826, 0.03521023941040039, 0.03502191925048828, 0.034874271392822266, 0.03485696029663086, 0.03482009506225586, 0.03487334442138672, 0.034885631561279294, 0.0349716796875, 0.035009502410888675, 0.03505766296386719, 0.0348487663269043, 0.03487539291381836, 0.034825214385986326, 0.034931713104248044, 0.03481804656982422, 0.034852863311767575, 0.12547481536865235, 0.03471462249755859, 0.03468492889404297, 0.034648063659667966, 0.034735103607177735, 0.03467264175415039, 0.034767871856689454, 0.0350013427734375, 0.03484672164916992, 0.03473612976074219, 0.03473920059204102, 0.03488256072998047, 0.03484979248046875, 0.03474534225463867, 0.034825214385986326, 0.03477913665771484, 0.03472793579101562, 0.03482828903198242, 0.034985984802246094, 0.03472895812988281, 0.03475558471679688, 0.034700286865234374, 0.03489689636230469, 0.03499318313598633, 0.03486102294921875, 0.03475046539306641, 0.034786304473876956, 0.03472281646728516, 0.03479142379760742, 0.03477196884155274, 0.03481292724609375, 0.03480473709106445, 0.03489177703857422, 0.03481497573852539, 0.03482316970825195, 0.03474534225463867, 0.03493478393554687, 0.03477407836914063, 0.03486611175537109, 0.03500032043457031, 0.03512627029418945, 0.035095550537109374, 0.035043327331542966, 0.035023872375488284, 0.035068958282470704, 0.03504431915283203, 0.03506073760986328, 0.035148799896240236, 0.0350750732421875, 0.035071041107177736, 0.03517331314086914, 0.03506892776489258, 0.03521535873413086, 0.03503615951538086, 0.034939903259277344, 0.03499724960327148, 0.03509664154052734, 0.034831295013427736, 0.035296257019042966, 0.03536076736450195, 0.035304447174072266, 0.035092479705810545, 0.03486617660522461, 0.12567040252685546, 0.03473100662231445, 0.03498086547851562, 0.03486105728149414, 0.03467468643188477, 0.03471155166625976, 0.03473100662231445, 0.03474227142333984, 0.034700286865234374, 0.03468185424804687, 0.03470336151123047, 0.03465523147583008, 0.03474534225463867, 0.034783233642578126, 0.03473100662231445, 0.03484979248046875, 0.03470848083496094, 0.03466854476928711, 0.03470131301879883, 0.03468185424804687, 0.0347586555480957, 0.03467161560058594, 0.034756607055664065, 0.034799617767333986, 0.0348109130859375, 0.034726879119873044, 0.0347955207824707, 0.034723838806152346, 0.03481702423095703, 0.03479040145874023, 0.03481497573852539, 0.03491328048706055, 0.03482726287841797, 0.03472588729858399, 0.03487744140625, 0.03488051223754883, 0.03499212646484375, 0.034769920349121096, 0.03505254364013672, 0.035093505859375, 0.034988033294677735, 0.03479449462890625, 0.03488051223754883, 0.034783233642578126, 0.035079166412353514, 0.03507814407348633, 0.03502284622192383, 0.034885631561279294, 0.03491635131835937, 0.03481087875366211, 0.03486207962036133, 0.0347852783203125, 0.03503513717651367, 0.03540383911132813, 0.035403713226318356, 0.03496550369262695, 0.034907135009765625, 0.03499929428100586, 0.03503308868408203, 0.03493686294555664, 0.03493475341796875, 0.03485184097290039, 0.034941951751708986, 0.12551273345947267, 0.03473507308959961, 0.034715648651123046, 0.034612224578857424, 0.0346798095703125, 0.034661376953125, 0.03468185424804687, 0.03479347229003906, 0.034713600158691404, 0.03471772766113281, 0.034717758178710936, 0.03465001678466797, 0.03472895812988281, 0.03467779159545899, 0.03470230484008789, 0.03476377487182617, 0.03481087875366211, 0.034699264526367186, 0.03474431991577148, 0.03471974563598633, 0.03479040145874023, 0.03469311904907227, 0.03477913665771484, 0.03542937469482422, 0.03522662353515625, 0.03483443069458008, 0.03482726287841797, 0.03500646209716797, 0.03478937530517578, 0.034909183502197266, 0.03505152130126953, 0.035033153533935546, 0.035034046173095704, 0.03484979248046875, 0.03483443069458008, 0.034800640106201174, 0.03502592086791992, 0.0347883529663086, 0.03484262466430664, 0.03506073760986328, 0.03504230499267578, 0.035076095581054685, 0.03481708908081055, 0.034758590698242185, 0.035225601196289064, 0.03481190490722656, 0.0349532470703125, 0.03495113754272461, 0.03489177703857422, 0.03486207962036133, 0.034945022583007815, 0.03486105728149414, 0.035056640625, 0.03486207962036133, 0.03486515045166016, 0.034887680053710936, 0.03490816116333008, 0.03487334442138672, 0.034933761596679686, 0.03485084915161133, 0.03488355255126953, 0.034852863311767575, 0.03487948989868164, 0.1255731201171875, 0.03466243362426758, 0.034718753814697266, 0.03467257690429688, 0.0346951675415039, 0.03467366409301758, 0.034677761077880856, 0.03476172637939453, 0.03475251388549805, 0.034625537872314455, 0.034683902740478514, 0.03469107055664063, 0.034726913452148435, 0.03467366409301758, 0.0347770881652832, 0.03484467315673828, 0.03479142379760742, 0.034813953399658204, 0.03479449462890625, 0.03477401733398437, 0.03475251388549805, 0.0351016960144043, 0.03500236892700195, 0.03482419204711914, 0.034718719482421875, 0.03472793579101562, 0.03482316970825195, 0.03473408126831055, 0.03477811050415039, 0.03469209671020508, 0.0348221435546875, 0.03484467315673828, 0.034772991180419925, 0.03480575942993164, 0.03481804656982422, 0.03478940963745117, 0.03485897445678711, 0.03477503967285156, 0.03490508651733398, 0.03496857452392578, 0.034887680053710936, 0.034799678802490235, 0.03488147354125976, 0.03484569549560547, 0.03491635131835937, 0.03486822509765625, 0.0349306869506836, 0.03492659378051758, 0.034895870208740236, 0.03488358306884766, 0.0349409294128418, 0.03492147064208984, 0.034890750885009765, 0.03487539291381836, 0.03492761611938477, 0.03498086547851562, 0.03492454528808594, 0.034909183502197266, 0.03498700714111328, 0.034857982635498046, 0.034923519134521484, 0.034841598510742186, 0.034928638458251955, 0.1257523193359375, 0.035064830780029296, 0.03537510299682617, 0.035378177642822264, 0.03510784149169922, 0.03473408126831055, 0.034738174438476564, 0.03501567840576172, 0.03498188781738281, 0.03481702423095703, 0.03481190490722656, 0.034732032775878906, 0.03487539291381836, 0.03477503967285156, 0.03500236892700195, 0.034977790832519534, 0.034956287384033204, 0.03479654312133789, 0.03480889511108398, 0.03498284912109375, 0.03504127883911133, 0.03495935821533203, 0.034874366760253905, 0.03476889419555664, 0.03484672164916992, 0.03483238220214844, 0.03480780792236328, 0.03479347229003906, 0.034842655181884764, 0.03487535858154297, 0.03479449462890625, 0.03484467315673828, 0.03486310577392578, 0.034890750885009765, 0.034974720001220705, 0.03501260757446289, 0.034955265045166016, 0.0349194221496582, 0.03486105728149414, 0.034885631561279294, 0.034950145721435545, 0.03498092651367188, 0.034893760681152346, 0.03479244613647461, 0.034854911804199216, 0.03503513717651367, 0.0349306869506836, 0.034939903259277344, 0.03488256072998047, 0.03484467315673828, 0.03496550369262695, 0.034854911804199216, 0.0349306869506836, 0.0348671989440918, 0.034931713104248044, 0.034953216552734374, 0.034941951751708986, 0.034923519134521484, 0.035007488250732424, 0.034948097229003904, 0.03496044921875, 0.034936767578125, 0.034907135009765625]",tokens/s,27.605657220121714,,,main,False,False,,,,,,,,,,,,,,,,,,,,,,, float16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,microsoft/rho-math-1b-v0.1,microsoft/rho-math-1b-v0.1,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1238.130688,3041.394688,0.0,2394.947584,2279.417856,s,10,0.26880492782592774,0.026880492782592774,0.0008987526138610299,0.026595680236816407,0.026978752708435055,0.028274463939666743,0.029311032924652102,"[0.02957017517089844, 0.02656800079345703, 0.026624223709106446, 0.026498048782348634, 0.02669081687927246, 0.02653455924987793, 0.026623359680175783, 0.026660736083984376, 0.026516223907470705, 0.026518783569335937]",tokens/s,9523.63493000322,kWh,3.126741712721287e-07,1.7133035382704246e-07,1.2450208372857195e-06,1.7290253623848905e-06,tokens/kWh,148060291.982584,MB,1238.519808,3041.394688,0.0,2394.947584,2279.420416,s,10,15.5807939453125,1.55807939453125,0.005216512186615283,1.5575559692382812,1.5656029785156251,1.5660965942382812,1.5664914868164062,"[1.5609205322265625, 1.5654932861328126, 1.561662353515625, 1.5665902099609375, 1.557181884765625, 1.5550050048828126, 1.5579300537109375, 1.5519298095703125, 1.55321826171875, 1.550862548828125]",tokens/s,40.43439648911705,kWh,1.8012090795692592e-05,9.867705139837371e-06,3.8868872364914635e-05,6.674866830044458e-05,tokens/kWh,943839.0548322053,,s,629,15.792589826583871,0.025107456004107888,0.003264401894865061,0.024826879501342772,0.025016524124145508,0.025216838455200195,0.05177774688720703,"[0.024412160873413087, 0.02388684844970703, 0.02385817527770996, 0.023840768814086914, 0.023856128692626953, 0.02386944007873535, 0.023968767166137696, 0.023797760009765623, 0.023783424377441405, 0.023767040252685546, 0.023969791412353517, 0.024361984252929687, 0.023824384689331055, 0.024758272171020508, 0.024879104614257814, 0.025027584075927735, 0.02490163230895996, 0.024863744735717775, 0.024832000732421877, 0.024099872589111327, 0.024890335083007812, 0.024851455688476562, 0.02529280090332031, 0.025255935668945313, 0.025250816345214845, 0.024894464492797853, 0.026001407623291017, 0.026261503219604493, 0.025214975357055663, 0.02508083152770996, 0.024953855514526366, 0.024921087265014647, 0.025012224197387696, 0.025011199951171875, 0.025042943954467774, 0.024936447143554686, 0.024970239639282226, 0.02492620849609375, 0.025629695892333985, 0.025084928512573244, 0.024816640853881834, 0.02491596794128418, 0.024893440246582032, 0.02483404731750488, 0.02492313575744629, 0.024061952590942383, 0.024167423248291017, 0.024824832916259764, 0.024826879501342772, 0.02488319969177246, 0.025025535583496093, 0.02498150444030762, 0.025061376571655275, 0.02512179183959961, 0.02505523109436035, 0.024887296676635744, 0.024947711944580078, 0.02508185577392578, 0.024972288131713868, 0.024036352157592773, 0.024808448791503908, 0.023994367599487306, 0.052125694274902344, 0.02493440055847168, 0.024770559310913084, 0.024303615570068358, 0.024801279067993166, 0.02483404731750488, 0.02483404731750488, 0.024774688720703125, 0.024644575119018554, 0.02490572738647461, 0.024028160095214843, 0.024809471130371095, 0.024835071563720702, 0.024818687438964843, 0.024962047576904296, 0.0249487361907959, 0.02496614456176758, 0.02487603187561035, 0.02473574447631836, 0.024835071563720702, 0.024853504180908204, 0.025457664489746092, 0.024824832916259764, 0.028212223052978515, 0.025609216690063476, 0.02507980728149414, 0.02476851272583008, 0.024641536712646486, 0.024608768463134766, 0.024506368637084962, 0.02408857536315918, 0.024813568115234375, 0.024793088912963866, 0.024794111251831053, 0.024810495376586913, 0.02486579132080078, 0.024019968032836913, 0.024184831619262694, 0.02476851272583008, 0.02503987121582031, 0.024816640853881834, 0.024791072845458985, 0.024844255447387695, 0.023839744567871093, 0.025226240158081056, 0.025016319274902343, 0.02503987121582031, 0.024958976745605467, 0.024790016174316407, 0.024827903747558593, 0.024738815307617186, 0.02472652816772461, 0.02453606414794922, 0.02494259262084961, 0.02487500762939453, 0.024927232742309572, 0.024946687698364257, 0.024151039123535157, 0.02488832092285156, 0.024975360870361327, 0.02492416000366211, 0.024902656555175783, 0.02490982437133789, 0.05210521697998047, 0.024751136779785157, 0.024843231201171875, 0.024861696243286133, 0.024839168548583986, 0.02486783981323242, 0.02497433662414551, 0.02495692825317383, 0.024007680892944337, 0.024584224700927734, 0.024926176071166994, 0.024943616867065428, 0.024954912185668945, 0.024878047943115236, 0.024885248184204102, 0.024845312118530274, 0.024589311599731444, 0.024983552932739257, 0.025022464752197264, 0.02446233558654785, 0.024020992279052734, 0.024847360610961915, 0.02391142463684082, 0.024878080368041993, 0.024819711685180663, 0.02490163230895996, 0.024793088912963866, 0.024999935150146483, 0.024837120056152344, 0.024506368637084962, 0.025018367767333984, 0.024065023422241212, 0.02465279960632324, 0.024007680892944337, 0.02447257614135742, 0.02495078468322754, 0.024859647750854492, 0.02488422393798828, 0.024995840072631836, 0.025009151458740234, 0.02496512031555176, 0.024862720489501954, 0.024863744735717775, 0.02487500762939453, 0.024945663452148437, 0.025001983642578125, 0.02493440055847168, 0.023961599349975587, 0.02487603187561035, 0.025052160263061524, 0.02493337631225586, 0.024839168548583986, 0.025028608322143556, 0.024869888305664063, 0.024869888305664063, 0.025030656814575194, 0.02490777587890625, 0.024962047576904296, 0.02484940719604492, 0.02489241600036621, 0.024919040679931642, 0.024978431701660156, 0.024213504791259766, 0.05179296112060547, 0.02481657600402832, 0.024852479934692383, 0.02490675163269043, 0.024778783798217772, 0.024894432067871095, 0.02482896041870117, 0.025025503158569336, 0.024422399520874022, 0.024992767333984374, 0.024378368377685547, 0.024847360610961915, 0.02487500762939453, 0.024999935150146483, 0.024856576919555663, 0.024987648010253907, 0.025062400817871092, 0.02488217544555664, 0.025001983642578125, 0.025218080520629883, 0.025987039566040038, 0.025017343521118163, 0.024895488739013674, 0.024832000732421877, 0.02492313575744629, 0.02488422393798828, 0.024773632049560547, 0.024792095184326172, 0.024828895568847657, 0.02488422393798828, 0.02498252868652344, 0.02491494369506836, 0.024781824111938477, 0.023945215225219727, 0.024795135498046874, 0.024893440246582032, 0.024928255081176756, 0.024847360610961915, 0.02492416000366211, 0.024999935150146483, 0.024879104614257814, 0.024944639205932616, 0.024848384857177733, 0.024847360610961915, 0.02493337631225586, 0.024828927993774414, 0.024806400299072266, 0.024819711685180663, 0.024583168029785156, 0.024525823593139647, 0.02484121513366699, 0.024802303314208983, 0.024800256729125978, 0.02473574447631836, 0.02492006492614746, 0.02490777587890625, 0.024904703140258787, 0.02487295913696289, 0.024845312118530274, 0.024771583557128905, 0.02484121513366699, 0.024809471130371095, 0.024800256729125978, 0.05202742385864258, 0.02481865692138672, 0.024795135498046874, 0.024838144302368165, 0.024809471130371095, 0.02492006492614746, 0.02475116729736328, 0.024717248916625977, 0.024571903228759767, 0.02395136070251465, 0.02489241600036621, 0.02494156837463379, 0.02490880012512207, 0.024998912811279295, 0.0249487361907959, 0.024951808929443358, 0.0249487361907959, 0.024871936798095705, 0.02486579132080078, 0.02527846336364746, 0.026267648696899414, 0.02528358459472656, 0.02490777587890625, 0.02505625534057617, 0.02488012886047363, 0.02488217544555664, 0.024838144302368165, 0.024763391494750975, 0.024772607803344726, 0.024623104095458984, 0.024543231964111328, 0.0248985595703125, 0.024559616088867187, 0.024816640853881834, 0.0248156795501709, 0.02498963165283203, 0.024824832916259764, 0.02491187286376953, 0.024800256729125978, 0.024913919448852538, 0.0248668155670166, 0.024851455688476562, 0.024754175186157225, 0.024798208236694336, 0.024777727127075197, 0.024792064666748048, 0.024770559310913084, 0.0247511043548584, 0.024801279067993166, 0.024817663192749022, 0.02470809555053711, 0.0244715518951416, 0.023734272003173826, 0.02390732765197754, 0.02379884719848633, 0.02382636833190918, 0.023888896942138672, 0.02392678451538086, 0.023868415832519533, 0.023840768814086914, 0.023931903839111326, 0.02504806327819824, 0.024816640853881834, 0.0512542724609375, 0.02474496078491211, 0.025034751892089844, 0.024895488739013674, 0.02386124801635742, 0.02387353515625, 0.024439807891845702, 0.024792064666748048, 0.024861696243286133, 0.024916032791137695, 0.024928192138671874, 0.02489139175415039, 0.024839168548583986, 0.024826879501342772, 0.024833023071289064, 0.024902656555175783, 0.02494156837463379, 0.024830976486206056, 0.024808448791503908, 0.024807424545288087, 0.024821760177612305, 0.024854528427124024, 0.024800256729125978, 0.024976383209228514, 0.024868864059448242, 0.024947711944580078, 0.02490777587890625, 0.024754175186157225, 0.024812543869018554, 0.024769535064697267, 0.02479315185546875, 0.024920000076293945, 0.02485862350463867, 0.024844287872314453, 0.024847360610961915, 0.024885248184204102, 0.025118719100952147, 0.02490777587890625, 0.02487603187561035, 0.024928255081176756, 0.024792064666748048, 0.02472550392150879, 0.02489036750793457, 0.024799232482910157, 0.024766464233398438, 0.024715263366699217, 0.024800256729125978, 0.024846336364746095, 0.02476748847961426, 0.02391347122192383, 0.024555519104003908, 0.024863744735717775, 0.024750080108642578, 0.02498252868652344, 0.02477670478820801, 0.02468454360961914, 0.023827455520629884, 0.02388582420349121, 0.023963647842407225, 0.023823360443115234, 0.02390630340576172, 0.023831552505493164, 0.023932928085327147, 0.051848190307617184, 0.023780351638793946, 0.024373247146606446, 0.0246824951171875, 0.024754175186157225, 0.02472038459777832, 0.024748031616210937, 0.024707071304321288, 0.0247511043548584, 0.024810495376586913, 0.024046592712402344, 0.024801279067993166, 0.024808448791503908, 0.02473472023010254, 0.02487094306945801, 0.024796127319335937, 0.02490982437133789, 0.024740863800048828, 0.02468351936340332, 0.024800256729125978, 0.02488217544555664, 0.024839168548583986, 0.024791040420532227, 0.024854528427124024, 0.024754175186157225, 0.02472243118286133, 0.024854528427124024, 0.024763391494750975, 0.024791040420532227, 0.024845312118530274, 0.024952831268310546, 0.023967744827270508, 0.023805952072143553, 0.02393600082397461, 0.024805376052856445, 0.024887296676635744, 0.024847360610961915, 0.02448896026611328, 0.02412031936645508, 0.024836095809936523, 0.024919040679931642, 0.02490163230895996, 0.02476748847961426, 0.024456192016601562, 0.023806976318359374, 0.023984128952026368, 0.024790016174316407, 0.026219520568847656, 0.025365503311157226, 0.024988672256469727, 0.025042943954467774, 0.024845312118530274, 0.02472857666015625, 0.025008127212524413, 0.0249036808013916, 0.02486783981323242, 0.024947711944580078, 0.024778751373291014, 0.02512281608581543, 0.02507980728149414, 0.024796159744262695, 0.024844287872314453, 0.02428108787536621, 0.05197516632080078, 0.024771583557128905, 0.02471014404296875, 0.024820735931396484, 0.02472857666015625, 0.024707071304321288, 0.024650751113891603, 0.02489753532409668, 0.02468454360961914, 0.024788991928100586, 0.024731647491455077, 0.023900159835815428, 0.02373222351074219, 0.023639039993286134, 0.023879680633544922, 0.024001535415649415, 0.024008703231811524, 0.023796735763549806, 0.023875583648681642, 0.023764991760253908, 0.023948287963867186, 0.023787519454956055, 0.02351411247253418, 0.023626752853393555, 0.024016895294189454, 0.024758272171020508, 0.024790016174316407, 0.02473369598388672, 0.024796159744262695, 0.02475315284729004, 0.024811519622802734, 0.024815616607666017, 0.024822784423828126, 0.024810495376586913, 0.024972288131713868, 0.02484121513366699, 0.0249036808013916, 0.024414239883422853, 0.024735712051391603, 0.025019392013549805, 0.025440256118774415, 0.024836095809936523, 0.02488832092285156, 0.02490060806274414, 0.02587750434875488, 0.02549247932434082, 0.02492620849609375, 0.024797183990478516, 0.024819711685180663, 0.02492416000366211, 0.02484940719604492, 0.02488630485534668, 0.024916959762573243, 0.024824832916259764, 0.02490060806274414, 0.024769535064697267, 0.024740863800048828, 0.024771583557128905, 0.02485660743713379, 0.02479203224182129, 0.024893440246582032, 0.024833023071289064, 0.024773632049560547, 0.05281587219238281, 0.024863744735717775, 0.024772607803344726, 0.024766464233398438, 0.024763391494750975, 0.024827903747558593, 0.024795135498046874, 0.02488319969177246, 0.02472857666015625, 0.024851455688476562, 0.024689664840698244, 0.024748031616210937, 0.024979455947875977, 0.024801279067993166, 0.024864767074584963, 0.024928287506103517, 0.024886240005493165, 0.024927232742309572, 0.024937471389770507, 0.024754175186157225, 0.024666112899780275, 0.02472755241394043, 0.02412646484375, 0.02432204818725586, 0.024697856903076174, 0.02493235206604004, 0.02508902359008789, 0.02387046432495117, 0.023801855087280274, 0.023829504013061522, 0.024846336364746095, 0.024953855514526366, 0.02486070442199707, 0.02477667236328125, 0.024755199432373046, 0.024793088912963866, 0.024863744735717775, 0.024861696243286133, 0.024802303314208983, 0.024521728515625, 0.023559167861938478, 0.023641088485717773, 0.023735296249389647, 0.023802879333496094, 0.023868415832519533, 0.023742464065551756, 0.023763967514038087, 0.024792064666748048, 0.024763423919677733, 0.024796127319335937, 0.024770559310913084, 0.024862720489501954, 0.02473472023010254, 0.024938495635986328, 0.024855552673339845, 0.02484121513366699, 0.024747007369995116, 0.024805408477783203, 0.024812511444091797, 0.024819711685180663, 0.02486579132080078, 0.024791040420532227, 0.024771583557128905, 0.051738624572753904, 0.02491494369506836, 0.024739839553833007, 0.02484121513366699, 0.024792064666748048, 0.02488934326171875, 0.02471833610534668, 0.02473779106140137, 0.024693792343139648, 0.024723424911499023, 0.024671232223510742, 0.024799232482910157, 0.02469068717956543, 0.024761344909667967, 0.02490060806274414, 0.024809471130371095, 0.024826879501342772, 0.024854528427124024, 0.02385305595397949, 0.023837696075439452, 0.02384588813781738, 0.02391244888305664, 0.02390425682067871, 0.024458240509033204, 0.024790016174316407, 0.024835071563720702, 0.024810495376586913, 0.025235456466674806, 0.02491801643371582, 0.024899616241455078, 0.024842208862304687, 0.024863744735717775, 0.024778751373291014, 0.024197120666503907, 0.023946239471435548, 0.02388275146484375, 0.0238919677734375, 0.024006656646728516, 0.023948320388793944, 0.0246015682220459, 0.025985023498535157, 0.025234432220458985, 0.024904703140258787, 0.02477670478820801, 0.024785951614379884, 0.02492720031738281, 0.02492313575744629, 0.02493337631225586, 0.02483404731750488, 0.024913919448852538, 0.024786943435668944, 0.024894464492797853, 0.024829952239990235, 0.02482585525512695, 0.024787967681884765, 0.024861696243286133, 0.025195520401000978, 0.023906368255615235, 0.024120256423950194, 0.023986175537109376, 0.023991296768188477, 0.023818239212036133, 0.023834623336791993]",tokens/s,39.82880622538531,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-2.7b,EleutherAI/pythia-2.7b,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1315.725312,6797.39392,0.0,6150.946816,5989.292032,s,10,0.4270291862487793,0.04270291862487793,0.0014589085893857718,0.042228033065795895,0.0427962459564209,0.044935611915588376,0.046647104682922366,"[0.04707497787475586, 0.04216880035400391, 0.042320831298828125, 0.04214396667480469, 0.04211011123657227, 0.04223136138916016, 0.04222470474243164, 0.04216451263427735, 0.04228224182128906, 0.04230767822265625]",tokens/s,5994.90639618387,kWh,5.012936111392275e-07,2.746845990513667e-07,2.6209189988051623e-06,3.3968972089957568e-06,tokens/kWh,75362892.73695235,MB,1315.725312,6797.39392,0.0,6150.946816,6115.763712,s,10,16.56944055175781,1.6569440551757815,0.022761192137948933,1.65922216796875,1.6744492065429688,1.6892416931152343,1.7010756823730468,"[1.6711619873046875, 1.652427978515625, 1.64447705078125, 1.6204678955078125, 1.641165771484375, 1.666016357421875, 1.631757080078125, 1.6691287841796876, 1.7040341796875, 1.668803466796875]",tokens/s,38.02180272967422,kWh,1.9981832101997237e-05,1.095027865999512e-05,6.577088571459451e-05,9.670299647658683e-05,tokens/kWh,651479.2953210421,,s,629,16.920467454910273,0.026900584189046544,0.005115355429982433,0.02596659278869629,0.027069645309448245,0.02723287010192871,0.0685291537475586,"[0.02694655990600586, 0.026814464569091798, 0.02693120002746582, 0.02692915153503418, 0.026862592697143556, 0.026812416076660156, 0.02688102340698242, 0.02592563247680664, 0.02592870330810547, 0.025809919357299805, 0.02593484878540039, 0.025873407363891602, 0.025997312545776367, 0.02588057518005371, 0.0272936954498291, 0.026875904083251953, 0.02679193687438965, 0.027023359298706053, 0.027052032470703126, 0.026602495193481446, 0.027470848083496095, 0.02771251106262207, 0.02629631996154785, 0.0259420166015625, 0.025964544296264647, 0.02590105628967285, 0.02587648010253906, 0.025852928161621092, 0.02630451202392578, 0.0271779842376709, 0.027049983978271484, 0.026813440322875977, 0.02689023971557617, 0.026818559646606444, 0.025903104782104492, 0.025821184158325194, 0.026123296737670897, 0.02581808090209961, 0.025996288299560546, 0.02593791961669922, 0.02590105628967285, 0.025847808837890625, 0.02595737648010254, 0.025977855682373048, 0.02593280029296875, 0.025874431610107423, 0.025995264053344725, 0.02590822410583496, 0.026630144119262695, 0.026810367584228514, 0.026786815643310546, 0.026857471466064452, 0.026884096145629883, 0.026830848693847657, 0.0269434871673584, 0.026879999160766603, 0.02689023971557617, 0.02688921546936035, 0.026798080444335938, 0.02688719940185547, 0.02674787139892578, 0.026824703216552736, 0.06859264373779297, 0.02592153549194336, 0.02589388847351074, 0.02591334342956543, 0.025802751541137696, 0.025866239547729493, 0.025819135665893556, 0.025935871124267578, 0.025808895111083984, 0.025874431610107423, 0.025775104522705077, 0.025842687606811524, 0.025791488647460937, 0.025793535232543945, 0.025857023239135742, 0.026072063446044923, 0.02570751953125, 0.025594879150390625, 0.025646080017089845, 0.02594713592529297, 0.025881599426269532, 0.02592153549194336, 0.02588979148864746, 0.026398719787597655, 0.028299264907836914, 0.027207679748535156, 0.026867712020874023, 0.02592153549194336, 0.02590822410583496, 0.025850879669189454, 0.02594099235534668, 0.025985023498535157, 0.025841663360595703, 0.02588057518005371, 0.02592870330810547, 0.025830400466918944, 0.025862144470214843, 0.02595840072631836, 0.026690559387207033, 0.02690559959411621, 0.02676531219482422, 0.026825727462768553, 0.026770431518554686, 0.02589286422729492, 0.025810943603515626, 0.025931776046752928, 0.026738687515258788, 0.026828800201416016, 0.02674073600769043, 0.027128896713256835, 0.02589689636230469, 0.025825279235839844, 0.025780223846435548, 0.025834495544433594, 0.025785343170166015, 0.026878976821899415, 0.026855424880981447, 0.02687286376953125, 0.02719024085998535, 0.026784767150878908, 0.026712064743041993, 0.027131904602050783, 0.02714112091064453, 0.06956755065917969, 0.026673088073730467, 0.026673152923583986, 0.026679296493530274, 0.02636595153808594, 0.025854976654052734, 0.02572083282470703, 0.02594713592529297, 0.025746431350708008, 0.025843711853027345, 0.025828351974487306, 0.025785375595092773, 0.026262496948242186, 0.026654720306396484, 0.0267007999420166, 0.02692095947265625, 0.02674892807006836, 0.026465343475341796, 0.0262225284576416, 0.026558464050292968, 0.02673459243774414, 0.026690559387207033, 0.026746879577636717, 0.026636287689208983, 0.025838592529296874, 0.025783296585083007, 0.025797632217407225, 0.025766912460327147, 0.025956352233886718, 0.026884096145629883, 0.026847232818603517, 0.025993215560913087, 0.025831424713134765, 0.025790464401245116, 0.025757696151733397, 0.025540607452392578, 0.025511936187744142, 0.02573107147216797, 0.027313152313232423, 0.027067392349243165, 0.026814464569091798, 0.02681548881530762, 0.0266844482421875, 0.02558051109313965, 0.025776128768920898, 0.025883647918701173, 0.025811967849731447, 0.0259368953704834, 0.02578124809265137, 0.02555084800720215, 0.025475072860717773, 0.025480192184448244, 0.02549247932434082, 0.025638912200927736, 0.025520128250122072, 0.025618431091308593, 0.02549558448791504, 0.025612255096435547, 0.025795583724975587, 0.025830400466918944, 0.02572800064086914, 0.02557542419433594, 0.025623552322387694, 0.06838988494873047, 0.02593894386291504, 0.025807872772216797, 0.02592153549194336, 0.025564159393310547, 0.02567475128173828, 0.025779199600219727, 0.025879552841186523, 0.025808895111083984, 0.025842687606811524, 0.026496000289916992, 0.02671615982055664, 0.02671513557434082, 0.026298368453979492, 0.025491455078125, 0.02565020751953125, 0.02555695915222168, 0.025552896499633788, 0.02549247932434082, 0.02555801582336426, 0.025520128250122072, 0.025539583206176757, 0.02557542419433594, 0.025554943084716796, 0.02548121643066406, 0.02549862480163574, 0.025531391143798828, 0.025579519271850586, 0.025758720397949218, 0.02576383972167969, 0.02571878433227539, 0.025899007797241212, 0.02573516845703125, 0.025779232025146485, 0.02552009582519531, 0.026063871383666993, 0.025487360000610353, 0.025463808059692384, 0.025524255752563476, 0.025710559844970702, 0.025479167938232423, 0.02551910400390625, 0.02553036880493164, 0.025615392684936525, 0.025540576934814454, 0.025581567764282227, 0.025540607452392578, 0.02589286422729492, 0.025589759826660157, 0.025580543518066406, 0.025620479583740235, 0.025840639114379883, 0.025511936187744142, 0.025604095458984375, 0.0255416316986084, 0.025734144210815428, 0.025619455337524414, 0.025705472946166992, 0.025825279235839844, 0.0259051513671875, 0.025787391662597657, 0.02570035171508789, 0.02551807975769043, 0.0685506591796875, 0.02550476837158203, 0.025664512634277343, 0.025569280624389647, 0.02552115249633789, 0.025617408752441406, 0.02568806457519531, 0.02564409637451172, 0.02559993553161621, 0.025769983291625977, 0.025787391662597657, 0.025817087173461914, 0.026792959213256837, 0.026770431518554686, 0.025584640502929686, 0.026158079147338868, 0.025973760604858398, 0.025651199340820312, 0.025845760345458983, 0.025684991836547853, 0.025614336013793947, 0.025851903915405275, 0.025858112335205078, 0.025695167541503906, 0.02548940849304199, 0.02575667190551758, 0.02579769515991211, 0.02584876823425293, 0.025842687606811524, 0.02593075180053711, 0.026785791397094725, 0.026895360946655275, 0.026755071640014647, 0.02675916862487793, 0.026730495452880858, 0.026785791397094725, 0.026771455764770507, 0.02673459243774414, 0.026764287948608398, 0.02597068786621094, 0.025878528594970703, 0.025866239547729493, 0.025971712112426756, 0.025826303482055665, 0.025846784591674804, 0.025832447052001953, 0.025819135665893556, 0.025928768157958984, 0.025840576171875, 0.025788415908813478, 0.025597951889038087, 0.025581567764282227, 0.02565225601196289, 0.025778144836425782, 0.025803775787353517, 0.02593791961669922, 0.025857023239135742, 0.02697216033935547, 0.026828800201416016, 0.026635263442993166, 0.026833919525146483, 0.026702848434448243, 0.0267007999420166, 0.06968013000488281, 0.02679193687438965, 0.026746879577636717, 0.0267007999420166, 0.02667830467224121, 0.026762208938598632, 0.026888191223144533, 0.025874431610107423, 0.026598400115966796, 0.026867712020874023, 0.02682368087768555, 0.026849279403686522, 0.026829824447631836, 0.026720256805419923, 0.026764287948608398, 0.026878976821899415, 0.02668339157104492, 0.026630144119262695, 0.026707008361816408, 0.025767871856689453, 0.025794559478759766, 0.025866239547729493, 0.025852928161621092, 0.026612735748291014, 0.02651238441467285, 0.02636288070678711, 0.02571161651611328, 0.02650931167602539, 0.02667519950866699, 0.027024383544921874, 0.027020288467407227, 0.02590822410583496, 0.025855039596557616, 0.026547136306762694, 0.02659744071960449, 0.02680415916442871, 0.02674176025390625, 0.025858047485351563, 0.025808895111083984, 0.02613043212890625, 0.026385408401489258, 0.025630720138549806, 0.025635839462280274, 0.025806848526000976, 0.025865215301513672, 0.025806848526000976, 0.026462207794189452, 0.02707967948913574, 0.0267007999420166, 0.02590412712097168, 0.025796607971191408, 0.025805824279785155, 0.026727424621582032, 0.0267509765625, 0.02648575973510742, 0.025935871124267578, 0.025750528335571288, 0.027829248428344725, 0.028071935653686524, 0.02731007957458496, 0.026452991485595705, 0.02589388847351074, 0.02574950408935547, 0.068663330078125, 0.02576380729675293, 0.02587552070617676, 0.027017152786254883, 0.02676736068725586, 0.025822208404541015, 0.025842687606811524, 0.025911296844482422, 0.025780223846435548, 0.026096639633178712, 0.02676940727233887, 0.025817087173461914, 0.025810943603515626, 0.025772064208984376, 0.02580271911621094, 0.025891839981079103, 0.025775104522705077, 0.025798688888549803, 0.026255392074584962, 0.026112960815429687, 0.025795583724975587, 0.025827327728271485, 0.025834495544433594, 0.02593075180053711, 0.02571878433227539, 0.02592051124572754, 0.02574950408935547, 0.025787391662597657, 0.025769983291625977, 0.025845760345458983, 0.025784320831298828, 0.025859071731567384, 0.025789440155029295, 0.025832447052001953, 0.025784320831298828, 0.025821184158325194, 0.025965568542480468, 0.025835519790649415, 0.025863168716430664, 0.026294271469116212, 0.025779199600219727, 0.025794559478759766, 0.025886720657348632, 0.025823232650756835, 0.025765888214111327, 0.025849855422973633, 0.02588057518005371, 0.025879552841186523, 0.02572697639465332, 0.025853952407836913, 0.02576486396789551, 0.025772031784057618, 0.0257259521484375, 0.025819135665893556, 0.025766912460327147, 0.02591948890686035, 0.025829376220703124, 0.025890815734863282, 0.025848831176757812, 0.025783296585083007, 0.025791488647460937, 0.02573516845703125, 0.025790464401245116, 0.0684738540649414, 0.025795583724975587, 0.025782272338867186, 0.025835519790649415, 0.025785343170166015, 0.025772031784057618, 0.025669631958007814, 0.025894912719726562, 0.02574336051940918, 0.02572083282470703, 0.025800703048706054, 0.026085376739501953, 0.02576076889038086, 0.025757696151733397, 0.025823232650756835, 0.025960447311401368, 0.025867263793945314, 0.02608332824707031, 0.025776128768920898, 0.02590105628967285, 0.02678169631958008, 0.027848703384399414, 0.026252288818359375, 0.026055679321289063, 0.025830432891845702, 0.025797599792480468, 0.026068992614746093, 0.026242048263549804, 0.02652774429321289, 0.026717216491699218, 0.02640176010131836, 0.026194976806640624, 0.025761760711669923, 0.02657587242126465, 0.02678374481201172, 0.02670796775817871, 0.026853376388549805, 0.02675814437866211, 0.026731552124023436, 0.027158496856689453, 0.02679091262817383, 0.02673971176147461, 0.026711040496826172, 0.026714111328125, 0.02750464057922363, 0.027048959732055664, 0.02673766326904297, 0.02688921546936035, 0.02673971176147461, 0.026814464569091798, 0.02717184066772461, 0.027020288467407227, 0.027054079055786134, 0.027064319610595702, 0.027115520477294923, 0.027223039627075195, 0.02699776077270508, 0.027555871963500976, 0.027730911254882813, 0.027246591567993163, 0.027041791915893554, 0.02712166404724121, 0.026928127288818358, 0.06986956787109375, 0.027034624099731445, 0.026528799057006835, 0.02694243240356445, 0.027089920043945313, 0.02709503936767578, 0.027173887252807616, 0.027380735397338866, 0.027069440841674806, 0.026999807357788085, 0.02698956871032715, 0.027031551361083983, 0.026945600509643553, 0.027084735870361327, 0.027026432037353516, 0.02714521598815918, 0.02696499252319336, 0.026987520217895508, 0.02695475196838379, 0.026640384674072266, 0.026605567932128905, 0.027001855850219726, 0.027014144897460936, 0.027193344116210938, 0.027025407791137695, 0.02654719924926758, 0.02729471969604492, 0.02718617630004883, 0.027045888900756834, 0.027026432037353516, 0.027021312713623048, 0.0272936954498291, 0.028247039794921876, 0.027448320388793947, 0.02716876792907715, 0.0269752311706543, 0.027014144897460936, 0.027099136352539063, 0.02699776077270508, 0.027193344116210938, 0.027032575607299804, 0.027158527374267577, 0.02700492858886719, 0.027066368103027344, 0.027025407791137695, 0.026944511413574217, 0.027060224533081056, 0.027262975692749023, 0.027056127548217773, 0.026944511413574217, 0.02697216033935547, 0.02697932815551758, 0.02657177543640137, 0.026603519439697267, 0.02707046318054199, 0.027052032470703126, 0.02695577621459961, 0.027059200286865235, 0.027107328414916993, 0.0265533447265625, 0.027289600372314454, 0.027166719436645507, 0.027062271118164064, 0.06987881469726563, 0.02695881652832031, 0.026427391052246094, 0.027094079971313478, 0.026931135177612305, 0.02692915153503418, 0.026409984588623047, 0.025798656463623046, 0.025858047485351563, 0.025911296844482422, 0.02611712074279785, 0.02697113609313965, 0.02696601676940918, 0.026587135314941408, 0.026426368713378907, 0.027239423751831054, 0.02711347198486328, 0.026993663787841796, 0.026918912887573244, 0.02691379165649414, 0.026990591049194337, 0.0269117431640625, 0.02646937561035156, 0.026529792785644532, 0.0263505916595459, 0.026673152923583986, 0.026933248519897462, 0.027206655502319335, 0.02697318458557129, 0.02696294403076172, 0.027057151794433593, 0.027284511566162108, 0.02696700859069824, 0.027061248779296877, 0.026994688034057617, 0.026945535659790038, 0.02734182357788086, 0.02709708786010742, 0.0268984317779541, 0.027123712539672853, 0.02696396827697754, 0.02597068786621094, 0.025853952407836913, 0.025862144470214843, 0.025827327728271485, 0.025882623672485353, 0.02590105628967285, 0.026035200119018553, 0.025842687606811524, 0.025894912719726562, 0.025816064834594726, 0.025865215301513672, 0.025833471298217774, 0.025818111419677735, 0.025818111419677735, 0.02596659278869629, 0.025874431610107423, 0.025923583984375, 0.02567475128173828, 0.02553036880493164, 0.025878528594970703, 0.025968639373779297, 0.025935871124267578]",tokens/s,37.173913881289714,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen2-beta-72B,Qwen/Qwen2-beta-72B,cuda,0,42,,,True,,,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",qwen2,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-0.5B,,cuda,0,42,,,,,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,217063f5c507ed7cc255df7e1f64c4333a0b4dfe,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,,MB,1292.283904,2227.699712,0.0,1581.252608,1426.269184,s,10,0.22845740699768066,0.022845740699768066,0.0011668354389555138,0.022283743858337403,0.02394346523284912,0.02501357221603393,0.02586965780258179,"[0.02608367919921875, 0.022581151962280274, 0.02220899200439453, 0.02228976058959961, 0.023705663681030272, 0.02218716812133789, 0.022176256179809572, 0.022223424911499024, 0.022277727127075195, 0.022723583221435546]",tokens/s,11205.589845576727,kWh,2.658592888180832e-07,1.4567710569802357e-07,7.140124313842353e-07,1.125548825900342e-06,tokens/kWh,227444597.79007992,MB,1292.611584,2227.699712,0.0,1581.252608,1426.271744,s,10,13.1124912109375,1.31124912109375,0.003834843099448192,1.30967822265625,1.3173670166015625,1.3180671752929687,1.3186273022460937,"[1.318767333984375, 1.308579345703125, 1.310361572265625, 1.313892578125, 1.31721142578125, 1.31084912109375, 1.308844970703125, 1.306833251953125, 1.308994873046875, 1.30815673828125]",tokens/s,48.04579006882378,kWh,1.5466314647416497e-05,8.475307297489297e-06,2.7254148832016058e-05,5.1195770776921855e-05,tokens/kWh,1230570.3975922808,,s,629,13.290261499404894,0.021129191572980772,0.0027457968275008066,0.020719615936279297,0.02117140464782715,0.021482905578613282,0.04338389083862305,"[0.021635072708129883, 0.02092748832702637, 0.020876287460327148, 0.020685823440551757, 0.020777984619140624, 0.020763647079467772, 0.020904991149902345, 0.02072060775756836, 0.020641792297363282, 0.02061724853515625, 0.02072060775756836, 0.020708351135253905, 0.02059878349304199, 0.02067148780822754, 0.020677631378173827, 0.020668415069580077, 0.020677631378173827, 0.020692991256713866, 0.020761600494384767, 0.020747264862060546, 0.020784128189086915, 0.020699136734008788, 0.02068377685546875, 0.020743167877197266, 0.020706304550170897, 0.020717567443847656, 0.021566463470458985, 0.021763071060180664, 0.021797887802124022, 0.02148659133911133, 0.021235712051391603, 0.020745248794555665, 0.02071651268005371, 0.02068070411682129, 0.020709375381469726, 0.020742143630981445, 0.020772863388061523, 0.020752384185791017, 0.020709375381469726, 0.02065920066833496, 0.020668415069580077, 0.020769792556762694, 0.02064384078979492, 0.02071244812011719, 0.020858879089355468, 0.02069196891784668, 0.020632575988769532, 0.020785152435302736, 0.02072166442871094, 0.02147225570678711, 0.021319679260253906, 0.020766719818115235, 0.02060492706298828, 0.020649984359741212, 0.0206376953125, 0.02065100860595703, 0.02063871955871582, 0.020661279678344725, 0.02067452812194824, 0.020695104598999023, 0.020852672576904298, 0.023468032836914062, 0.04402073669433594, 0.020642816543579103, 0.020616191864013672, 0.020695039749145508, 0.020761600494384767, 0.021575679779052736, 0.020737024307250978, 0.02087014389038086, 0.020798463821411133, 0.020717567443847656, 0.020786176681518553, 0.02068992042541504, 0.020734975814819336, 0.02067046356201172, 0.020708351135253905, 0.020699136734008788, 0.020763647079467772, 0.02067251205444336, 0.020818944931030273, 0.020727807998657227, 0.020788223266601562, 0.020699136734008788, 0.02081177520751953, 0.020710399627685547, 0.02070528030395508, 0.020909055709838868, 0.020787200927734374, 0.020753408432006838, 0.020822015762329102, 0.02085478401184082, 0.020769792556762694, 0.02106060791015625, 0.020980735778808594, 0.020724735260009765, 0.021160959243774414, 0.020768768310546876, 0.020722688674926756, 0.020808704376220705, 0.020682815551757813, 0.02072876739501953, 0.02062848091125488, 0.0206878719329834, 0.02068070411682129, 0.02084454345703125, 0.020722688674926756, 0.020682752609252928, 0.02081996726989746, 0.020722688674926756, 0.020702207565307617, 0.020694015502929687, 0.020760576248168947, 0.0206878719329834, 0.02069708824157715, 0.020767744064331056, 0.020781055450439453, 0.020624383926391602, 0.02068070411682129, 0.020757503509521484, 0.02064384078979492, 0.020658176422119142, 0.020688928604125977, 0.020733919143676758, 0.020694015502929687, 0.04339507293701172, 0.020734975814819336, 0.02066431999206543, 0.02084454345703125, 0.02071244812011719, 0.020706304550170897, 0.02062848091125488, 0.020707328796386718, 0.020677631378173827, 0.020746240615844725, 0.020723712921142577, 0.02129408073425293, 0.021073919296264648, 0.020999168395996092, 0.0212992000579834, 0.02128998374938965, 0.020747264862060546, 0.02063667106628418, 0.02102374458312988, 0.020751359939575196, 0.020736000061035157, 0.020719615936279297, 0.02086809539794922, 0.020682752609252928, 0.020711423873901368, 0.0207391357421875, 0.02073798370361328, 0.020908031463623047, 0.020754432678222655, 0.020733951568603515, 0.020699136734008788, 0.020779008865356444, 0.02068992042541504, 0.020733951568603515, 0.020776960372924806, 0.020722688674926756, 0.020760576248168947, 0.021448703765869142, 0.02087424087524414, 0.02068377685546875, 0.020736000061035157, 0.020783103942871094, 0.02087424087524414, 0.020739072799682616, 0.020720640182495118, 0.020708351135253905, 0.020719615936279297, 0.020750335693359375, 0.020694015502929687, 0.02090188789367676, 0.020941823959350587, 0.02066534423828125, 0.020755456924438476, 0.020706304550170897, 0.020682752609252928, 0.02070425605773926, 0.02067251205444336, 0.020864000320434572, 0.020722688674926756, 0.02063871955871582, 0.02066739273071289, 0.020806655883789063, 0.020734975814819336, 0.0434595832824707, 0.020619264602661135, 0.02091110420227051, 0.020763647079467772, 0.020706304550170897, 0.02068172836303711, 0.020653055191040038, 0.020642816543579103, 0.0206561279296875, 0.02064793586730957, 0.020669439315795898, 0.02065407943725586, 0.020686847686767578, 0.02063871955871582, 0.020752384185791017, 0.021515264511108398, 0.02066431999206543, 0.020686847686767578, 0.020771839141845702, 0.020740095138549804, 0.020787200927734374, 0.0208855037689209, 0.020786176681518553, 0.020724735260009765, 0.020806655883789063, 0.020733951568603515, 0.020743167877197266, 0.020792320251464845, 0.020685823440551757, 0.020690944671630858, 0.02069811248779297, 0.02147737693786621, 0.022746112823486327, 0.021777408599853516, 0.021429248809814453, 0.02142310333251953, 0.020750335693359375, 0.020683807373046877, 0.020765663146972656, 0.02065920066833496, 0.02068172836303711, 0.020701183319091796, 0.020860960006713867, 0.02072060775756836, 0.020773887634277344, 0.020831232070922853, 0.020990976333618162, 0.020726783752441406, 0.020726783752441406, 0.020771839141845702, 0.020957183837890626, 0.02072166442871094, 0.020710399627685547, 0.02069606399536133, 0.020701183319091796, 0.020749343872070312, 0.021204959869384764, 0.020743167877197266, 0.020742143630981445, 0.020790271759033203, 0.020714496612548827, 0.02084864044189453, 0.021172224044799806, 0.04560182571411133, 0.020725727081298828, 0.020707328796386718, 0.020728832244873048, 0.02110873603820801, 0.02146713638305664, 0.020751359939575196, 0.02082099151611328, 0.02085683250427246, 0.020970495223999023, 0.020769792556762694, 0.020714496612548827, 0.020758527755737305, 0.02081177520751953, 0.020742143630981445, 0.02069606399536133, 0.020750335693359375, 0.020809728622436522, 0.020761600494384767, 0.020714559555053712, 0.02081171226501465, 0.020761600494384767, 0.020758527755737305, 0.020684799194335936, 0.020832256317138673, 0.02088860893249512, 0.020798463821411133, 0.020745183944702147, 0.020759552001953126, 0.02126233673095703, 0.020971519470214844, 0.02089472007751465, 0.02079641532897949, 0.02069606399536133, 0.020883455276489257, 0.020815872192382814, 0.020930559158325195, 0.020855808258056642, 0.02085478401184082, 0.020714496612548827, 0.02067865562438965, 0.020735008239746094, 0.02070217514038086, 0.021810176849365235, 0.021212223052978516, 0.02072159957885742, 0.020703231811523438, 0.02067967987060547, 0.020658176422119142, 0.02066534423828125, 0.020719615936279297, 0.02068070411682129, 0.02067046356201172, 0.02121625518798828, 0.02270207977294922, 0.021942272186279296, 0.02109542465209961, 0.0208855037689209, 0.02066537666320801, 0.02112816047668457, 0.021287935256958008, 0.0208035831451416, 0.02067967987060547, 0.04335513687133789, 0.02064896011352539, 0.020695039749145508, 0.020701183319091796, 0.021019647598266602, 0.020726783752441406, 0.02064076805114746, 0.020711423873901368, 0.020599807739257812, 0.020686847686767578, 0.02062233543395996, 0.020714559555053712, 0.020706239700317382, 0.021015552520751952, 0.020624383926391602, 0.020678688049316406, 0.02068988800048828, 0.020720640182495118, 0.021614591598510743, 0.02088140869140625, 0.020642816543579103, 0.021337087631225587, 0.020726783752441406, 0.020632575988769532, 0.02069811248779297, 0.020688928604125977, 0.020846559524536134, 0.021163007736206055, 0.020750335693359375, 0.020720640182495118, 0.020718591690063477, 0.020764671325683593, 0.020677631378173827, 0.020760576248168947, 0.020799488067626954, 0.021243904113769533, 0.021769216537475586, 0.021540864944458008, 0.020913152694702147, 0.020694015502929687, 0.020743167877197266, 0.020726783752441406, 0.020612096786499022, 0.02062233543395996, 0.0206376953125, 0.020577280044555665, 0.02064691162109375, 0.020624383926391602, 0.020753408432006838, 0.02079539108276367, 0.020724767684936522, 0.02070012855529785, 0.020701183319091796, 0.020754432678222655, 0.020722688674926756, 0.020749311447143554, 0.020727807998657227, 0.020926464080810548, 0.021121023178100586, 0.020695039749145508, 0.020773887634277344, 0.020798463821411133, 0.02081996726989746, 0.04332339096069336, 0.02192793655395508, 0.022320127487182616, 0.02130745506286621, 0.02075539207458496, 0.02067356872558594, 0.020776927947998045, 0.020749311447143554, 0.020707328796386718, 0.020649984359741212, 0.021081087112426757, 0.021355520248413085, 0.02045849609375, 0.020587520599365236, 0.020513792037963868, 0.020563968658447264, 0.020552703857421875, 0.020690944671630858, 0.020671552658081054, 0.020661184310913086, 0.02066739273071289, 0.020730880737304686, 0.02046668815612793, 0.020686847686767578, 0.020768768310546876, 0.020624383926391602, 0.020694015502929687, 0.020674560546875, 0.02065407943725586, 0.020685823440551757, 0.02065100860595703, 0.020736000061035157, 0.020769792556762694, 0.020823040008544923, 0.02067251205444336, 0.020658176422119142, 0.020661247253417968, 0.02064076805114746, 0.020757503509521484, 0.020626432418823244, 0.02063052749633789, 0.02061008071899414, 0.020661216735839844, 0.020676607131958007, 0.0207227840423584, 0.021127071380615235, 0.02090291213989258, 0.020774911880493165, 0.020634624481201173, 0.020682752609252928, 0.020753408432006838, 0.020738048553466795, 0.020671552658081054, 0.020666303634643553, 0.020642816543579103, 0.020666368484497072, 0.020722688674926756, 0.02067148780822754, 0.02067967987060547, 0.020580352783203124, 0.020624383926391602, 0.020644863128662108, 0.021565439224243164, 0.04346992111206055, 0.020659103393554687, 0.020684799194335936, 0.02068070411682129, 0.020728832244873048, 0.02063155174255371, 0.020711423873901368, 0.020759552001953126, 0.02070528030395508, 0.020666368484497072, 0.020568063735961914, 0.021062719345092774, 0.02063046455383301, 0.02068377685546875, 0.02065100860595703, 0.020682752609252928, 0.02065407943725586, 0.02045030403137207, 0.020653055191040038, 0.020750335693359375, 0.020722688674926756, 0.02066329574584961, 0.02144358444213867, 0.02105241584777832, 0.020699136734008788, 0.020719615936279297, 0.020545536041259766, 0.020720640182495118, 0.020777984619140624, 0.021196800231933592, 0.021387264251708983, 0.020744192123413087, 0.020772863388061523, 0.02071244812011719, 0.02079539108276367, 0.020814847946166993, 0.020731903076171874, 0.020899839401245117, 0.020756479263305663, 0.020685823440551757, 0.020707359313964845, 0.020652000427246093, 0.02066431999206543, 0.020716543197631835, 0.02066227149963379, 0.02064896011352539, 0.020732927322387695, 0.02070528030395508, 0.020694015502929687, 0.020720640182495118, 0.020703231811523438, 0.020730880737304686, 0.02066431999206543, 0.020783103942871094, 0.020790271759033203, 0.02067251205444336, 0.02063052749633789, 0.02082713508605957, 0.02068377685546875, 0.020641792297363282, 0.02063871955871582, 0.020686847686767578, 0.02043503952026367, 0.04344412612915039, 0.020682752609252928, 0.02064384078979492, 0.020717567443847656, 0.020668415069580077, 0.02067251205444336, 0.020624383926391602, 0.02065203285217285, 0.02061516761779785, 0.02065100860595703, 0.02046156883239746, 0.020427776336669923, 0.02067046356201172, 0.020715520858764647, 0.02069708824157715, 0.021188608169555666, 0.02065100860595703, 0.02068172836303711, 0.020798463821411133, 0.02083839988708496, 0.020711423873901368, 0.02068172836303711, 0.02065203285217285, 0.02058956718444824, 0.02065920066833496, 0.020715520858764647, 0.020624383926391602, 0.020724735260009765, 0.02066329574584961, 0.02063155174255371, 0.020620288848876952, 0.022112255096435548, 0.021897216796875, 0.021122047424316406, 0.02082211112976074, 0.020707231521606445, 0.020700159072875975, 0.021617664337158202, 0.020822080612182617, 0.02076460838317871, 0.02067865562438965, 0.02063158416748047, 0.020653024673461914, 0.021243904113769533, 0.02068172836303711, 0.020661247253417968, 0.020709375381469726, 0.02068889617919922, 0.020694015502929687, 0.020709375381469726, 0.02069606399536133, 0.020756479263305663, 0.020702207565307617, 0.020779008865356444, 0.020757503509521484, 0.020909055709838868, 0.020758527755737305, 0.020676607131958007, 0.020716543197631835, 0.02067046356201172, 0.020702207565307617, 0.020934656143188478, 0.020567039489746093, 0.043998207092285156, 0.02067967987060547, 0.020823040008544923, 0.021171199798583985, 0.02106985664367676, 0.02070729637145996, 0.020738048553466795, 0.02068172836303711, 0.020701183319091796, 0.02066534423828125, 0.020737024307250978, 0.020708351135253905, 0.02067046356201172, 0.020672544479370118, 0.02068783950805664, 0.02070528030395508, 0.02065715217590332, 0.020644863128662108, 0.020695072174072265, 0.02067555236816406, 0.02070425605773926, 0.02083737564086914, 0.02069606399536133, 0.02063260841369629, 0.020738048553466795, 0.020673503875732423, 0.020754432678222655, 0.020569087982177735, 0.02043801689147949, 0.02043084716796875, 0.02046463966369629, 0.020741119384765624, 0.020743167877197266, 0.020487167358398437, 0.020445184707641603, 0.020463615417480468, 0.02064691162109375, 0.020618240356445314, 0.02045337677001953, 0.02061516761779785, 0.02061516761779785, 0.02064384078979492, 0.020719615936279297, 0.02105855941772461, 0.021343231201171875, 0.02142617607116699, 0.02063871955871582, 0.02063974380493164, 0.02065203285217285, 0.02087014389038086, 0.02064588737487793, 0.02151219177246094, 0.020706304550170897, 0.020709375381469726, 0.0210565128326416, 0.02074015998840332, 0.02071443176269531, 0.020706304550170897, 0.02068889617919922, 0.021211135864257814, 0.021339136123657225, 0.02125209617614746, 0.020779008865356444]",tokens/s,47.32788741802895,,,main,False,False,,,,,,,,,,,,,,,,,,,,,,, float16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-1.8B,,cuda,0,42,,,,,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,217063f5c507ed7cc255df7e1f64c4333a0b4dfe,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,,MB,1299.443712,5222.432768,0.0,4575.985664,4408.404992,s,10,0.2515818881988526,0.02515818881988526,0.0004316832491674847,0.02500271987915039,0.025320387840270995,0.025869905948638916,0.02630952043533325,"[0.026419424057006837, 0.02496672058105469, 0.025182367324829102, 0.02489097595214844, 0.024919551849365236, 0.024940799713134766, 0.024999679565429686, 0.025005760192871093, 0.02505833625793457, 0.025198272705078125]",tokens/s,10175.613269809603,kWh,2.9668963177892634e-07,1.6257227317216066e-07,1.3936045679599308e-06,1.8528664729110178e-06,tokens/kWh,138164300.41923165,MB,1299.443712,5222.432768,0.0,4575.985664,4408.407552,s,10,13.746141967773436,1.3746141967773435,0.012068877601918208,1.3695853881835938,1.3918320556640624,1.3927660034179687,1.3935131616210938,"[1.39162451171875, 1.3634466552734374, 1.368049560546875, 1.3711212158203125, 1.361060302734375, 1.363426025390625, 1.3650748291015624, 1.380155029296875, 1.38848388671875, 1.393699951171875]",tokens/s,45.83104128249054,kWh,1.6011149473747676e-05,8.771077954546157e-06,4.0418401103841396e-05,6.520062853213523e-05,tokens/kWh,966248.3540162406,,s,629,13.947795446395862,0.022174555558657986,0.0030496775723597627,0.02167193603515625,0.022245785522460936,0.02251919288635254,0.046950276489257814,"[0.02288640022277832, 0.02246246337890625, 0.022433792114257813, 0.02268671989440918, 0.022524927139282228, 0.022582271575927734, 0.0224849910736084, 0.02244095993041992, 0.022395904541015626, 0.022575103759765625, 0.023621631622314454, 0.02288332748413086, 0.02253107261657715, 0.022656000137329102, 0.022479936599731444, 0.022508480072021483, 0.02260383987426758, 0.02281465530395508, 0.02229862403869629, 0.022204416275024414, 0.02211123275756836, 0.022552576065063477, 0.02287820816040039, 0.02267852783203125, 0.022592512130737305, 0.022412288665771486, 0.022122495651245116, 0.021751808166503905, 0.021548032760620117, 0.021643264770507813, 0.021566463470458985, 0.022872064590454103, 0.022042623519897463, 0.021644351959228515, 0.02156332778930664, 0.021745664596557617, 0.02162483215332031, 0.021583871841430666, 0.02162380790710449, 0.021588991165161133, 0.021570592880249023, 0.021568479537963866, 0.021530624389648437, 0.021544960021972655, 0.021576704025268553, 0.02151628875732422, 0.021704704284667968, 0.021578752517700195, 0.021510143280029297, 0.021531648635864258, 0.021539840698242187, 0.02209587287902832, 0.022054912567138672, 0.021630975723266603, 0.021630975723266603, 0.021567487716674806, 0.021604352951049805, 0.021615615844726564, 0.021495872497558594, 0.021494720458984377, 0.021606399536132814, 0.021622783660888673, 0.04694732666015625, 0.02149990463256836, 0.021506080627441405, 0.02151420783996582, 0.021542911529541017, 0.021522432327270507, 0.02147020721435547, 0.021599231719970705, 0.02152448081970215, 0.02144972801208496, 0.021313535690307618, 0.021501951217651367, 0.02156550407409668, 0.021486528396606447, 0.021595136642456055, 0.021566463470458985, 0.021576704025268553, 0.021612543106079102, 0.021534719467163087, 0.021538816452026367, 0.021534719467163087, 0.021542911529541017, 0.02156342315673828, 0.021603296279907227, 0.021554176330566405, 0.021531648635864258, 0.021561344146728514, 0.021571584701538086, 0.021606399536132814, 0.021547008514404296, 0.021507072448730468, 0.02161465644836426, 0.021548992156982423, 0.021581823348999024, 0.021605375289916993, 0.021571584701538086, 0.02164121627807617, 0.021559295654296876, 0.021578752517700195, 0.02168320083618164, 0.02164531135559082, 0.02163711929321289, 0.021549055099487305, 0.021629951477050782, 0.021558271408081055, 0.021584896087646483, 0.021611520767211914, 0.02163302421569824, 0.021555200576782226, 0.021561344146728514, 0.02150092887878418, 0.021549055099487305, 0.021548032760620117, 0.021817407608032226, 0.023067583084106447, 0.02248806381225586, 0.022123519897460937, 0.02169241523742676, 0.021929983139038087, 0.021817344665527344, 0.02186444854736328, 0.021950464248657226, 0.021619712829589844, 0.04780441665649414, 0.022130687713623046, 0.02162073516845703, 0.02165452766418457, 0.021505023956298826, 0.021571584701538086, 0.021646335601806642, 0.02231500816345215, 0.021991424560546875, 0.021728256225585937, 0.021742591857910155, 0.021569536209106444, 0.021737472534179687, 0.02166169548034668, 0.021529600143432616, 0.021643264770507813, 0.021597183227539063, 0.02187980842590332, 0.022001663208007814, 0.022384639739990234, 0.02250649642944336, 0.021992448806762696, 0.022120447158813478, 0.02169241523742676, 0.021916671752929686, 0.022411264419555665, 0.021980159759521483, 0.02147327995300293, 0.021531648635864258, 0.021511167526245118, 0.021463071823120117, 0.021717983245849608, 0.021606399536132814, 0.021571584701538086, 0.02150092887878418, 0.021519359588623048, 0.021781503677368166, 0.02147532844543457, 0.02169036865234375, 0.02165555191040039, 0.0214968318939209, 0.02148249626159668, 0.02149068832397461, 0.021522432327270507, 0.021711872100830077, 0.021832704544067383, 0.021578752517700195, 0.02165452766418457, 0.02149171257019043, 0.02152448081970215, 0.021578752517700195, 0.021590015411376954, 0.021534719467163087, 0.021545984268188476, 0.021597183227539063, 0.02167807960510254, 0.021617664337158202, 0.021510143280029297, 0.02161664009094238, 0.021584896087646483, 0.02164838409423828, 0.021595136642456055, 0.02168012809753418, 0.046889984130859375, 0.021547008514404296, 0.021567487716674806, 0.021604352951049805, 0.021572608947753907, 0.021583871841430666, 0.021702655792236326, 0.02168524742126465, 0.022210559844970702, 0.022600704193115235, 0.021905439376831055, 0.021652448654174806, 0.0217262077331543, 0.0218787841796875, 0.021635072708129883, 0.02169343948364258, 0.02164531135559082, 0.02168832015991211, 0.022055936813354493, 0.021537792205810546, 0.021582847595214845, 0.021642240524291992, 0.02153267288208008, 0.021587968826293946, 0.021598207473754884, 0.021595136642456055, 0.021614591598510743, 0.021954559326171876, 0.021597183227539063, 0.021543935775756837, 0.02152448081970215, 0.02185215950012207, 0.021642240524291992, 0.021595136642456055, 0.021554176330566405, 0.021547008514404296, 0.021773311614990236, 0.021793792724609375, 0.023221248626708983, 0.0222873592376709, 0.021884927749633788, 0.021941247940063476, 0.021567487716674806, 0.021533695220947266, 0.021602304458618164, 0.02167091178894043, 0.02150912094116211, 0.021547008514404296, 0.022218751907348632, 0.02185215950012207, 0.021635072708129883, 0.021598207473754884, 0.021592063903808592, 0.02242355155944824, 0.02231603240966797, 0.021988351821899413, 0.021639167785644533, 0.022130687713623046, 0.02162073516845703, 0.021561344146728514, 0.02144256019592285, 0.021734399795532225, 0.021617664337158202, 0.04705587387084961, 0.021587968826293946, 0.021554176330566405, 0.021530624389648437, 0.021789695739746092, 0.0215285758972168, 0.021600255966186522, 0.02147123146057129, 0.021745664596557617, 0.02168832015991211, 0.02162483215332031, 0.021579776763916016, 0.021394432067871092, 0.021385215759277345, 0.021514240264892577, 0.021553152084350585, 0.021517311096191406, 0.021525503158569336, 0.021779455184936524, 0.02148249626159668, 0.021490751266479494, 0.02153571128845215, 0.0215316162109375, 0.021610496520996093, 0.021584896087646483, 0.021557247161865235, 0.021909503936767577, 0.021564416885375977, 0.02143129539489746, 0.021749759674072267, 0.021559295654296876, 0.021574655532836915, 0.02169139289855957, 0.02170163154602051, 0.02168627166748047, 0.021600255966186522, 0.02169036865234375, 0.02149171257019043, 0.02167091178894043, 0.021646335601806642, 0.02166579246520996, 0.021570592880249023, 0.02155926322937012, 0.021577728271484374, 0.021585920333862304, 0.02165350341796875, 0.021742591857910155, 0.02158697509765625, 0.021603296279907227, 0.021541887283325196, 0.021615615844726564, 0.021603328704833984, 0.021566463470458985, 0.0216494083404541, 0.021555200576782226, 0.02189004707336426, 0.021558271408081055, 0.021566463470458985, 0.021549055099487305, 0.021561344146728514, 0.021567487716674806, 0.021535743713378908, 0.021533695220947266, 0.04710297775268555, 0.02148249626159668, 0.02148761558532715, 0.021515264511108398, 0.021536767959594725, 0.021536767959594725, 0.021521408081054686, 0.021559295654296876, 0.02148044776916504, 0.021742591857910155, 0.021617664337158202, 0.021578752517700195, 0.021394432067871092, 0.021424127578735352, 0.021557247161865235, 0.02168627166748047, 0.021574655532836915, 0.021572608947753907, 0.021819391250610352, 0.02167193603515625, 0.021550079345703126, 0.021364736557006835, 0.021547040939331054, 0.02171286392211914, 0.022006784439086914, 0.021372928619384765, 0.021578752517700195, 0.02162380790710449, 0.021766143798828123, 0.021506080627441405, 0.021520351409912108, 0.021539840698242187, 0.02165247917175293, 0.021548032760620117, 0.021588991165161133, 0.021544960021972655, 0.021587968826293946, 0.02186956787109375, 0.021591039657592775, 0.021573631286621094, 0.021603328704833984, 0.021506048202514647, 0.021579776763916016, 0.021536767959594725, 0.021521408081054686, 0.021617664337158202, 0.02187468719482422, 0.02211020851135254, 0.02210406494140625, 0.021610496520996093, 0.021733375549316408, 0.02207027244567871, 0.02162892723083496, 0.021630975723266603, 0.021613567352294923, 0.021758975982666014, 0.021992448806762696, 0.021586944580078125, 0.021575679779052736, 0.02166169548034668, 0.021985279083251954, 0.021586944580078125, 0.021592063903808592, 0.04695142364501953, 0.021402624130249022, 0.021393407821655275, 0.021329919815063478, 0.0216494083404541, 0.021761024475097656, 0.021736448287963867, 0.021651456832885742, 0.02166374397277832, 0.02167091178894043, 0.021582847595214845, 0.021535743713378908, 0.02161664009094238, 0.021603328704833984, 0.0216627197265625, 0.021794815063476563, 0.021598207473754884, 0.021742591857910155, 0.021610496520996093, 0.021943296432495117, 0.021743616104125976, 0.021569536209106444, 0.021557247161865235, 0.02165452766418457, 0.021906431198120118, 0.021873664855957032, 0.02147942352294922, 0.0214517765045166, 0.021725183486938478, 0.02150399971008301, 0.02148249626159668, 0.021580799102783203, 0.02147532844543457, 0.021807104110717773, 0.02186751937866211, 0.02149990463256836, 0.021493759155273438, 0.021574655532836915, 0.022009855270385743, 0.021703680038452147, 0.021573631286621094, 0.021517311096191406, 0.021741567611694337, 0.021621759414672852, 0.021556224822998047, 0.021541887283325196, 0.021651456832885742, 0.021849088668823242, 0.021514240264892577, 0.021526527404785157, 0.02145484733581543, 0.021402624130249022, 0.021586944580078125, 0.021543935775756837, 0.02186956787109375, 0.021578752517700195, 0.02182963180541992, 0.021800960540771484, 0.02211123275756836, 0.02230067253112793, 0.02210406494140625, 0.021929983139038087, 0.021629951477050782, 0.04704051208496094, 0.021749759674072267, 0.02152448081970215, 0.021741567611694337, 0.021511167526245118, 0.02151219177246094, 0.021717023849487305, 0.021791711807250976, 0.02191257667541504, 0.02146611213684082, 0.021710847854614256, 0.021521408081054686, 0.021536767959594725, 0.021530624389648437, 0.02163199996948242, 0.021901311874389647, 0.021593088150024413, 0.021679103851318358, 0.021976064682006836, 0.02185318374633789, 0.02164121627807617, 0.021597183227539063, 0.021615615844726564, 0.021634048461914062, 0.02171392059326172, 0.021596160888671875, 0.021541887283325196, 0.021608448028564452, 0.02173651123046875, 0.021968832015991212, 0.022009855270385743, 0.022152191162109376, 0.02209280014038086, 0.02209791946411133, 0.022026239395141603, 0.022037504196166992, 0.02208665657043457, 0.02208051109313965, 0.02211840057373047, 0.022226943969726562, 0.022162431716918944, 0.02204876708984375, 0.022108160018920898, 0.02208153533935547, 0.022076416015625, 0.022510591506958007, 0.022375423431396483, 0.02227507209777832, 0.02227097511291504, 0.022240255355834963, 0.022125568389892578, 0.022125600814819336, 0.022187999725341797, 0.02207539176940918, 0.022187007904052734, 0.022016000747680665, 0.021898239135742188, 0.02171801567077637, 0.021828607559204103, 0.022218751907348632, 0.022039583206176758, 0.022118368148803712, 0.022023231506347656, 0.047406017303466795, 0.022054912567138672, 0.021940223693847655, 0.021799936294555664, 0.02190336036682129, 0.02205286407470703, 0.021950464248657226, 0.022010879516601564, 0.02186956787109375, 0.022033407211303712, 0.02226380729675293, 0.021755903244018555, 0.022030336380004883, 0.022012992858886717, 0.022056896209716795, 0.02184806442260742, 0.021938175201416017, 0.02186342430114746, 0.02147532844543457, 0.02188287925720215, 0.021884927749633788, 0.02186854362487793, 0.02208051109313965, 0.02209382438659668, 0.022071296691894532, 0.022041599273681642, 0.022137855529785155, 0.022200319290161134, 0.022192127227783204, 0.02202726364135742, 0.02212761688232422, 0.02222489547729492, 0.02206515121459961, 0.02211737632751465, 0.022133760452270508, 0.021832704544067383, 0.021775360107421874, 0.021948415756225585, 0.02209689521789551, 0.021999616622924805, 0.022034431457519533, 0.022161407470703123, 0.022040576934814454, 0.022032384872436524, 0.022171648025512695, 0.022165504455566407, 0.022204416275024414, 0.02225971221923828, 0.022090751647949217, 0.021938175201416017, 0.022208511352539064, 0.022153280258178712, 0.02202412796020508, 0.021946367263793946, 0.02211123275756836, 0.02222390365600586, 0.021888032913208007, 0.022071231842041017, 0.022125568389892578, 0.022154239654541014, 0.022191104888916017, 0.02211942481994629, 0.022167552947998048, 0.04878131103515625, 0.022565887451171874, 0.022185983657836913, 0.02208358383178711, 0.022173696517944336, 0.022063104629516602, 0.022143999099731446, 0.02189004707336426, 0.022107135772705077, 0.02220953559875488, 0.02229350471496582, 0.022143999099731446, 0.022038528442382813, 0.022071296691894532, 0.02209382438659668, 0.02218297576904297, 0.022261695861816408, 0.02203545570373535, 0.022129663467407225, 0.02188287925720215, 0.021986303329467775, 0.02162892723083496, 0.021578752517700195, 0.021572608947753907, 0.02169753646850586, 0.02171801567077637, 0.021803007125854493, 0.02169343948364258, 0.021685312271118164, 0.023793600082397462, 0.022803455352783202, 0.02231500816345215, 0.021643264770507813, 0.02168320083618164, 0.021646335601806642, 0.02204569625854492, 0.02231091117858887, 0.022163455963134765, 0.02205081558227539, 0.022231039047241212, 0.021991424560546875, 0.0222423038482666, 0.022098943710327147, 0.022153215408325197, 0.02210918426513672, 0.022151199340820313, 0.02233955192565918, 0.022163455963134765, 0.022586368560791017, 0.0220579833984375, 0.022339584350585938, 0.0220446720123291, 0.022211584091186523, 0.02209791946411133, 0.022107135772705077, 0.02234060859680176, 0.02211327934265137, 0.022185983657836913, 0.021766143798828123, 0.02215116882324219, 0.0220897274017334, 0.022131711959838866, 0.02209280014038086]",tokens/s,45.09673248488415,,,main,False,False,,,,,,,,,,,,,,,,,,,,,,, float16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-70m,,cuda,0,42,,,,,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,217063f5c507ed7cc255df7e1f64c4333a0b4dfe,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,,MB,1315.168256,883.42528,0.0,236.978176,215.486464,s,28,0.182021791934967,0.006500778283391679,0.0002263058519311169,0.006460000038146972,0.006561971139907837,0.006944868779182433,0.007283769564628601,"[0.007343039989471435, 0.0061948800086975096, 0.006518208026885987, 0.006405087947845459, 0.007123519897460937, 0.006466559886932373, 0.006432960033416748, 0.0064646081924438475, 0.006444159984588623, 0.006486303806304931, 0.006613088130950928, 0.006523007869720459, 0.006525951862335205, 0.00644163179397583, 0.0064856958389282226, 0.006450975894927978, 0.006482336044311523, 0.0064481601715087895, 0.006441343784332276, 0.00642790412902832, 0.006089024066925049, 0.006455904006958008, 0.006467936038970948, 0.006464096069335937, 0.00642137622833252, 0.006437888145446777, 0.006426080226898194, 0.0065400638580322265]",tokens/s,39379.90019657093,kWh,7.492711197067531e-08,4.105308570907706e-08,1.423468693202274e-07,2.583270669999798e-07,tokens/kWh,990991780.1994014,MB,1315.168256,883.42528,0.0,236.978176,220.639744,s,28,10.072622283935546,0.35973651014055524,0.0032474302938083083,0.3602442169189453,0.36207228088378907,0.36254206085205076,0.3633073748779297,"[0.35180557250976563, 0.35722259521484373, 0.3613737487792969, 0.3610284423828125, 0.36098431396484376, 0.361478515625, 0.3602765197753906, 0.3598533325195313, 0.363503662109375, 0.3603602600097656, 0.36210635375976563, 0.36130404663085935, 0.35905242919921876, 0.36277667236328126, 0.3601375122070313, 0.3604914245605469, 0.3609216003417969, 0.3592059326171875, 0.3594628601074219, 0.36017001342773436, 0.34657196044921873, 0.36127410888671874, 0.3620576782226563, 0.3598563537597656, 0.35982952880859376, 0.3601278991699219, 0.3602119140625, 0.35917703247070315]",tokens/s,175.12817916475817,kWh,4.016594237704222e-06,2.200914088621858e-06,6.214890017231273e-06,1.2432398343557352e-05,tokens/kWh,5067405.19882453,,s,1763,10.241831935882582,0.005809320440092212,0.0008596758214108518,0.005706751823425293,0.005765939044952393,0.005850594997406006,0.012617625446319579,"[0.006372352123260498, 0.0061521921157836916, 0.006253568172454834, 0.0061931519508361815, 0.006154304027557373, 0.006063039779663086, 0.006088704109191895, 0.006377471923828125, 0.005947391986846923, 0.005459968090057373, 0.005411839962005615, 0.005409791946411133, 0.005471231937408447, 0.0053944320678710935, 0.005417984008789062, 0.005388288021087646, 0.005396512031555176, 0.005360608100891113, 0.005402624130249023, 0.005549056053161621, 0.005410816192626953, 0.005486591815948487, 0.005395455837249756, 0.0054282240867614745, 0.005409791946411133, 0.005449728012084961, 0.005434368133544922, 0.005405695915222168, 0.005393407821655274, 0.005425151824951172, 0.005386240005493164, 0.005386240005493164, 0.0054579200744628905, 0.0054271998405456545, 0.005385216236114502, 0.005473279953002929, 0.005421055793762207, 0.0054271998405456545, 0.005405695915222168, 0.0053944320678710935, 0.005408768177032471, 0.005406720161437988, 0.005391359806060791, 0.005441535949707031, 0.005416959762573242, 0.00541593599319458, 0.005429247856140137, 0.005388288021087646, 0.005380095958709717, 0.005439487934112549, 0.005395455837249756, 0.005385216236114502, 0.005440512180328369, 0.005370944023132324, 0.005768127918243408, 0.005731328010559082, 0.005734399795532226, 0.005772287845611572, 0.005727231979370117, 0.005789696216583252, 0.005772287845611572, 0.005768191814422607, 0.012103679656982422, 0.005413887977600097, 0.005389311790466309, 0.005434368133544922, 0.005403647899627686, 0.005401599884033203, 0.005440512180328369, 0.005386240005493164, 0.005412864208221436, 0.00538316822052002, 0.00538316822052002, 0.005429247856140137, 0.005403647899627686, 0.005393407821655274, 0.005437439918518067, 0.0057118721008300784, 0.005705728054046631, 0.0057487359046936035, 0.0056780800819396975, 0.005673984050750733, 0.005631999969482422, 0.005668863773345947, 0.005658624172210694, 0.00565561580657959, 0.005704639911651611, 0.005701632022857666, 0.005755904197692871, 0.0057497601509094234, 0.005739520072937012, 0.005742591857910156, 0.005708799839019775, 0.005765120029449463, 0.0057190399169921875, 0.005739520072937012, 0.005725183963775635, 0.0057415680885314945, 0.005715968132019043, 0.0057077760696411135, 0.005754879951477051, 0.005723135948181152, 0.005948416233062744, 0.005747712135314941, 0.005775360107421875, 0.00577023983001709, 0.005799935817718506, 0.005787648200988769, 0.005753856182098388, 0.005846015930175781, 0.005713920116424561, 0.005777408123016357, 0.005987328052520752, 0.005815296173095703, 0.005740543842315674, 0.0057825279235839844, 0.005747712135314941, 0.005776447772979736, 0.005711840152740478, 0.005697504043579101, 0.0057487359046936035, 0.00570470380783081, 0.005728256225585937, 0.0057118721008300784, 0.005742591857910156, 0.012722175598144531, 0.005698560237884521, 0.00572211217880249, 0.005716991901397705, 0.005723135948181152, 0.005755904197692871, 0.005698560237884521, 0.005752831935882568, 0.005699584007263184, 0.005716991901397705, 0.0057825279235839844, 0.005701632022857666, 0.005745664119720459, 0.005691391944885254, 0.005728256225585937, 0.005701632022857666, 0.005731328010559082, 0.005730368137359619, 0.005705664157867432, 0.005756927967071533, 0.005691391944885254, 0.005718016147613526, 0.005725183963775635, 0.005747712135314941, 0.005716991901397705, 0.00568832015991211, 0.005753856182098388, 0.005700607776641845, 0.00572108793258667, 0.005701632022857666, 0.005706751823425293, 0.005718016147613526, 0.00566476821899414, 0.00572211217880249, 0.005713920116424561, 0.005699584007263184, 0.005687295913696289, 0.005702655792236328, 0.0057118721008300784, 0.005764095783233642, 0.005780479907989502, 0.005753856182098388, 0.005745664119720459, 0.0057292799949646, 0.005747712135314941, 0.005731328010559082, 0.005769216060638428, 0.005743680000305176, 0.005701568126678467, 0.005751808166503906, 0.0060764479637146, 0.005811168193817139, 0.005715968132019043, 0.0057118721008300784, 0.005706751823425293, 0.005708799839019775, 0.005706783771514893, 0.005695456027984619, 0.005783552169799804, 0.0057118721008300784, 0.0057190399169921875, 0.00568012809753418, 0.005710847854614258, 0.012568575859069824, 0.005693439960479736, 0.005744639873504639, 0.005720128059387207, 0.005714879989624023, 0.005783552169799804, 0.005687295913696289, 0.005726208209991455, 0.005681215763092041, 0.005734335899353027, 0.005693439960479736, 0.005694464206695556, 0.005739520072937012, 0.005687295913696289, 0.005698560237884521, 0.005718016147613526, 0.005713920116424561, 0.005726240158081055, 0.005709792137145996, 0.005733376026153565, 0.005720064163208007, 0.0057292799949646, 0.005774335861206055, 0.005739520072937012, 0.005770304203033447, 0.005724095821380616, 0.005740543842315674, 0.005691391944885254, 0.005731328010559082, 0.005710847854614258, 0.005798912048339844, 0.005723135948181152, 0.005681151866912842, 0.00571289587020874, 0.005718016147613526, 0.00572108793258667, 0.005701632022857666, 0.00573747205734253, 0.005724160194396972, 0.005689343929290771, 0.005740543842315674, 0.005699584007263184, 0.005730368137359619, 0.005703616142272949, 0.005725183963775635, 0.00568012809753418, 0.00570470380783081, 0.005780479907989502, 0.005709824085235596, 0.005718016147613526, 0.005694464206695556, 0.005718016147613526, 0.005705728054046631, 0.005690368175506591, 0.005724160194396972, 0.005684224128723145, 0.005696512222290039, 0.005706751823425293, 0.005725183963775635, 0.005703680038452149, 0.0056935038566589356, 0.005767104148864746, 0.005894144058227539, 0.013527039527893067, 0.005755904197692871, 0.005740543842315674, 0.005651519775390625, 0.005641151905059814, 0.005647359848022461, 0.0056258559226989744, 0.005629951953887939, 0.005715968132019043, 0.005724160194396972, 0.005702655792236328, 0.005750783920288086, 0.005746687889099121, 0.005695487976074219, 0.005730303764343261, 0.005699584007263184, 0.005740543842315674, 0.005734399795532226, 0.0057712640762329105, 0.005734399795532226, 0.005668863773345947, 0.005715968132019043, 0.005733376026153565, 0.00571289587020874, 0.005727231979370117, 0.00571289587020874, 0.005765120029449463, 0.005703680038452149, 0.005780479907989502, 0.00568832015991211, 0.0057190399169921875, 0.005716991901397705, 0.006060031890869141, 0.005944320201873779, 0.005662720203399658, 0.005744639873504639, 0.0056638078689575195, 0.005727168083190918, 0.005679103851318359, 0.0057784318923950195, 0.005679103851318359, 0.00568012809753418, 0.0057292799949646, 0.005661695957183838, 0.005698560237884521, 0.005687295913696289, 0.005706751823425293, 0.00570470380783081, 0.00566476821899414, 0.005599232196807862, 0.00562278413772583, 0.005709824085235596, 0.0057077760696411135, 0.00568832015991211, 0.005751808166503906, 0.005693439960479736, 0.005752831935882568, 0.005952511787414551, 0.005780479907989502, 0.0057415680885314945, 0.005787648200988769, 0.005708799839019775, 0.005696512222290039, 0.012692480087280274, 0.005734399795532226, 0.005725183963775635, 0.005746687889099121, 0.005699584007263184, 0.005720064163208007, 0.005718016147613526, 0.005683199882507324, 0.005720064163208007, 0.005606400012969971, 0.005699584007263184, 0.005858304023742676, 0.006145023822784424, 0.005715968132019043, 0.0057149438858032225, 0.005731328010559082, 0.005715968132019043, 0.005720064163208007, 0.005713920116424561, 0.005742591857910156, 0.005700607776641845, 0.005724160194396972, 0.005708799839019775, 0.005738495826721191, 0.005725183963775635, 0.005706751823425293, 0.005757952213287353, 0.005707839965820313, 0.005749695777893067, 0.005740543842315674, 0.005720064163208007, 0.005757952213287353, 0.005689439773559571, 0.005719967842102051, 0.005691391944885254, 0.00571289587020874, 0.006141952037811279, 0.005720064163208007, 0.005695487976074219, 0.0057487359046936035, 0.005700607776641845, 0.005699584007263184, 0.005710847854614258, 0.005693439960479736, 0.00571289587020874, 0.005706751823425293, 0.005716991901397705, 0.00571289587020874, 0.005704768180847168, 0.005711808204650879, 0.005708799839019775, 0.005744639873504639, 0.00570470380783081, 0.005739520072937012, 0.005718016147613526, 0.00570470380783081, 0.00572211217880249, 0.005694464206695556, 0.005710847854614258, 0.0057077760696411135, 0.005702655792236328, 0.005698560237884521, 0.005694464206695556, 0.012620800018310547, 0.005733376026153565, 0.005692416191101074, 0.0057292799949646, 0.0056852478981018065, 0.0057118721008300784, 0.0057149438858032225, 0.005702655792236328, 0.005705728054046631, 0.005692416191101074, 0.005700607776641845, 0.005687295913696289, 0.00573747205734253, 0.005716063976287842, 0.005716896057128907, 0.0057077760696411135, 0.0057077760696411135, 0.005750783920288086, 0.005695487976074219, 0.005727231979370117, 0.005717023849487305, 0.005713888168334961, 0.00572108793258667, 0.005724160194396972, 0.005746687889099121, 0.005673984050750733, 0.005728256225585937, 0.00572211217880249, 0.005723135948181152, 0.005695487976074219, 0.005693439960479736, 0.005750783920288086, 0.005699584007263184, 0.005738495826721191, 0.005696512222290039, 0.005755904197692871, 0.0057190399169921875, 0.005698560237884521, 0.005724160194396972, 0.005687295913696289, 0.005734399795532226, 0.00570470380783081, 0.005731391906738281, 0.005693376064300537, 0.0056852478981018065, 0.005775360107421875, 0.00572108793258667, 0.005692416191101074, 0.005705728054046631, 0.005716991901397705, 0.00568832015991211, 0.005685311794281006, 0.005719999790191651, 0.005694464206695556, 0.005730303764343261, 0.005700640201568604, 0.005730271816253662, 0.005702655792236328, 0.005679103851318359, 0.0057190399169921875, 0.005689343929290771, 0.005702655792236328, 0.005702655792236328, 0.012592127799987793, 0.00571289587020874, 0.005699584007263184, 0.005695487976074219, 0.005691391944885254, 0.005668863773345947, 0.005709824085235596, 0.005702655792236328, 0.005701632022857666, 0.005701632022857666, 0.005690368175506591, 0.005698560237884521, 0.005696512222290039, 0.005715968132019043, 0.005692416191101074, 0.005873663902282715, 0.005747712135314941, 0.005703680038452149, 0.005690368175506591, 0.005670944213867188, 0.005711840152740478, 0.005703680038452149, 0.005691391944885254, 0.005693439960479736, 0.00569862413406372, 0.0057066879272460935, 0.005690368175506591, 0.00572211217880249, 0.005697535991668701, 0.005706751823425293, 0.005693439960479736, 0.005698560237884521, 0.005730303764343261, 0.005700672149658203, 0.0057363839149475095, 0.00570470380783081, 0.0057118721008300784, 0.005706751823425293, 0.005676032066345215, 0.005697535991668701, 0.005684224128723145, 0.005730303764343261, 0.005686272144317627, 0.005690368175506591, 0.00568012809753418, 0.00568012809753418, 0.005728256225585937, 0.005676032066345215, 0.005715968132019043, 0.005716991901397705, 0.0056780800819396975, 0.005705728054046631, 0.005687295913696289, 0.005720064163208007, 0.0057149438858032225, 0.005733376026153565, 0.005710847854614258, 0.005709824085235596, 0.005700607776641845, 0.005684224128723145, 0.0057292799949646, 0.00572108793258667, 0.00572108793258667, 0.012615679740905762, 0.005723135948181152, 0.005742688179016114, 0.0057281599044799805, 0.0057190399169921875, 0.005716991901397705, 0.005694464206695556, 0.005735424041748047, 0.005713920116424561, 0.005718016147613526, 0.005730303764343261, 0.005910592079162598, 0.005893087863922119, 0.005851103782653809, 0.006043647766113281, 0.006368256092071533, 0.0058726401329040525, 0.006154240131378174, 0.005738495826721191, 0.005795839786529541, 0.005760000228881836, 0.005716991901397705, 0.005780479907989502, 0.005716991901397705, 0.005776383876800537, 0.005743616104125977, 0.005774335861206055, 0.0057077760696411135, 0.005689343929290771, 0.005720064163208007, 0.005700607776641845, 0.0057149438858032225, 0.005698560237884521, 0.005713920116424561, 0.005701632022857666, 0.0056852478981018065, 0.0057149438858032225, 0.005696512222290039, 0.005725183963775635, 0.005703680038452149, 0.005814271926879883, 0.006278143882751465, 0.005765120029449463, 0.005702655792236328, 0.005698560237884521, 0.005696512222290039, 0.005694464206695556, 0.00572211217880249, 0.005691391944885254, 0.005734399795532226, 0.00568832015991211, 0.005696512222290039, 0.005789696216583252, 0.00571289587020874, 0.005775392055511475, 0.005739488124847412, 0.005719103813171387, 0.005683135986328125, 0.005709824085235596, 0.005702655792236328, 0.005676032066345215, 0.005690368175506591, 0.005693439960479736, 0.012670975685119629, 0.00572211217880249, 0.005690368175506591, 0.0057077760696411135, 0.005693439960479736, 0.005696512222290039, 0.005731328010559082, 0.005683199882507324, 0.005720064163208007, 0.005711935997009278, 0.005705664157867432, 0.0057292799949646, 0.005723135948181152, 0.005734399795532226, 0.005698560237884521, 0.005708799839019775, 0.0057118721008300784, 0.005728256225585937, 0.0057149438858032225, 0.005670911788940429, 0.005744639873504639, 0.005682176113128662, 0.005732351779937744, 0.005691391944885254, 0.005705728054046631, 0.005705728054046631, 0.005703680038452149, 0.005726208209991455, 0.005700704097747803, 0.005716896057128907, 0.005700607776641845, 0.00572211217880249, 0.005692416191101074, 0.005702655792236328, 0.0057149438858032225, 0.00571289587020874, 0.005713920116424561, 0.005700607776641845, 0.00570470380783081, 0.005733439922332764, 0.0056728959083557125, 0.005714015960693359, 0.005615520000457764, 0.005659647941589355, 0.005616703987121582, 0.00562988805770874, 0.005605375766754151, 0.005619711875915527, 0.005650432109832764, 0.005706751823425293, 0.005731328010559082, 0.0056893758773803715, 0.005713888168334961, 0.005804031848907471, 0.005682176113128662, 0.005715968132019043, 0.005710847854614258, 0.005710847854614258, 0.005712959766387939, 0.006025152206420898, 0.005801983833312988, 0.005794816017150879, 0.0059064321517944334, 0.013111295700073243, 0.006303743839263916, 0.005900288105010986, 0.005789696216583252, 0.005760000228881836, 0.005795839786529541, 0.0057497601509094234, 0.005733376026153565, 0.005775360107421875, 0.005745664119720459, 0.005791744232177734, 0.005760000228881836, 0.00577023983001709, 0.005744639873504639, 0.005740543842315674, 0.005706751823425293, 0.005696512222290039, 0.005723135948181152, 0.005689343929290771, 0.005716991901397705, 0.005971968173980713, 0.00582041597366333, 0.005742591857910156, 0.00563097620010376, 0.005686272144317627, 0.005631008148193359, 0.005719007968902588, 0.005723135948181152, 0.005743648052215576, 0.0058633918762207035, 0.005684224128723145, 0.005852159976959229, 0.00568832015991211, 0.0057415680885314945, 0.005696512222290039, 0.005691391944885254, 0.005596159934997558, 0.005596159934997558, 0.005706751823425293, 0.005595136165618897, 0.005646336078643799, 0.005606400012969971, 0.005586944103240967, 0.00566374397277832, 0.0056217598915100095, 0.005718016147613526, 0.005700607776641845, 0.005735424041748047, 0.005696512222290039, 0.005681151866912842, 0.005718016147613526, 0.005679103851318359, 0.005698560237884521, 0.005693439960479736, 0.005715968132019043, 0.005701632022857666, 0.005822495937347412, 0.005765088081359863, 0.0057077760696411135, 0.005693439960479736, 0.0057292799949646, 0.006123519897460937, 0.005690368175506591, 0.012659711837768555, 0.005683199882507324, 0.005696512222290039, 0.005803008079528809, 0.005697535991668701, 0.00571289587020874, 0.005736447811126709, 0.005682176113128662, 0.005699584007263184, 0.005766143798828125, 0.005716991901397705, 0.005732351779937744, 0.005693439960479736, 0.005726208209991455, 0.005677055835723877, 0.005757952213287353, 0.005725183963775635, 0.0057497920989990235, 0.005881824016571045, 0.005698560237884521, 0.005745664119720459, 0.0057292799949646, 0.0057415680885314945, 0.00571289587020874, 0.005775360107421875, 0.005720064163208007, 0.0057077760696411135, 0.005689343929290771, 0.005702655792236328, 0.005706751823425293, 0.0056780800819396975, 0.005764095783233642, 0.005786623954772949, 0.005733439922332764, 0.005713856220245362, 0.005708799839019775, 0.00577023983001709, 0.005693439960479736, 0.005716991901397705, 0.005697535991668701, 0.005720064163208007, 0.005703680038452149, 0.005681151866912842, 0.0057149438858032225, 0.005670976161956787, 0.005872576236724853, 0.005731328010559082, 0.005756927967071533, 0.005728256225585937, 0.005696512222290039, 0.005727231979370117, 0.005672959804534912, 0.0057118721008300784, 0.005686272144317627, 0.005715968132019043, 0.005694464206695556, 0.005694464206695556, 0.0057415680885314945, 0.005690368175506591, 0.005709824085235596, 0.005730303764343261, 0.006089727878570556, 0.005726208209991455, 0.012700672149658204, 0.005724192142486572, 0.005703648090362549, 0.005756927967071533, 0.005699584007263184, 0.00568832015991211, 0.0057415680885314945, 0.00577023983001709, 0.00577023983001709, 0.005694464206695556, 0.005709824085235596, 0.005692416191101074, 0.005684224128723145, 0.005725183963775635, 0.005672959804534912, 0.005710847854614258, 0.00566476821899414, 0.005724160194396972, 0.00571289587020874, 0.005681151866912842, 0.005715968132019043, 0.005698560237884521, 0.005746687889099121, 0.005715968132019043, 0.0057118721008300784, 0.005683199882507324, 0.005677087783813477, 0.005719007968902588, 0.005690368175506591, 0.005706751823425293, 0.00570470380783081, 0.005693439960479736, 0.005703680038452149, 0.005710847854614258, 0.005839871883392334, 0.005718016147613526, 0.005743616104125977, 0.005702655792236328, 0.005695519924163818, 0.00568726396560669, 0.005677055835723877, 0.005702655792236328, 0.005692416191101074, 0.005822463989257813, 0.005743616104125977, 0.005758975982666016, 0.005703680038452149, 0.005724160194396972, 0.005769279956817627, 0.005630911827087403, 0.005608448028564453, 0.005596159934997558, 0.005576704025268555, 0.005598207950592041, 0.005590015888214111, 0.005620736122131348, 0.005597184181213379, 0.005560319900512695, 0.005623807907104492, 0.005573631763458252, 0.005610496044158936, 0.005657599925994873, 0.005668863773345947, 0.012621824264526366, 0.005708799839019775, 0.00566374397277832, 0.005703680038452149, 0.005686272144317627, 0.005701632022857666, 0.005713920116424561, 0.006104063987731933, 0.005943295955657959, 0.005904416084289551, 0.006235104084014892, 0.0064174079895019534, 0.005808127880096436, 0.005780479907989502, 0.005720064163208007, 0.005773312091827393, 0.005745664119720459, 0.00570470380783081, 0.005796864032745362, 0.005710847854614258, 0.00576204776763916, 0.005726208209991455, 0.0057784318923950195, 0.005723135948181152, 0.005745664119720459, 0.00569862413406372, 0.005684160232543946, 0.00576204776763916, 0.005691391944885254, 0.005716991901397705, 0.005708799839019775, 0.005710847854614258, 0.005703680038452149, 0.00568832015991211, 0.005705728054046631, 0.005753856182098388, 0.0057118721008300784, 0.005702655792236328, 0.0057118721008300784, 0.0057242240905761715, 0.005671872138977051, 0.0057292799949646, 0.006073344230651856, 0.005693439960479736, 0.005667840003967285, 0.005706751823425293, 0.005682176113128662, 0.0056780800819396975, 0.005742591857910156, 0.005696512222290039, 0.005735424041748047, 0.005723135948181152, 0.005691391944885254, 0.005776383876800537, 0.0056514558792114256, 0.005697535991668701, 0.005668863773345947, 0.005696512222290039, 0.005676032066345215, 0.005691391944885254, 0.005692416191101074, 0.005673984050750733, 0.00568832015991211, 0.012633088111877442, 0.005667840003967285, 0.005695487976074219, 0.005691391944885254, 0.005735424041748047, 0.005683199882507324, 0.005684224128723145, 0.005701632022857666, 0.005681151866912842, 0.00570470380783081, 0.005694464206695556, 0.005699584007263184, 0.0057118721008300784, 0.005724160194396972, 0.005724160194396972, 0.005656576156616211, 0.005706751823425293, 0.005690368175506591, 0.0057077760696411135, 0.005696512222290039, 0.005691391944885254, 0.005699584007263184, 0.005702655792236328, 0.005710847854614258, 0.005701632022857666, 0.005703680038452149, 0.005676032066345215, 0.005672959804534912, 0.0057292799949646, 0.005665791988372803, 0.005682176113128662, 0.005697535991668701, 0.005686272144317627, 0.005738495826721191, 0.005689343929290771, 0.005698560237884521, 0.005679103851318359, 0.005723135948181152, 0.005695487976074219, 0.005683199882507324, 0.005718016147613526, 0.005670911788940429, 0.005764095783233642, 0.005750783920288086, 0.00574675178527832, 0.005731264114379883, 0.005768288135528565, 0.005759903907775879, 0.005708799839019775, 0.005790719985961914, 0.005700607776641845, 0.005764095783233642, 0.005769216060638428, 0.00573747205734253, 0.005743616104125977, 0.005716991901397705, 0.005732351779937744, 0.005697535991668701, 0.0057712640762329105, 0.005703743934631348, 0.005718976020812988, 0.005687295913696289, 0.005672959804534912, 0.012626943588256835, 0.005686272144317627, 0.005701632022857666, 0.00571395206451416, 0.005684192180633545, 0.005629951953887939, 0.005623807907104492, 0.00562278413772583, 0.005602303981781006, 0.005590015888214111, 0.00562278413772583, 0.00561356782913208, 0.005631999969482422, 0.005656576156616211, 0.005698560237884521, 0.005735424041748047, 0.005687295913696289, 0.005718016147613526, 0.005703680038452149, 0.005725183963775635, 0.005690368175506591, 0.00568832015991211, 0.005709824085235596, 0.005718016147613526, 0.0057118721008300784, 0.005699584007263184, 0.00623308801651001, 0.005840896129608154, 0.005816383838653565, 0.005765056133270264, 0.005695487976074219, 0.005752831935882568, 0.005676032066345215, 0.005706751823425293, 0.005693439960479736, 0.005727231979370117, 0.005675007820129394, 0.005702655792236328, 0.005742591857910156, 0.005705728054046631, 0.005734399795532226, 0.00576204776763916, 0.00572211217880249, 0.005695487976074219, 0.005677055835723877, 0.005716991901397705, 0.00566374397277832, 0.005757984161376953, 0.005708767890930176, 0.005934080123901367, 0.00572211217880249, 0.005694464206695556, 0.005752927780151367, 0.005754784107208252, 0.005744639873504639, 0.005736447811126709, 0.00572108793258667, 0.005710847854614258, 0.005708799839019775, 0.0057292799949646, 0.005690368175506591, 0.005718016147613526, 0.0057077760696411135, 0.012640255928039551, 0.00572108793258667, 0.005691391944885254, 0.005690368175506591, 0.005681248188018799, 0.005674911975860596, 0.00570470380783081, 0.00567193603515625, 0.005705728054046631, 0.005697535991668701, 0.005691455841064453, 0.005723072052001953, 0.0057077760696411135, 0.005728256225585937, 0.005675007820129394, 0.005706751823425293, 0.00570470380783081, 0.005724160194396972, 0.005691487789154053, 0.005702559947967529, 0.005743616104125977, 0.00570470380783081, 0.005710847854614258, 0.00570470380783081, 0.005716991901397705, 0.005697535991668701, 0.005684224128723145, 0.005705728054046631, 0.005670911788940429, 0.006087679862976075, 0.00571289587020874, 0.005718016147613526, 0.005659647941589355, 0.005689343929290771, 0.005732351779937744, 0.005705728054046631, 0.005701632022857666, 0.005684224128723145, 0.00570470380783081, 0.005713920116424561, 0.005695487976074219, 0.005724160194396972, 0.005652480125427246, 0.005687295913696289, 0.0056780800819396975, 0.00571289587020874, 0.005716991901397705, 0.005691391944885254, 0.005791744232177734, 0.0057231678962707516, 0.00572822380065918, 0.005677055835723877, 0.006113279819488526, 0.006120448112487793, 0.005776383876800537, 0.00572211217880249, 0.0057077760696411135, 0.005695487976074219, 0.005690368175506591, 0.005718016147613526, 0.005670911788940429, 0.005751808166503906, 0.005619711875915527, 0.012601344108581543, 0.005726208209991455, 0.005702655792236328, 0.00570470380783081, 0.0057292799949646, 0.005692416191101074, 0.005725183963775635, 0.005703680038452149, 0.0057415680885314945, 0.005710847854614258, 0.005679103851318359, 0.005726208209991455, 0.005687295913696289, 0.005705728054046631, 0.00570681619644165, 0.005716928005218506, 0.005695487976074219, 0.005723135948181152, 0.0057487359046936035, 0.00566476821899414, 0.005733376026153565, 0.0057190399169921875, 0.005727231979370117, 0.0057077760696411135, 0.005705728054046631, 0.005755904197692871, 0.005696512222290039, 0.0057190399169921875, 0.005709824085235596, 0.005736447811126709, 0.005705728054046631, 0.005695487976074219, 0.005753856182098388, 0.005682176113128662, 0.00573747205734253, 0.005723135948181152, 0.0056852478981018065, 0.005608448028564453, 0.005628928184509278, 0.005646336078643799, 0.005585919857025146, 0.005612544059753418, 0.005616640090942383, 0.005609471797943116, 0.005705728054046631, 0.005694464206695556, 0.005732351779937744, 0.005710847854614258, 0.005710847854614258, 0.005695487976074219, 0.005697535991668701, 0.005728256225585937, 0.005691391944885254, 0.005726208209991455, 0.00570470380783081, 0.005731328010559082, 0.005713920116424561, 0.005670911788940429, 0.0056555519104003905, 0.005600255966186523, 0.005650432109832764, 0.005631999969482422, 0.005615615844726562, 0.01267199993133545, 0.005700607776641845, 0.005696512222290039, 0.0057415680885314945, 0.005698560237884521, 0.005729343891143799, 0.005696447849273681, 0.005736447811126709, 0.0056852478981018065, 0.005853184223175049, 0.005776383876800537, 0.005726208209991455, 0.005735424041748047, 0.005681151866912842, 0.005706751823425293, 0.0057149438858032225, 0.0057621440887451175, 0.005716896057128907, 0.005690368175506591, 0.005751808166503906, 0.005738495826721191, 0.005709824085235596, 0.005705728054046631, 0.005718016147613526, 0.005702655792236328, 0.005662720203399658, 0.005757952213287353, 0.00571289587020874, 0.0057487359046936035, 0.00572108793258667, 0.005728256225585937, 0.005705728054046631, 0.005687295913696289, 0.005723135948181152, 0.0057077760696411135, 0.005689343929290771, 0.00561356782913208, 0.005627903938293457, 0.00563097620010376, 0.0055920639038085935, 0.005638144016265869, 0.005665791988372803, 0.0057497601509094234, 0.005708799839019775, 0.005699584007263184, 0.005753856182098388, 0.005672959804534912, 0.005706751823425293, 0.0057149438858032225, 0.005746687889099121, 0.005692416191101074, 0.005709824085235596, 0.005640192031860352, 0.005598207950592041, 0.005673984050750733, 0.005617663860321045, 0.005590015888214111, 0.005653503894805908, 0.005636096000671387, 0.005646336078643799, 0.005669888019561768, 0.005731328010559082, 0.005683199882507324, 0.012584959983825684, 0.005726208209991455, 0.005718016147613526, 0.0057118721008300784, 0.005769216060638428, 0.005681151866912842, 0.005709824085235596, 0.005701632022857666, 0.005710847854614258, 0.0057077760696411135, 0.006062143802642822, 0.00573740816116333, 0.005682176113128662, 0.0057077760696411135, 0.0056780800819396975, 0.005697535991668701, 0.005743616104125977, 0.005684224128723145, 0.005701632022857666, 0.005697535991668701, 0.005709824085235596, 0.005705728054046631, 0.0057149438858032225, 0.005697535991668701, 0.005694464206695556, 0.005718016147613526, 0.005668863773345947, 0.005694464206695556, 0.005696576118469238, 0.005715904235839844, 0.005689343929290771, 0.005682176113128662, 0.0056852478981018065, 0.0056780800819396975, 0.005706751823425293, 0.005691391944885254, 0.005720064163208007, 0.005701632022857666, 0.0057077760696411135, 0.005725183963775635, 0.005677055835723877, 0.005709824085235596, 0.005696512222290039, 0.005702655792236328, 0.005691391944885254, 0.005695487976074219, 0.005857279777526855, 0.005661695957183838, 0.00572108793258667, 0.005701632022857666, 0.005706751823425293, 0.005696512222290039, 0.0056780800819396975, 0.005713920116424561, 0.005709824085235596, 0.00568832015991211, 0.005690368175506591, 0.0056934719085693355, 0.005711840152740478, 0.00568832015991211, 0.005723135948181152, 0.005691391944885254, 0.005702655792236328, 0.011913215637207031, 0.005354496002197265, 0.005343232154846191, 0.005369855880737305, 0.005352447986602784, 0.005386240005493164, 0.005378047943115235, 0.00537497615814209, 0.00538316822052002, 0.005453824043273926, 0.005386240005493164, 0.005399551868438721, 0.0054282240867614745, 0.005423168182373047, 0.005303232192993164, 0.0053340158462524415, 0.005303296089172363, 0.005345280170440674, 0.0053340158462524415, 0.00531660795211792, 0.0053350400924682614, 0.005359615802764893, 0.005369855880737305, 0.005345280170440674, 0.005403647899627686, 0.005346303939819336, 0.005349376201629639, 0.005426176071166992, 0.00537395191192627, 0.005387263774871826, 0.0053606400489807126, 0.005356544017791748, 0.0053821439743042, 0.00536575984954834, 0.005325823783874512, 0.005405695915222168, 0.005353536128997803, 0.005369791984558106, 0.005694464206695556, 0.005699584007263184, 0.005697535991668701, 0.005708799839019775, 0.005677055835723877, 0.005715968132019043, 0.005692416191101074, 0.005690368175506591, 0.0056780800819396975, 0.005652480125427246, 0.005617663860321045, 0.0056145920753479005, 0.005624832153320313, 0.005585951805114746, 0.005626848220825195, 0.005710847854614258, 0.005694464206695556, 0.005693439960479736, 0.00571289587020874, 0.005693439960479736, 0.005839871883392334, 0.005732351779937744, 0.005733376026153565, 0.005681151866912842, 0.005705728054046631, 0.012626943588256835, 0.005684224128723145, 0.005738495826721191, 0.00568832015991211, 0.005703680038452149, 0.00570470380783081, 0.005713920116424561, 0.005710847854614258, 0.005690368175506591, 0.005708799839019775, 0.005703680038452149, 0.005725183963775635, 0.005694464206695556, 0.0056852478981018065, 0.00572211217880249, 0.005725183963775635, 0.005833727836608887, 0.00582144021987915, 0.005735424041748047, 0.005730303764343261, 0.005760000228881836, 0.005758975982666016, 0.00572211217880249, 0.005752831935882568, 0.005701632022857666, 0.005774335861206055, 0.005726208209991455, 0.005765120029449463, 0.005746687889099121, 0.005696512222290039, 0.005713920116424561, 0.005687295913696289, 0.005746687889099121, 0.005689343929290771, 0.005816319942474365, 0.005713920116424561, 0.005896192073822021, 0.005706751823425293, 0.005687295913696289, 0.005724160194396972, 0.005695487976074219, 0.005744639873504639, 0.005682176113128662, 0.005726208209991455, 0.005706751823425293, 0.005683199882507324, 0.005720064163208007, 0.005715968132019043, 0.005731328010559082, 0.00571289587020874, 0.005713920116424561, 0.005716991901397705, 0.005690368175506591, 0.005716991901397705, 0.005681151866912842, 0.00571289587020874, 0.005667840003967285, 0.005661695957183838, 0.006114304065704346, 0.00572108793258667, 0.0057190399169921875, 0.005676032066345215, 0.005713920116424561, 0.012643327713012695, 0.005683199882507324, 0.005696512222290039, 0.005700607776641845, 0.005723135948181152, 0.005708799839019775, 0.005692416191101074, 0.005756927967071533, 0.005666816234588623, 0.005701632022857666, 0.005699584007263184, 0.005698560237884521, 0.005724160194396972, 0.005706751823425293, 0.00576204776763916, 0.005665791988372803, 0.005708831787109375, 0.005668831825256348, 0.005690368175506591, 0.005726208209991455, 0.005691391944885254, 0.005720064163208007, 0.005698560237884521, 0.005725183963775635, 0.005699584007263184, 0.005686272144317627, 0.005720064163208007, 0.005668863773345947, 0.0057118721008300784, 0.005706751823425293, 0.0057149438858032225, 0.005705728054046631, 0.005720064163208007, 0.00572211217880249, 0.005698560237884521, 0.005770304203033447, 0.005747647762298584, 0.005698560237884521, 0.005691391944885254, 0.005727231979370117, 0.005687295913696289, 0.005714975833892823, 0.006147136211395263, 0.006640543937683105, 0.006212607860565185, 0.005801983833312988, 0.005744639873504639, 0.005700607776641845, 0.005785600185394287, 0.005687295913696289, 0.0057292799949646, 0.005675007820129394, 0.005699584007263184, 0.005696512222290039, 0.005689343929290771, 0.0057487359046936035, 0.005682176113128662, 0.005733376026153565, 0.005723135948181152, 0.005695487976074219, 0.005690368175506591, 0.005716991901397705, 0.00571289587020874, 0.012638208389282226, 0.005659647941589355, 0.005746687889099121, 0.005679103851318359, 0.005713920116424561, 0.005698560237884521, 0.005743616104125977, 0.005709824085235596, 0.005692416191101074, 0.005709824085235596, 0.005677055835723877, 0.00572211217880249, 0.005691391944885254, 0.005702655792236328, 0.005689343929290771, 0.005695487976074219, 0.005735424041748047, 0.005687295913696289, 0.005703680038452149, 0.0056852478981018065, 0.0057190399169921875, 0.00572108793258667, 0.005692416191101074, 0.005727231979370117, 0.005700607776641845, 0.005732351779937744, 0.005732351779937744, 0.005718016147613526, 0.0057118721008300784, 0.005683199882507324, 0.005734399795532226, 0.005687295913696289, 0.0057077760696411135, 0.005699584007263184, 0.005724160194396972, 0.0057149438858032225, 0.0057190399169921875, 0.0057415680885314945, 0.005702655792236328, 0.005706751823425293, 0.00571289587020874, 0.005772287845611572, 0.005746687889099121, 0.005697535991668701, 0.0057292799949646, 0.005681151866912842, 0.00572108793258667, 0.00570470380783081, 0.005734399795532226, 0.005730303764343261, 0.005679103851318359, 0.005718016147613526, 0.005679103851318359, 0.00570470380783081, 0.00570470380783081, 0.0057149438858032225, 0.00570470380783081, 0.005694464206695556, 0.005730303764343261, 0.005676032066345215, 0.005647359848022461, 0.00561356782913208, 0.005586944103240967, 0.012614656448364257, 0.005718016147613526, 0.005689343929290771, 0.005713920116424561, 0.005687295913696289, 0.005702655792236328, 0.005681151866912842, 0.005702655792236328, 0.0057077760696411135, 0.00566476821899414, 0.00572211217880249, 0.005687295913696289, 0.005734399795532226, 0.005692416191101074, 0.0057190399169921875, 0.005709824085235596, 0.005679103851318359, 0.005720064163208007, 0.005709824085235596, 0.005723135948181152, 0.005696512222290039, 0.005747712135314941, 0.005699584007263184, 0.005683199882507324, 0.005709856033325195, 0.005699552059173584, 0.005691391944885254, 0.005706751823425293, 0.005734464168548584, 0.005689280033111572, 0.005672959804534912, 0.005763072013854981, 0.005670911788940429, 0.005705728054046631, 0.005676032066345215, 0.005725183963775635, 0.005682176113128662, 0.005702655792236328, 0.006097919940948486, 0.005662720203399658, 0.005710847854614258, 0.005682240009307862, 0.005692351818084717, 0.005702655792236328, 0.005682176113128662, 0.005716991901397705, 0.005669919967651367, 0.005711840152740478, 0.005694464206695556, 0.00571292781829834, 0.005713888168334961, 0.005676032066345215, 0.005691391944885254, 0.005677055835723877, 0.005692480087280273, 0.005659584045410156, 0.005670911788940429, 0.005687295913696289, 0.00566476821899414, 0.005695551872253418, 0.005700543880462646, 0.005705728054046631, 0.005693439960479736, 0.012593152046203614, 0.005732351779937744, 0.0057149438858032225, 0.005692416191101074, 0.005745664119720459, 0.005853184223175049, 0.00573747205734253, 0.005679103851318359, 0.0057149438858032225, 0.005703680038452149, 0.00572211217880249, 0.005696512222290039, 0.005686272144317627, 0.005699584007263184, 0.005701632022857666, 0.00572108793258667, 0.005713920116424561, 0.005682176113128662, 0.005713920116424561, 0.005700607776641845, 0.005750783920288086, 0.005701632022857666, 0.005709824085235596, 0.005683199882507324, 0.005691391944885254, 0.005727231979370117, 0.005676032066345215, 0.005725183963775635, 0.005699584007263184, 0.0057077760696411135, 0.005689343929290771, 0.005695487976074219, 0.005794816017150879, 0.005668863773345947, 0.005694464206695556, 0.005684224128723145, 0.005708799839019775, 0.0057118721008300784, 0.005698560237884521, 0.005730303764343261, 0.005687295913696289, 0.005713920116424561, 0.005681151866912842, 0.0057292799949646, 0.005701632022857666, 0.005699584007263184, 0.005716991901397705, 0.0057118721008300784, 0.005709824085235596, 0.0057292799949646, 0.005726208209991455, 0.005704736232757569, 0.005695456027984619, 0.005733376026153565, 0.005666816234588623, 0.005728256225585937, 0.00570470380783081, 0.00570470380783081, 0.005683199882507324, 0.005677055835723877, 0.00570470380783081, 0.00567193603515625, 0.005746687889099121, 0.012634112358093261, 0.005675007820129394, 0.00572211217880249, 0.0056852478981018065, 0.00572108793258667, 0.005726208209991455, 0.005691391944885254, 0.005893119812011719, 0.005691391944885254, 0.0057292799949646, 0.005796864032745362, 0.005693439960479736, 0.005672959804534912, 0.005675007820129394, 0.005733376026153565, 0.005692416191101074, 0.005708831787109375, 0.005715936183929443, 0.005734399795532226, 0.005701632022857666, 0.005705728054046631, 0.005752831935882568, 0.0056852478981018065, 0.005740543842315674, 0.005734399795532226, 0.005701632022857666, 0.005684224128723145, 0.0056780800819396975, 0.005726208209991455, 0.0056780800819396975, 0.005724160194396972, 0.005694464206695556, 0.005696512222290039, 0.005701632022857666, 0.005702655792236328, 0.005742591857910156, 0.005718016147613526, 0.005720064163208007, 0.005705728054046631, 0.005730303764343261, 0.005624832153320313, 0.005617663860321045, 0.005653503894805908, 0.005637119770050049, 0.005646336078643799, 0.005718016147613526, 0.005695487976074219, 0.005739520072937012, 0.005697535991668701, 0.005745664119720459, 0.0057077760696411135, 0.005736447811126709, 0.005709824085235596, 0.005699584007263184, 0.005736512184143066, 0.005831615924835205, 0.005772287845611572, 0.005693439960479736, 0.00572108793258667, 0.005698560237884521, 0.005699584007263184, 0.005703680038452149, 0.00568012809753418, 0.012715007781982422, 0.005692416191101074, 0.005691391944885254, 0.005709824085235596, 0.0057077760696411135, 0.005715968132019043, 0.005701632022857666, 0.0057118721008300784, 0.005716991901397705, 0.00568012809753418, 0.005698560237884521, 0.005706751823425293, 0.0056780800819396975, 0.005690368175506591, 0.005667840003967285, 0.005727231979370117, 0.005691391944885254, 0.005694464206695556, 0.00568832015991211, 0.005690368175506591, 0.005700607776641845, 0.0056852478981018065, 0.005772287845611572, 0.005683199882507324, 0.0057077760696411135, 0.006051839828491211, 0.005708799839019775, 0.005679135799407959, 0.005703648090362549, 0.005699584007263184, 0.005649407863616943, 0.00570470380783081, 0.00566374397277832, 0.00570579195022583, 0.005678016185760498, 0.005668863773345947, 0.005708799839019775, 0.005682176113128662, 0.00572108793258667, 0.005680160045623779, 0.005696479797363281, 0.005698560237884521, 0.005693439960479736, 0.005743680000305176, 0.0056698241233825684, 0.005695487976074219, 0.005684224128723145, 0.005711904048919678, 0.005674975872039795, 0.005683199882507324, 0.00570470380783081, 0.005609471797943116, 0.005605440139770507, 0.005574592113494873, 0.005578752040863037, 0.0056145920753479005, 0.005596159934997558, 0.0057487359046936035, 0.005701632022857666, 0.00570470380783081, 0.005703680038452149, 0.0057118721008300784, 0.0057149438858032225]",tokens/s,172.1371734116507,,,main,False,False,,,,,,,,,,,,,,,,,,,,,,, float16-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,google/gemma-2b,google/gemma-2b,cuda,0,42,,,True,,,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.223-212.873.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.0,,,,1.21.4,,,,0.12.0,,,gemma,MB,1548.746752,6797.39392,0.0,6150.946816,5419.87328,s,10,0.3111448307037354,0.03111448307037354,5.5504299295999784e-05,0.031113007545471194,0.03116964817047119,0.031190311908721926,0.03120684289932251,"[0.030986431121826172, 0.031112319946289064, 0.031210975646972658, 0.03108620834350586, 0.031117055892944338, 0.03109724807739258, 0.031165056228637696, 0.031104768753051758, 0.03111369514465332, 0.031151071548461914]",tokens/s,8227.679676406291,kWh,3.6658460442250196e-07,2.008701039057232e-07,1.9744124842414904e-06,2.5418671925697154e-06,tokens/kWh,100713365.6503884,MB,1610.981376,6797.39392,0.0,6150.946816,5419.87584,s,10,14.270213378906249,1.427021337890625,0.009567950981069422,1.4271953124999999,1.4386142944335938,1.4386188537597657,1.438622501220703,"[1.4255933837890624, 1.4287972412109375, 1.4152265625, 1.4386234130859374, 1.43861328125, 1.43568603515625, 1.4083038330078126, 1.4229683837890625, 1.433564208984375, 1.4228370361328124]",tokens/s,44.147903277413135,kWh,1.6662386055368294e-05,9.130898602801004e-06,5.4451132656158225e-05,8.02444173143275e-05,tokens/kWh,785101.3454708139,,s,630,14.2685245513916,0.022648451668875556,0.0004943414891973842,0.02290380859375,0.023064677619934082,0.023164006614685057,0.023568353996276856,"[0.022970367431640625, 0.023136255264282226, 0.023379968643188476, 0.023199743270874023, 0.02310860824584961, 0.02194534492492676, 0.022709247589111328, 0.023167999267578124, 0.021976064682006836, 0.021898239135742188, 0.022373376846313478, 0.0231014404296875, 0.023164928436279295, 0.023186431884765626, 0.021959680557250977, 0.021968896865844727, 0.02191974449157715, 0.02208665657043457, 0.021916671752929686, 0.021955583572387697, 0.02204364776611328, 0.021972991943359374, 0.021781503677368166, 0.021798912048339843, 0.02190336036682129, 0.02186649513244629, 0.02188390350341797, 0.021940223693847655, 0.023406591415405274, 0.0232806396484375, 0.023174144744873046, 0.023183359146118163, 0.021999616622924805, 0.022064128875732423, 0.021737472534179687, 0.02186240005493164, 0.022148096084594726, 0.022434816360473633, 0.022072320938110353, 0.021955583572387697, 0.02209587287902832, 0.023198720932006835, 0.02310553550720215, 0.023088127136230468, 0.024705024719238283, 0.023388160705566406, 0.02308710479736328, 0.02292531204223633, 0.02289151954650879, 0.022944768905639647, 0.02292736053466797, 0.02263039970397949, 0.022979583740234375, 0.022981632232666017, 0.023085056304931642, 0.023018495559692383, 0.023012351989746094, 0.022964223861694336, 0.022975488662719725, 0.022952959060668944, 0.02289356803894043, 0.02294169616699219, 0.023000064849853515, 0.021964799880981444, 0.021934080123901366, 0.022022144317626953, 0.02209791946411133, 0.02285260772705078, 0.02289971160888672, 0.022973440170288087, 0.022929407119750975, 0.022987775802612305, 0.02295910453796387, 0.022993919372558593, 0.022931455612182617, 0.02309017562866211, 0.022969343185424804, 0.02287718391418457, 0.022658048629760744, 0.022973440170288087, 0.023000064849853515, 0.02290995216369629, 0.02303385543823242, 0.02308403205871582, 0.022890495300292968, 0.023030784606933592, 0.022932479858398438, 0.02293350410461426, 0.02288025665283203, 0.02307891273498535, 0.02287718391418457, 0.0229171199798584, 0.022924287796020508, 0.022930431365966796, 0.0229171199798584, 0.02290892791748047, 0.022854656219482423, 0.022915071487426757, 0.023004159927368165, 0.021974016189575195, 0.021970943450927736, 0.021935104370117187, 0.021986303329467775, 0.021970943450927736, 0.022046720504760742, 0.021931007385253908, 0.021953535079956055, 0.021873664855957032, 0.021977088928222657, 0.021816320419311523, 0.02183475112915039, 0.022958080291748048, 0.022787071228027343, 0.022915071487426757, 0.022948863983154297, 0.022948863983154297, 0.022968320846557616, 0.023009279251098632, 0.02292736053466797, 0.022790143966674805, 0.022830080032348633, 0.023003135681152344, 0.022583295822143554, 0.02254745674133301, 0.023034879684448242, 0.022967296600341795, 0.023160831451416015, 0.02290483283996582, 0.021902336120605468, 0.021958656311035156, 0.021976064682006836, 0.022662208557128905, 0.021956544876098633, 0.021918720245361328, 0.021960704803466798, 0.021936128616333008, 0.021954559326171876, 0.02191564750671387, 0.022007808685302735, 0.021933055877685546, 0.021942272186279296, 0.021909503936767577, 0.02191257667541504, 0.02191974449157715, 0.02185625648498535, 0.022928384780883788, 0.022970367431640625, 0.02294169616699219, 0.022936639785766603, 0.022855615615844725, 0.02290995216369629, 0.022897663116455077, 0.023061504364013673, 0.021970943450927736, 0.021935104370117187, 0.0219289608001709, 0.021966848373413086, 0.021955583572387697, 0.021962751388549806, 0.02194534492492676, 0.022016000747680665, 0.02191974449157715, 0.02191155242919922, 0.02191980743408203, 0.021920703887939454, 0.022013952255249023, 0.021952512741088868, 0.021926912307739257, 0.022999040603637694, 0.022974464416503908, 0.02306764793395996, 0.0229171199798584, 0.022953983306884765, 0.02286796760559082, 0.022947839736938477, 0.023010303497314453, 0.023006208419799806, 0.022957056045532227, 0.022944768905639647, 0.02288332748413086, 0.022930431365966796, 0.022947839736938477, 0.022944768905639647, 0.023019519805908203, 0.02306764793395996, 0.023170047760009766, 0.02306662368774414, 0.022951936721801756, 0.022991872787475585, 0.02207744026184082, 0.022025215148925782, 0.0230328311920166, 0.022984703063964843, 0.022929407119750975, 0.022998016357421876, 0.023162879943847657, 0.022983680725097655, 0.022966272354125978, 0.02292736053466797, 0.022964223861694336, 0.022946815490722656, 0.02293452835083008, 0.02310041618347168, 0.023249919891357423, 0.023029760360717775, 0.02289151954650879, 0.022948863983154297, 0.022979583740234375, 0.022936576843261718, 0.02263859176635742, 0.022837247848510742, 0.02190438461303711, 0.0219289608001709, 0.02227609634399414, 0.022999040603637694, 0.021951488494873047, 0.02196793556213379, 0.022519744873046876, 0.02293350410461426, 0.023021568298339845, 0.022937599182128905, 0.022969343185424804, 0.022976512908935546, 0.023188480377197264, 0.022972415924072266, 0.02306252861022949, 0.023005184173583985, 0.02301644706726074, 0.022971391677856445, 0.023109632492065428, 0.022236160278320313, 0.022041599273681642, 0.022832128524780275, 0.022964223861694336, 0.022956031799316406, 0.02292633628845215, 0.022961151123046874, 0.022957056045532227, 0.0230645751953125, 0.022980607986450196, 0.023047168731689452, 0.022973440170288087, 0.022994943618774414, 0.022922239303588866, 0.022874111175537108, 0.02287308883666992, 0.022986751556396484, 0.023004159927368165, 0.022921215057373046, 0.02285260772705078, 0.02294988822937012, 0.02287513542175293, 0.023570432662963867, 0.023600128173828124, 0.02311065673828125, 0.023212032318115236, 0.023051263809204102, 0.022981632232666017, 0.022935552597045897, 0.02285055923461914, 0.02304204750061035, 0.0229171199798584, 0.02285772705078125, 0.022931455612182617, 0.02292531204223633, 0.022998016357421876, 0.022920192718505858, 0.022960128784179686, 0.02288844871520996, 0.02289151954650879, 0.022945791244506835, 0.022853631973266602, 0.022953983306884765, 0.02290278434753418, 0.022988800048828126, 0.02291097640991211, 0.02289254379272461, 0.021950464248657226, 0.021951488494873047, 0.02185523223876953, 0.021995519638061522, 0.02208153533935547, 0.021960704803466798, 0.023563264846801758, 0.023596031188964844, 0.02309836769104004, 0.023069696426391603, 0.022967296600341795, 0.02282803153991699, 0.02304102325439453, 0.02291097640991211, 0.022895616531372072, 0.02290483283996582, 0.022001663208007814, 0.021925888061523437, 0.021833728790283204, 0.02247270393371582, 0.022634496688842775, 0.02286899185180664, 0.023155712127685548, 0.02293350410461426, 0.0228853759765625, 0.02292633628845215, 0.022937599182128905, 0.022968320846557616, 0.023005184173583985, 0.02306559944152832, 0.02309529685974121, 0.022572032928466795, 0.022743040084838868, 0.02285977554321289, 0.022966272354125978, 0.022960128784179686, 0.02288844871520996, 0.022983680725097655, 0.022009855270385743, 0.022649856567382814, 0.023029760360717775, 0.023013376235961915, 0.022990848541259764, 0.0229171199798584, 0.022943744659423827, 0.022980607986450196, 0.022978559494018554, 0.023051263809204102, 0.02264371109008789, 0.02285977554321289, 0.02290892791748047, 0.022961151123046874, 0.022946815490722656, 0.022992895126342772, 0.022930431365966796, 0.02291302490234375, 0.022863872528076173, 0.022947839736938477, 0.02290176010131836, 0.022974464416503908, 0.0229171199798584, 0.022948863983154297, 0.02282803153991699, 0.022958080291748048, 0.023037952423095705, 0.023004159927368165, 0.021947391510009767, 0.021968896865844727, 0.021918720245361328, 0.022642688751220705, 0.022952959060668944, 0.022971391677856445, 0.02290483283996582, 0.02288025665283203, 0.02288332748413086, 0.021876735687255858, 0.022026239395141603, 0.021993471145629884, 0.021894144058227538, 0.022715391159057616, 0.02287615966796875, 0.02287513542175293, 0.02288128089904785, 0.022922239303588866, 0.022866943359375, 0.022683647155761717, 0.022947839736938477, 0.0228853759765625, 0.022923263549804687, 0.02285670471191406, 0.02313932800292969, 0.022961151123046874, 0.02292531204223633, 0.02287308883666992, 0.022854656219482423, 0.02291916847229004, 0.022890495300292968, 0.02290176010131836, 0.022945791244506835, 0.022866943359375, 0.022832128524780275, 0.021982208251953125, 0.02192793655395508, 0.021929983139038087, 0.022536191940307617, 0.022931455612182617, 0.022962175369262695, 0.023208959579467774, 0.02307891273498535, 0.022895616531372072, 0.02285977554321289, 0.021924863815307616, 0.02188390350341797, 0.021952512741088868, 0.022205440521240235, 0.022006784439086914, 0.021971967697143553, 0.021961727142333985, 0.021963775634765623, 0.022047744750976563, 0.021988351821899413, 0.021933055877685546, 0.022999040603637694, 0.023052288055419923, 0.0219238395690918, 0.021965887069702147, 0.022114240646362304, 0.02192793655395508, 0.02193715286254883, 0.022214656829833986, 0.02201190376281738, 0.022074367523193358, 0.021943296432495117, 0.021949440002441405, 0.022005760192871093, 0.021748735427856446, 0.021757951736450197, 0.021948415756225585, 0.021938175201416017, 0.021948415756225585, 0.02192793655395508, 0.021893119812011717, 0.021947391510009767, 0.021909503936767577, 0.021931007385253908, 0.023020544052124024, 0.022968320846557616, 0.023008256912231444, 0.02290073585510254, 0.023000064849853515, 0.02292736053466797, 0.022947839736938477, 0.022982656478881838, 0.022914047241210937, 0.022965248107910157, 0.023026687622070312, 0.022897663116455077, 0.02285977554321289, 0.022940671920776368, 0.02284236717224121, 0.022887424468994142, 0.021925888061523437, 0.022006784439086914, 0.021902336120605468, 0.021943296432495117, 0.02186444854736328, 0.02193715286254883, 0.021943296432495117, 0.02187059211730957, 0.02189516830444336, 0.021926912307739257, 0.021727231979370116, 0.02170982360839844, 0.021935104370117187, 0.02233241653442383, 0.02286489677429199, 0.02290176010131836, 0.02309939193725586, 0.022977535247802734, 0.02290176010131836, 0.022819839477539062, 0.021695487976074217, 0.021702655792236326, 0.021909503936767577, 0.021957632064819335, 0.021993471145629884, 0.022321151733398437, 0.02202726364135742, 0.021974016189575195, 0.021958656311035156, 0.02227097511291504, 0.022031360626220704, 0.023762943267822266, 0.023179264068603517, 0.023113727569580078, 0.023061504364013673, 0.023172096252441408, 0.02295910453796387, 0.023001087188720702, 0.022966272354125978, 0.02352742385864258, 0.024390655517578123, 0.023187456130981447, 0.02304921531677246, 0.022923263549804687, 0.02280963134765625, 0.02288739204406738, 0.023012351989746094, 0.022831104278564454, 0.022932479858398438, 0.022821887969970703, 0.022601728439331056, 0.02280243110656738, 0.022847488403320314, 0.022973440170288087, 0.022947839736938477, 0.022987775802612305, 0.022977535247802734, 0.02309529685974121, 0.022708223342895507, 0.022791168212890626, 0.0231147518157959, 0.02293350410461426, 0.021999616622924805, 0.021959680557250977, 0.021959680557250977, 0.022017023086547852, 0.021740543365478517, 0.022004735946655272, 0.022114303588867186, 0.022305791854858398, 0.022947839736938477, 0.02285670471191406, 0.02306355285644531, 0.022957056045532227, 0.022965248107910157, 0.022916095733642578, 0.022956031799316406, 0.021988351821899413, 0.021949440002441405, 0.02189516830444336, 0.021981184005737304, 0.021991424560546875, 0.022361087799072265, 0.022949920654296876, 0.02294166374206543, 0.022897663116455077, 0.02291916847229004, 0.02294272041320801, 0.022938623428344726, 0.022965248107910157, 0.022947839736938477, 0.02294272041320801, 0.022928384780883788, 0.022923263549804687, 0.022906879425048828, 0.022957056045532227, 0.02292633628845215, 0.0229171199798584, 0.022732799530029296, 0.022750207901000977, 0.022967296600341795, 0.022887424468994142, 0.02188287925720215, 0.021987327575683592, 0.02285875129699707, 0.02386227226257324, 0.023181312561035155, 0.02307788848876953, 0.023008256912231444, 0.022995967864990235, 0.022935552597045897, 0.02295091247558594, 0.022937599182128905, 0.02292633628845215, 0.022956031799316406, 0.02295091247558594, 0.021932031631469725, 0.02193715286254883, 0.022966272354125978, 0.022897663116455077, 0.02305433654785156, 0.023440383911132814, 0.023036928176879884, 0.022982656478881838, 0.022887424468994142, 0.022922239303588866, 0.022657024383544923, 0.02291814422607422, 0.022906879425048828, 0.021974016189575195, 0.02191257667541504, 0.021979135513305666, 0.021959680557250977, 0.021947391510009767, 0.021938175201416017, 0.02272153663635254, 0.0228853759765625, 0.022889471054077147, 0.02266214370727539, 0.021979135513305666, 0.02201907157897949, 0.022033407211303712, 0.02202726364135742, 0.022026239395141603, 0.021975040435791016, 0.02222591972351074, 0.02208665657043457, 0.022006784439086914, 0.021971967697143553, 0.021997568130493163, 0.021969919204711915, 0.021933055877685546, 0.021997568130493163, 0.021972991943359374, 0.021999616622924805, 0.021943296432495117, 0.02207027244567871, 0.0229171199798584, 0.02294169616699219, 0.02289254379272461, 0.02293452835083008, 0.02304102325439453, 0.023009279251098632, 0.023197696685791015, 0.023034879684448242, 0.022840320587158205, 0.022707199096679686, 0.02289971160888672, 0.022931455612182617, 0.022971391677856445, 0.02287001609802246, 0.022946815490722656, 0.02289356803894043, 0.022849536895751952, 0.023006208419799806, 0.022976512908935546, 0.023020544052124024, 0.022930431365966796, 0.02327244758605957, 0.022972415924072266, 0.022970367431640625, 0.02291916847229004, 0.022957056045532227, 0.022985727310180663, 0.023043071746826172, 0.022982656478881838, 0.022974464416503908, 0.022937599182128905, 0.022922239303588866, 0.023120895385742187, 0.023006208419799806, 0.02268569564819336]",tokens/s,44.15312863855685,,,,,,MB,933.494784,6768.033792,0.0,6138.363904,6060.931072,s,1,6.9301474609375,6.9301474609375,0.0,6.9301474609375,6.9301474609375,6.9301474609375,6.9301474609375,[6.9301474609375],,kWh,1.691654853474726e-06,8.749000186634714e-07,1.764723634001264e-06,4.3312785061394615e-06, float16-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,s,s,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/s/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6694734e-239c6f3a377a159a27825dff;f03a233b-84fe-4c0d-88a5-1579097ed82b) Repository Not Found for url: https://huggingface.co/s/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: s is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-40b,tiiuae/falcon-40b,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ self.load_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 250, in load_model_with_no_weights self.pretrained_model = self.automodel_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 559, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3710, in from_pretrained model = cls(config, *model_args, **model_kwargs) File ""/root/.cache/huggingface/modules/transformers_modules/tiiuae/falcon-40b/4a70170c215b36a3cce4b4253f6d0612bb7d4146/modeling_falcon.py"", line 843, in __init__ self.transformer = FalconModel(config) File ""/root/.cache/huggingface/modules/transformers_modules/tiiuae/falcon-40b/4a70170c215b36a3cce4b4253f6d0612bb7d4146/modeling_falcon.py"", line 650, in __init__ self.h = nn.ModuleList([FalconDecoderLayer(config) for _ in range(config.num_hidden_layers)]) File ""/root/.cache/huggingface/modules/transformers_modules/tiiuae/falcon-40b/4a70170c215b36a3cce4b4253f6d0612bb7d4146/modeling_falcon.py"", line 650, in self.h = nn.ModuleList([FalconDecoderLayer(config) for _ in range(config.num_hidden_layers)]) File ""/root/.cache/huggingface/modules/transformers_modules/tiiuae/falcon-40b/4a70170c215b36a3cce4b4253f6d0612bb7d4146/modeling_falcon.py"", line 420, in __init__ self.mlp = FalconMLP(config) File ""/root/.cache/huggingface/modules/transformers_modules/tiiuae/falcon-40b/4a70170c215b36a3cce4b4253f6d0612bb7d4146/modeling_falcon.py"", line 403, in __init__ self.dense_h_to_4h = FalconLinear(hidden_size, 4 * hidden_size, bias=config.bias) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/linear.py"", line 98, in __init__ self.weight = Parameter(torch.empty((out_features, in_features), **factory_kwargs)) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_device.py"", line 78, in __torch_function__ return func(*args, **kwargs) torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 512.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,google/gemma-7b,google/gemma-7b,cuda,0,42,,,True,,,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.223-212.873.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.0,,,,1.21.4,,,,0.12.0,,,gemma,MB,1562.554368,19300.614144,0.0,18654.16704,17596.665856,s,10,0.9506052551269532,0.09506052551269531,9.581950041984465e-05,0.09504988861083985,0.09517140808105469,0.09521050567626953,0.09524178375244141,"[0.095111328125, 0.09490370941162109, 0.0951627197265625, 0.09503404998779297, 0.0950564193725586, 0.09524960327148438, 0.09500150299072266, 0.09494755554199219, 0.09509500885009765, 0.09504335784912109]",tokens/s,2693.021089661568,kWh,1.1234152975563459e-06,6.15574724968647e-07,6.625815049075546e-06,8.364805071600539e-06,tokens/kWh,30604419.087917425,MB,1634.275328,19302.711296,0.0,18656.264192,17596.668416,s,10,27.72810009765625,2.772810009765625,0.0006644198257382086,2.7725189208984373,2.7736293701171872,2.7737610961914063,2.773866477050781,"[2.773892822265625, 2.7715537109375, 2.77253466796875, 2.773227783203125, 2.772401123046875, 2.77250048828125, 2.773388427734375, 2.772497802734375, 2.77360009765625, 2.772503173828125]",tokens/s,22.720633501076097,kWh,3.272707584959614e-05,1.7935850741894317e-05,0.00014161377576472374,0.00019227670235621416,tokens/kWh,327652.8005108255,,s,630,27.72617106246949,0.044009795337253146,0.0001611674955760164,0.04398694229125977,0.04412764129638672,0.04431482963562012,0.04471432319641113,"[0.044022785186767575, 0.04384972763061523, 0.043807743072509765, 0.043824127197265625, 0.04388249588012695, 0.04382207870483398, 0.043858943939208986, 0.04401152038574219, 0.04386406326293945, 0.04381491088867188, 0.043840511322021485, 0.0438476791381836, 0.04394905471801758, 0.04394291305541992, 0.04442214584350586, 0.04415692901611328, 0.04389376068115235, 0.04386918258666992, 0.04392243194580078, 0.04388556671142578, 0.04392652893066406, 0.04390707015991211, 0.04389068984985352, 0.04402380752563476, 0.04392755126953125, 0.04420198440551758, 0.043930622100830076, 0.04395827102661133, 0.04403507232666016, 0.043990016937255856, 0.04395622253417969, 0.044045310974121094, 0.04396851348876953, 0.04401561737060547, 0.04425830459594727, 0.04404326248168945, 0.04400537490844727, 0.04405964660644531, 0.04451635360717773, 0.044050430297851564, 0.043979774475097655, 0.0439920654296875, 0.04399923324584961, 0.044521472930908204, 0.04405350494384765, 0.04401971054077149, 0.044014591217041016, 0.04406169509887695, 0.044012542724609374, 0.044077056884765625, 0.044065792083740236, 0.044009471893310545, 0.044050430297851564, 0.044022785186767575, 0.04405350494384765, 0.044065792083740236, 0.044047359466552735, 0.044068862915039066, 0.04414156723022461, 0.044058624267578124, 0.04412723159790039, 0.04485222244262695, 0.044219390869140625, 0.0440709114074707, 0.04384153747558594, 0.04379033660888672, 0.043824127197265625, 0.04379955291748047, 0.04387123107910156, 0.04381491088867188, 0.04392345428466797, 0.043835391998291014, 0.04387635040283203, 0.043826175689697267, 0.04383846282958984, 0.04386816024780273, 0.043894783020019534, 0.04382720184326172, 0.04397568130493164, 0.04389068984985352, 0.044066814422607424, 0.043979774475097655, 0.04396441650390625, 0.04402687835693359, 0.04392550277709961, 0.04389273452758789, 0.04400435256958008, 0.04397158432006836, 0.043954177856445314, 0.04391424179077148, 0.043936767578125, 0.04392652893066406, 0.04395110321044922, 0.04393471908569336, 0.04413542556762695, 0.04394905471801758, 0.04433100891113281, 0.044022785186767575, 0.04397568130493164, 0.04393164825439453, 0.04398796844482422, 0.04392959976196289, 0.044060672760009766, 0.04400128173828125, 0.044335105895996096, 0.04398899078369141, 0.04483583831787109, 0.044047359466552735, 0.044009471893310545, 0.0439736328125, 0.04407398223876953, 0.0439818229675293, 0.04400332641601563, 0.04397158432006836, 0.0439920654296875, 0.044071937561035154, 0.04402380752563476, 0.04401561737060547, 0.044088321685791014, 0.0440412483215332, 0.044035041809082034, 0.044047359466552735, 0.04403200149536133, 0.044065792083740236, 0.04405452728271484, 0.04414361572265625, 0.04410572814941406, 0.043837440490722655, 0.04391526412963867, 0.04384972763061523, 0.043815937042236325, 0.043853824615478515, 0.04382207870483398, 0.04394496154785156, 0.04388454437255859, 0.04390707015991211, 0.04389273452758789, 0.04388147354125976, 0.043930622100830076, 0.04390092849731445, 0.04395622253417969, 0.04395212936401367, 0.04389068984985352, 0.043936767578125, 0.04386611175537109, 0.04388044738769531, 0.043911167144775394, 0.04435763168334961, 0.04392345428466797, 0.04399923324584961, 0.04391321563720703, 0.04452867126464844, 0.04406371307373047, 0.04396646499633789, 0.04406988906860351, 0.04395315170288086, 0.04394291305541992, 0.04401049423217773, 0.043979774475097655, 0.04396543884277344, 0.04395110321044922, 0.0439552001953125, 0.04397260665893555, 0.0439920654296875, 0.04398284912109375, 0.04403302383422852, 0.04400844955444336, 0.04405657577514648, 0.044036094665527346, 0.04399411010742187, 0.04404121780395508, 0.04401561737060547, 0.044036094665527346, 0.04418560028076172, 0.04416716766357422, 0.044131328582763675, 0.04402483367919922, 0.04398387145996094, 0.04439449691772461, 0.04404326248168945, 0.04403200149536133, 0.04406476974487305, 0.04405145645141602, 0.04414566421508789, 0.04405452728271484, 0.044065792083740236, 0.044058624267578124, 0.04416819381713867, 0.04408627319335937, 0.04472115325927734, 0.04386406326293945, 0.0438476791381836, 0.04379852676391602, 0.04389376068115235, 0.04387430572509766, 0.043843582153320314, 0.044440574645996093, 0.04389068984985352, 0.04393574523925781, 0.043894783020019534, 0.043910144805908206, 0.043976703643798826, 0.04395212936401367, 0.043850753784179686, 0.04398489761352539, 0.04387123107910156, 0.04388351821899414, 0.04390707015991211, 0.043886592864990234, 0.043889663696289063, 0.04392550277709961, 0.04389376068115235, 0.043996158599853515, 0.0439285774230957, 0.043960319519042966, 0.04409958267211914, 0.04397568130493164, 0.044075008392333984, 0.0439818229675293, 0.044044288635253906, 0.044147712707519535, 0.04395110321044922, 0.04397875213623047, 0.043957248687744144, 0.04396441650390625, 0.04397055816650391, 0.04398489761352539, 0.043996158599853515, 0.044112895965576174, 0.04399513626098633, 0.04402483367919922, 0.04401152038574219, 0.044027904510498046, 0.04402483367919922, 0.04405657577514648, 0.04401049423217773, 0.04435865783691406, 0.044006401062011716, 0.04404326248168945, 0.04402073669433594, 0.044025856018066405, 0.04405452728271484, 0.04405759811401367, 0.044475391387939454, 0.04414976119995117, 0.044022785186767575, 0.04408729553222656, 0.04410572814941406, 0.04409446334838867, 0.044147712707519535, 0.044104705810546874, 0.04408115386962891, 0.044009471893310545, 0.043835391998291014, 0.04388044738769531, 0.043837440490722655, 0.04386713409423828, 0.04383129501342774, 0.04381184005737305, 0.04395110321044922, 0.04384153747558594, 0.04386304092407227, 0.043842559814453126, 0.04386816024780273, 0.04483891296386719, 0.04398591995239258, 0.043842559814453126, 0.0439285774230957, 0.04391628646850586, 0.04389376068115235, 0.04388044738769531, 0.04389785766601562, 0.043930622100830076, 0.04389785766601562, 0.04391219329833984, 0.04400332641601563, 0.04391219329833984, 0.043940864562988284, 0.043910144805908206, 0.044388351440429685, 0.043974655151367184, 0.04395315170288086, 0.04392243194580078, 0.044014591217041016, 0.043947006225585936, 0.04394188690185547, 0.04394803237915039, 0.04417433547973633, 0.044229633331298826, 0.04398080062866211, 0.04397260665893555, 0.044034046173095705, 0.043960319519042966, 0.04397568130493164, 0.04395008087158203, 0.043993087768554685, 0.044006401062011716, 0.044009471893310545, 0.044025856018066405, 0.04406476974487305, 0.04409139251708984, 0.04416819381713867, 0.04409241485595703, 0.044025856018066405, 0.04408627319335937, 0.04404326248168945, 0.04402073669433594, 0.044085247039794925, 0.04405145645141602, 0.044044288635253906, 0.044030975341796875, 0.04458905410766602, 0.04415590286254883, 0.04409139251708984, 0.044045310974121094, 0.04476006317138672, 0.04382822418212891, 0.04382822418212891, 0.04382515335083008, 0.04380876922607422, 0.043834369659423826, 0.04383129501342774, 0.04395008087158203, 0.04391424179077148, 0.043875328063964845, 0.04384972763061523, 0.043835391998291014, 0.04389580917358398, 0.04390707015991211, 0.043902976989746094, 0.04397158432006836, 0.043870208740234375, 0.04401663970947266, 0.044096511840820314, 0.043878398895263675, 0.04391526412963867, 0.04393881607055664, 0.04391424179077148, 0.04399513626098633, 0.043919361114501954, 0.043963390350341795, 0.043963390350341795, 0.04394905471801758, 0.0439552001953125, 0.0439920654296875, 0.044071937561035154, 0.044055553436279295, 0.043979774475097655, 0.04396441650390625, 0.043960319519042966, 0.04396236801147461, 0.04400230407714844, 0.04398387145996094, 0.0439920654296875, 0.04402380752563476, 0.04400537490844727, 0.04478668975830078, 0.044095489501953126, 0.043963390350341795, 0.04403200149536133, 0.04404940795898438, 0.04405452728271484, 0.044058624267578124, 0.04402175903320313, 0.04403302383422852, 0.04405964660644531, 0.04400332641601563, 0.04407398223876953, 0.04431257629394531, 0.04405657577514648, 0.044058624267578124, 0.044065792083740236, 0.044068862915039066, 0.044077056884765625, 0.04403200149536133, 0.04406476974487305, 0.044080127716064454, 0.04408422470092774, 0.04469760131835938, 0.04384153747558594, 0.04380364990234375, 0.043817985534667966, 0.04379443359375, 0.043843582153320314, 0.043832321166992184, 0.04396441650390625, 0.043908096313476565, 0.04392755126953125, 0.04391526412963867, 0.04388147354125976, 0.043908096313476565, 0.0439183349609375, 0.04388761520385742, 0.04398694229125977, 0.043856895446777344, 0.04390604782104492, 0.043894783020019534, 0.04391628646850586, 0.043957248687744144, 0.043993087768554685, 0.04388454437255859, 0.045415424346923826, 0.043963390350341795, 0.04396543884277344, 0.04392652893066406, 0.04396441650390625, 0.04398796844482422, 0.04396953582763672, 0.04394291305541992, 0.04436172866821289, 0.04405452728271484, 0.044200958251953124, 0.043963390350341795, 0.0439285774230957, 0.043943935394287106, 0.04397260665893555, 0.04399923324584961, 0.044065792083740236, 0.043963390350341795, 0.04399923324584961, 0.0439736328125, 0.04401971054077149, 0.04397568130493164, 0.044028926849365234, 0.044030975341796875, 0.044183551788330076, 0.04401152038574219, 0.04403712081909179, 0.04402687835693359, 0.04403200149536133, 0.0440893440246582, 0.044066814422607424, 0.044045310974121094, 0.04413337707519531, 0.044077056884765625, 0.044085247039794925, 0.04407295989990234, 0.04409958267211914, 0.04407910537719727, 0.044098560333251956, 0.044111873626708986, 0.04407295989990234, 0.043829246520996096, 0.04389580917358398, 0.04381695938110351, 0.044385280609130856, 0.04394803237915039, 0.043834369659423826, 0.043921409606933595, 0.043832321166992184, 0.04383846282958984, 0.043850753784179686, 0.043878398895263675, 0.04390707015991211, 0.04393369674682617, 0.04389785766601562, 0.043974655151367184, 0.04389990234375, 0.043875328063964845, 0.043858943939208986, 0.04392345428466797, 0.04391424179077148, 0.04391628646850586, 0.043902976989746094, 0.0445030403137207, 0.043976703643798826, 0.04394496154785156, 0.043908096313476565, 0.043974655151367184, 0.044336128234863284, 0.04403814315795898, 0.043930622100830076, 0.04402175903320313, 0.043979774475097655, 0.04393369674682617, 0.0439183349609375, 0.0441610221862793, 0.04405350494384765, 0.04399513626098633, 0.043937793731689455, 0.044050430297851564, 0.043996158599853515, 0.04400128173828125, 0.04407398223876953, 0.04402687835693359, 0.04401049423217773, 0.04401971054077149, 0.04404326248168945, 0.04405759811401367, 0.0440709114074707, 0.044050430297851564, 0.04402073669433594, 0.044075008392333984, 0.044349441528320314, 0.04403302383422852, 0.04400537490844727, 0.044095489501953126, 0.044052478790283206, 0.044047359466552735, 0.044037185668945315, 0.0442305908203125, 0.04411904144287109, 0.04408115386962891, 0.044050430297851564, 0.04414156723022461, 0.04387635040283203, 0.04385177612304687, 0.043842559814453126, 0.043848705291748044, 0.043850753784179686, 0.04385279846191406, 0.04395622253417969, 0.043943935394287106, 0.04391526412963867, 0.04439449691772461, 0.043883583068847654, 0.04391417694091797, 0.0439183349609375, 0.043937793731689455, 0.044478462219238284, 0.04403507232666016, 0.04396543884277344, 0.04390195083618164, 0.043875328063964845, 0.04391731262207031, 0.044025856018066405, 0.04392550277709961, 0.04398284912109375, 0.04390911865234375, 0.04393574523925781, 0.04396134567260742, 0.04391526412963867, 0.04396646499633789, 0.043998207092285156, 0.044014591217041016, 0.044491775512695314, 0.0440074234008789, 0.04398489761352539, 0.044047359466552735, 0.04396236801147461, 0.043974655151367184, 0.04396748733520508, 0.043979774475097655, 0.044170238494873046, 0.044045310974121094, 0.044014591217041016, 0.04399513626098633, 0.04397875213623047, 0.04401766586303711, 0.04401152038574219, 0.044036094665527346, 0.044068862915039066, 0.044039169311523435, 0.04403712081909179, 0.04401971054077149, 0.044042240142822264, 0.044088321685791014, 0.04412211227416992, 0.04405145645141602, 0.04414054489135742, 0.04408627319335937, 0.04465971374511719, 0.044112895965576174, 0.044071937561035154, 0.04409241485595703, 0.044058624267578124, 0.04410060882568359, 0.04426335906982422, 0.04382720184326172, 0.04380672073364258, 0.043826175689697267, 0.043796478271484376, 0.043865089416503904, 0.04383027267456055, 0.04393267059326172, 0.043888641357421876, 0.043930622100830076, 0.04390911865234375, 0.04392652893066406, 0.04392038345336914, 0.04395008087158203, 0.043886592864990234, 0.04400435256958008, 0.04398694229125977, 0.043910144805908206, 0.04390195083618164, 0.04388761520385742, 0.04391731262207031, 0.04393983840942383, 0.04394803237915039, 0.04403302383422852, 0.04391219329833984, 0.04397055816650391, 0.043943935394287106, 0.04398080062866211, 0.0439818229675293, 0.043976703643798826, 0.04396441650390625, 0.04405964660644531, 0.04397875213623047, 0.04396748733520508, 0.043960319519042966, 0.04391424179077148, 0.0440074234008789, 0.04396134567260742, 0.04396236801147461, 0.044660736083984375, 0.04401971054077149, 0.04403200149536133, 0.04403507232666016, 0.04402483367919922, 0.04402380752563476, 0.044036094665527346, 0.04400844955444336, 0.04410163116455078, 0.044009471893310545, 0.04406784057617188, 0.04405350494384765, 0.0440709114074707, 0.04405350494384765, 0.044115966796875, 0.044045310974121094, 0.04412416076660156, 0.044224510192871096, 0.04407603073120117, 0.044050430297851564, 0.04425523376464844, 0.04406272125244141, 0.04410879898071289, 0.044316673278808595]",tokens/s,22.722214278363754,,,,,,MB,935.223296,19279.642624,0.0,18649.972736,18648.241152,s,1,7.0098154296875,7.0098154296875,0.0,7.0098154296875,7.0098154296875,7.0098154296875,7.0098154296875,[7.0098154296875],,kWh,2.4941218027789904e-06,1.2600992074067204e-06,8.220839909998945e-06,1.1975060920184656e-05, float16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-125m,,cuda,0,42,,,,,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,d35829e539df8480b726c647eeabf91e41eae047,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,,MB,1198.358528,1019.74016,0.0,373.293056,302.523392,s,23,0.19381337833404544,0.008426668623219365,0.0002819819868018276,0.0083372802734375,0.00853413143157959,0.00866414737701416,0.009456716499328615,"[0.009676927566528321, 0.008425344467163086, 0.008417535781860351, 0.008675968170166016, 0.008270208358764648, 0.008324000358581543, 0.008313920021057129, 0.008333312034606934, 0.008371935844421387, 0.008271424293518067, 0.008366559982299805, 0.008431455612182618, 0.008307040214538574, 0.0083372802734375, 0.00855776023864746, 0.008357184410095215, 0.00829798412322998, 0.008380096435546875, 0.008328127861022949, 0.008439616203308105, 0.008334976196289062, 0.008324607849121094, 0.008270112037658692]",tokens/s,30379.739781697564,kWh,1.0079676041695311e-07,5.523167517390536e-08,2.3742387215359032e-07,3.934523077444488e-07,tokens/kWh,650650650.5644251,MB,1198.686208,1019.74016,0.0,373.293056,312.415744,s,23,10.091357208251953,0.4387546612283458,0.005665811472044785,0.43642874145507815,0.44653612060546877,0.4503826171875,0.4552436303710937,"[0.4565221252441406, 0.45071078491210936, 0.44213473510742185, 0.4374317626953125, 0.4350805969238281, 0.4375250549316406, 0.4402088623046875, 0.4383892211914063, 0.43448800659179687, 0.43642874145507815, 0.4362821655273437, 0.43523263549804686, 0.43606253051757815, 0.4474291076660156, 0.4365704040527344, 0.4339928283691406, 0.4396087646484375, 0.43543649291992187, 0.4351995544433594, 0.43567752075195315, 0.43433328247070313, 0.4336478576660156, 0.44296417236328123]",tokens/s,143.58821812541893,kWh,5.1311319713792945e-06,2.8116294429456024e-06,8.951218744456647e-06,1.6893980158781544e-05,tokens/kWh,3729138.983701979,,s,1448,10.267983878612535,0.007091149087439585,0.0010863371160162378,0.006908927917480469,0.007153971242904663,0.007261788797378539,0.01566075898170471,"[0.007955455780029297, 0.007928832054138184, 0.007986176013946533, 0.008102911949157715, 0.007932928085327149, 0.0070830078125, 0.006937600135803222, 0.006919167995452881, 0.00694374418258667, 0.0069212160110473635, 0.006934591770172119, 0.0068720002174377445, 0.006945792198181152, 0.006917119979858399, 0.0076574721336364745, 0.007366655826568603, 0.007294976234436036, 0.007325695991516113, 0.007302144050598145, 0.007311359882354736, 0.007290880203247071, 0.007333888053894043, 0.007372799873352051, 0.007305280208587646, 0.007324607849121094, 0.007300096035003662, 0.007300096035003662, 0.007286784172058106, 0.00729804801940918, 0.007312384128570557, 0.0073062400817871095, 0.007314432144165039, 0.007296000003814697, 0.007324672222137451, 0.007384064197540283, 0.007428095817565918, 0.007174143791198731, 0.0071905279159545895, 0.007174143791198731, 0.007200767993927002, 0.007170048236846924, 0.007145472049713135, 0.00719974422454834, 0.007124991893768311, 0.007124991893768311, 0.007088128089904785, 0.007090176105499267, 0.007116799831390381, 0.0069959678649902345, 0.0069632000923156735, 0.006957056045532227, 0.006945792198181152, 0.0069027838706970214, 0.007124991893768311, 0.00713321590423584, 0.007165919780731201, 0.007264319896697998, 0.007127999782562256, 0.007102464199066162, 0.007155712127685547, 0.006917119979858399, 0.0069110078811645504, 0.015962080001831056, 0.006993919849395752, 0.007019519805908203, 0.007121920108795166, 0.007122943878173828, 0.007046144008636474, 0.007121920108795166, 0.0071198720932006835, 0.007168000221252442, 0.006982656002044678, 0.007122943878173828, 0.00719974422454834, 0.0071485438346862796, 0.007130112171173096, 0.007194623947143554, 0.007035903930664063, 0.007207935810089112, 0.007342080116271973, 0.007350272178649903, 0.007365632057189942, 0.007450623989105225, 0.007463935852050781, 0.007402495861053467, 0.007188479900360107, 0.007110655784606934, 0.007208960056304932, 0.007180287837982178, 0.0071916160583496095, 0.0071699838638305665, 0.007060480117797851, 0.007126016139984131, 0.007077888011932373, 0.007001088142395019, 0.00695091199874878, 0.006964223861694336, 0.006916096210479736, 0.00694271993637085, 0.006903808116912841, 0.007769087791442871, 0.007549952030181885, 0.007214079856872559, 0.007160831928253173, 0.0071198720932006835, 0.007153664112091064, 0.0071485438346862796, 0.007139328002929687, 0.00714137601852417, 0.007116799831390381, 0.007102464199066162, 0.007004159927368164, 0.0074035201072692874, 0.00725708818435669, 0.0070860800743103025, 0.007113728046417236, 0.007134208202362061, 0.007109632015228271, 0.007094272136688232, 0.007180287837982178, 0.007154687881469727, 0.007109632015228271, 0.007030784130096435, 0.006892543792724609, 0.006936575889587402, 0.015773695945739748, 0.006959104061126709, 0.006918144226074219, 0.00689356803894043, 0.00692633581161499, 0.006923264026641846, 0.006910975933074951, 0.006959104061126709, 0.007044095993041993, 0.007007232189178467, 0.006914048194885254, 0.006843391895294189, 0.006845439910888672, 0.006814720153808594, 0.006812672138214112, 0.00684441614151001, 0.006884352207183838, 0.006814720153808594, 0.006901823997497559, 0.006914015769958496, 0.006908895969390869, 0.006896639823913574, 0.006924287796020508, 0.006973440170288086, 0.0068884482383728025, 0.006884352207183838, 0.0069027838706970214, 0.006912000179290771, 0.007041024208068848, 0.007107583999633789, 0.007095295906066895, 0.007094272136688232, 0.007094304084777832, 0.006995935916900634, 0.006944767951965332, 0.006927360057830811, 0.006890560150146484, 0.006874048233032227, 0.006913023948669434, 0.006903840065002441, 0.0068576960563659665, 0.006899712085723877, 0.006909952163696289, 0.006897664070129395, 0.006917119979858399, 0.006923264026641846, 0.00690176010131836, 0.006904831886291504, 0.006876160144805908, 0.007795711994171142, 0.007635968208312988, 0.007323647975921631, 0.008150015830993652, 0.007303167819976806, 0.007164927959442138, 0.007114751815795899, 0.007163904190063477, 0.007177216053009033, 0.007130112171173096, 0.007071743965148926, 0.007112703800201416, 0.007097343921661377, 0.007169023990631103, 0.016236543655395508, 0.007109632015228271, 0.007098368167877197, 0.007159808158874512, 0.00693555212020874, 0.0071157760620117185, 0.007180287837982178, 0.007103487968444824, 0.007109632015228271, 0.0071495680809020995, 0.0070860800743103025, 0.007027711868286133, 0.006953983783721924, 0.006964223861694336, 0.006870016098022461, 0.006873087882995605, 0.006924287796020508, 0.006946815967559815, 0.006890528202056885, 0.00689353609085083, 0.006909952163696289, 0.00687820816040039, 0.0068853759765625, 0.0069324798583984375, 0.006914048194885254, 0.006923264026641846, 0.006852608203887939, 0.006930431842803955, 0.006909952163696289, 0.006873087882995605, 0.006896639823913574, 0.0069027838706970214, 0.006884352207183838, 0.006894591808319092, 0.0069283838272094726, 0.006907904148101806, 0.006899712085723877, 0.006872064113616944, 0.006892543792724609, 0.006907904148101806, 0.006912000179290771, 0.0069212160110473635, 0.006920191764831543, 0.006894591808319092, 0.006881279945373535, 0.006895616054534912, 0.00692633581161499, 0.006889472007751465, 0.006858751773834228, 0.006909952163696289, 0.006884352207183838, 0.006855679988861084, 0.006900735855102539, 0.006894591808319092, 0.006881279945373535, 0.006929408073425293, 0.0068853759765625, 0.006882304191589355, 0.006845439910888672, 0.006889472007751465, 0.006894591808319092, 0.006883359909057617, 0.00690070390701294, 0.015736831665039062, 0.0069550080299377445, 0.006904831886291504, 0.006859776020050049, 0.006910975933074951, 0.0069283838272094726, 0.006897664070129395, 0.006867968082427979, 0.0069027838706970214, 0.0069212160110473635, 0.00693555212020874, 0.006883327960968018, 0.00683622407913208, 0.00682809591293335, 0.0070061440467834475, 0.00713318395614624, 0.007106560230255127, 0.007017471790313721, 0.006865920066833496, 0.006908927917480469, 0.006909952163696289, 0.006876160144805908, 0.006907904148101806, 0.0068884482383728025, 0.006899712085723877, 0.006918144226074219, 0.006872064113616944, 0.00692633581161499, 0.006913023948669434, 0.006884352207183838, 0.006910975933074951, 0.006924287796020508, 0.006900735855102539, 0.006882304191589355, 0.006933504104614257, 0.00689356803894043, 0.006896639823913574, 0.006851583957672119, 0.006817791938781738, 0.0068280320167541505, 0.0067870721817016606, 0.006841343879699707, 0.006818816184997558, 0.006854720115661621, 0.0068904318809509275, 0.006959104061126709, 0.006880256175994873, 0.006978559970855713, 0.006892543792724609, 0.006905888080596924, 0.006881247997283935, 0.006864895820617676, 0.00690176010131836, 0.00688640022277832, 0.006864895820617676, 0.006880256175994873, 0.006887423992156983, 0.006870016098022461, 0.00688640022277832, 0.006846464157104492, 0.006894591808319092, 0.006778880119323731, 0.00687718391418457, 0.01567129611968994, 0.006971392154693603, 0.0068986878395080565, 0.006896639823913574, 0.0069918718338012695, 0.006910975933074951, 0.006884352207183838, 0.006864895820617676, 0.006905856132507324, 0.006884352207183838, 0.0068915200233459475, 0.006919167995452881, 0.007035903930664063, 0.0069550080299377445, 0.006889472007751465, 0.006905856132507324, 0.006848512172698974, 0.006797311782836914, 0.006907904148101806, 0.007085055828094483, 0.0070256638526916505, 0.00690176010131836, 0.007136256217956543, 0.006934591770172119, 0.0069836158752441405, 0.0068986878395080565, 0.0068915200233459475, 0.007080959796905518, 0.007147520065307617, 0.007058432102203369, 0.0070860800743103025, 0.007211008071899414, 0.0070891518592834475, 0.007048192024230957, 0.007010303974151612, 0.006987808227539062, 0.00690992021560669, 0.006897664070129395, 0.006895616054534912, 0.006936575889587402, 0.006909952163696289, 0.006897664070129395, 0.006903808116912841, 0.006918144226074219, 0.006890495777130127, 0.006879231929779053, 0.0068986878395080565, 0.00688640022277832, 0.006879231929779053, 0.006870016098022461, 0.006910975933074951, 0.0068884482383728025, 0.0068577280044555666, 0.0069027838706970214, 0.0069253120422363285, 0.006855679988861084, 0.006895616054534912, 0.006899712085723877, 0.00689356803894043, 0.006910975933074951, 0.006915071964263916, 0.006919167995452881, 0.006905856132507324, 0.0156559362411499, 0.0068915200233459475, 0.006912000179290771, 0.006922304153442383, 0.006922175884246826, 0.006903840065002441, 0.006915040016174316, 0.006904831886291504, 0.007027711868286133, 0.006970367908477783, 0.007138304233551026, 0.007107583999633789, 0.007010303974151612, 0.00693555212020874, 0.00695091199874878, 0.006930431842803955, 0.007301119804382325, 0.007008255958557129, 0.007076863765716553, 0.007074816226959229, 0.007103487968444824, 0.007095295906066895, 0.0071188478469848635, 0.007229440212249756, 0.007097343921661377, 0.0071188478469848635, 0.007097343921661377, 0.00714035177230835, 0.007256063938140869, 0.007069695949554444, 0.00709939193725586, 0.007180287837982178, 0.007165952205657959, 0.007018496036529541, 0.006996992111206054, 0.006952960014343262, 0.006845439910888672, 0.00682700777053833, 0.0068076162338256835, 0.006837183952331543, 0.006843391895294189, 0.0070348801612854, 0.006990848064422607, 0.007021567821502686, 0.007013376235961914, 0.007000063896179199, 0.006927360057830811, 0.006843391895294189, 0.006814720153808594, 0.006817791938781738, 0.006852608203887939, 0.006862847805023193, 0.006816768169403077, 0.006802432060241699, 0.00682912015914917, 0.006804416179656983, 0.006863872051239014, 0.006838272094726563, 0.007211008071899414, 0.007020544052124023, 0.0069027838706970214, 0.006905856132507324, 0.00693452787399292, 0.015681535720825195, 0.006899712085723877, 0.0070225920677185055, 0.006938623905181885, 0.00689356803894043, 0.006882304191589355, 0.006927360057830811, 0.006900735855102539, 0.006896704196929932, 0.007182271957397461, 0.008265727996826172, 0.00748748779296875, 0.007187456130981445, 0.007197696208953858, 0.00710041618347168, 0.0070225920677185055, 0.006994944095611572, 0.006895616054534912, 0.006905856132507324, 0.006903808116912841, 0.006918144226074219, 0.006900735855102539, 0.006905856132507324, 0.006875135898590088, 0.0068986878395080565, 0.006916096210479736, 0.00683622407913208, 0.006907904148101806, 0.006941696166992187, 0.0068853759765625, 0.006864895820617676, 0.006916096210479736, 0.006899712085723877, 0.006858751773834228, 0.006899712085723877, 0.006889472007751465, 0.006897664070129395, 0.006880256175994873, 0.006892543792724609, 0.006958079814910889, 0.006899712085723877, 0.006882304191589355, 0.006910975933074951, 0.006953983783721924, 0.006899712085723877, 0.006894591808319092, 0.0068853759765625, 0.006889472007751465, 0.006859776020050049, 0.0068986878395080565, 0.00688640022277832, 0.006870016098022461, 0.0068986878395080565, 0.006905856132507324, 0.006917183876037598, 0.006861760139465332, 0.006892543792724609, 0.006917119979858399, 0.006864895820617676, 0.006879231929779053, 0.006896639823913574, 0.006947840213775635, 0.006904831886291504, 0.01567129611968994, 0.006894591808319092, 0.006916096210479736, 0.006880256175994873, 0.006894591808319092, 0.006976511955261231, 0.00689356803894043, 0.006889472007751465, 0.006883327960968018, 0.006945792198181152, 0.006906879901885986, 0.006922239780426025, 0.006947840213775635, 0.006977536201477051, 0.006884352207183838, 0.0068853759765625, 0.006922239780426025, 0.006907904148101806, 0.006871039867401123, 0.006874144077301025, 0.006907872200012207, 0.0068884482383728025, 0.006904831886291504, 0.00690176010131836, 0.006883327960968018, 0.0068884482383728025, 0.006856704235076904, 0.00688640022277832, 0.006887423992156983, 0.006870016098022461, 0.006879231929779053, 0.006897664070129395, 0.0068884482383728025, 0.006892543792724609, 0.006909952163696289, 0.006895616054534912, 0.00694374418258667, 0.006881279945373535, 0.006889472007751465, 0.006896639823913574, 0.00687718391418457, 0.006904831886291504, 0.006944767951965332, 0.006860799789428711, 0.006862847805023193, 0.0069283838272094726, 0.0068884482383728025, 0.006862912178039551, 0.0068863358497619626, 0.0068351998329162595, 0.006790143966674805, 0.006803455829620361, 0.006820864200592041, 0.006874112129211426, 0.006806528091430664, 0.006910975933074951, 0.006900735855102539, 0.00687820816040039, 0.00688640022277832, 0.006889472007751465, 0.006903808116912841, 0.006952960014343262, 0.0068403840065002445, 0.01559443187713623, 0.0069283838272094726, 0.006866943836212158, 0.006890495777130127, 0.0069621758460998535, 0.007044095993041993, 0.007037983894348144, 0.007015391826629639, 0.007105535984039306, 0.007098368167877197, 0.006976511955261231, 0.006951935768127441, 0.006973440170288086, 0.006969344139099121, 0.007015423774719238, 0.006922239780426025, 0.007013376235961914, 0.0069621758460998535, 0.006957056045532227, 0.006990848064422607, 0.006982656002044678, 0.006968319892883301, 0.0069816322326660156, 0.006876160144805908, 0.006903808116912841, 0.006890495777130127, 0.006905856132507324, 0.0068853759765625, 0.006896639823913574, 0.006882304191589355, 0.006887423992156983, 0.006912000179290771, 0.006890528202056885, 0.006869984149932861, 0.006908927917480469, 0.006903808116912841, 0.00690176010131836, 0.0068884482383728025, 0.006905856132507324, 0.006909952163696289, 0.0068884482383728025, 0.0068884482383728025, 0.006900735855102539, 0.006922239780426025, 0.006904831886291504, 0.00692633581161499, 0.0068853759765625, 0.006887423992156983, 0.0068392958641052244, 0.0068915200233459475, 0.00690176010131836, 0.006783999919891357, 0.006814720153808594, 0.006865920066833496, 0.006905856132507324, 0.0068689918518066405, 0.006905856132507324, 0.006889472007751465, 0.006874112129211426, 0.006872064113616944, 0.006903808116912841, 0.006865920066833496, 0.006881279945373535, 0.015689727783203124, 0.006905856132507324, 0.007403552055358887, 0.007221216201782226, 0.007077888011932373, 0.007154687881469727, 0.006957056045532227, 0.006904831886291504, 0.0068884482383728025, 0.0068884482383728025, 0.0069253120422363285, 0.006892543792724609, 0.0068853759765625, 0.006906879901885986, 0.00687820816040039, 0.006908927917480469, 0.006922239780426025, 0.0069212160110473635, 0.006909952163696289, 0.006855679988861084, 0.006915071964263916, 0.006916096210479736, 0.0068618240356445315, 0.006884352207183838, 0.006870016098022461, 0.006910975933074951, 0.0068700799942016605, 0.006903744220733642, 0.006949888229370117, 0.0069027838706970214, 0.006855679988861084, 0.006892543792724609, 0.006914048194885254, 0.00689356803894043, 0.006894591808319092, 0.006895616054534912, 0.006948863983154297, 0.006821887969970703, 0.006810624122619629, 0.006841343879699707, 0.006800384044647217, 0.0067870721817016606, 0.006845439910888672, 0.006837279796600342, 0.006791135787963867, 0.006823935985565186, 0.0073062400817871095, 0.007062528133392334, 0.0069212160110473635, 0.006933504104614257, 0.0068915200233459475, 0.0068853759765625, 0.006873087882995605, 0.006939648151397705, 0.006936575889587402, 0.006906879901885986, 0.006881279945373535, 0.0068884482383728025, 0.006909952163696289, 0.006899712085723877, 0.006899712085723877, 0.006920224189758301, 0.006910975933074951, 0.015768608093261718, 0.006906816005706787, 0.0069027838706970214, 0.006967296123504638, 0.006880256175994873, 0.00695091199874878, 0.006930431842803955, 0.006908927917480469, 0.006913023948669434, 0.006904831886291504, 0.006919167995452881, 0.00693452787399292, 0.006883327960968018, 0.006949888229370117, 0.006956031799316406, 0.006813695907592773, 0.006874112129211426, 0.006837247848510742, 0.006809599876403808, 0.006823935985565186, 0.006856704235076904, 0.006860799789428711, 0.006821887969970703, 0.006830080032348633, 0.006816768169403077, 0.0068618240356445315, 0.00682700777053833, 0.00682700777053833, 0.006822944164276123, 0.006815775871276856, 0.00683513593673706, 0.006865920066833496, 0.006915071964263916, 0.0069027838706970214, 0.006912000179290771, 0.006946815967559815, 0.00688640022277832, 0.006898719787597657, 0.0069263682365417484, 0.006970304012298584, 0.006874112129211426, 0.006938623905181885, 0.006905856132507324, 0.006889472007751465, 0.006937600135803222, 0.0069283838272094726, 0.006892543792724609, 0.0068915200233459475, 0.00693555212020874, 0.006837247848510742, 0.006862847805023193, 0.006946815967559815, 0.006964223861694336, 0.007005184173583984, 0.006958079814910889, 0.0071188478469848635, 0.007121920108795166, 0.006938687801361084, 0.006880191802978516, 0.006880256175994873, 0.00693555212020874, 0.006931456089019775, 0.0068884482383728025, 0.015653887748718263, 0.006910975933074951, 0.006928415775299073, 0.006914015769958496, 0.006896639823913574, 0.006937600135803222, 0.00689356803894043, 0.006918144226074219, 0.006883327960968018, 0.006907904148101806, 0.006922239780426025, 0.0069110078811645504, 0.00693449592590332, 0.006951935768127441, 0.006864895820617676, 0.0069027838706970214, 0.00694271993637085, 0.006879263877868652, 0.006906847953796386, 0.006906879901885986, 0.006917119979858399, 0.0069550080299377445, 0.006895616054534912, 0.006917119979858399, 0.006917119979858399, 0.006940671920776367, 0.0068915200233459475, 0.006895679950714112, 0.006911935806274414, 0.006904831886291504, 0.006916096210479736, 0.006947840213775635, 0.006931456089019775, 0.006874112129211426, 0.00694377613067627, 0.006957056045532227, 0.00690070390701294, 0.006914048194885254, 0.006965248107910156, 0.0069283838272094726, 0.006905856132507324, 0.006897727966308594, 0.006917056083679199, 0.00693555212020874, 0.006903808116912841, 0.006918144226074219, 0.006937600135803222, 0.00693452787399292, 0.00688640022277832, 0.0069253120422363285, 0.006957056045532227, 0.006894591808319092, 0.006933504104614257, 0.006944767951965332, 0.006938623905181885, 0.006933504104614257, 0.006919167995452881, 0.00693555212020874, 0.006927360057830811, 0.006905856132507324, 0.006860799789428711, 0.006852608203887939, 0.006849535942077637, 0.015653887748718263, 0.006933504104614257, 0.00693452787399292, 0.006966271877288818, 0.006870016098022461, 0.006952960014343262, 0.006974463939666748, 0.006879231929779053, 0.006914048194885254, 0.007013440132141113, 0.007114687919616699, 0.0071833600997924804, 0.00723967981338501, 0.007176191806793213, 0.007108607769012451, 0.0071823358535766605, 0.007208960056304932, 0.0070830078125, 0.007197696208953858, 0.007138304233551026, 0.006966271877288818, 0.007107583999633789, 0.007177279949188232, 0.007164864063262939, 0.007106560230255127, 0.007218175888061523, 0.007169023990631103, 0.0071157760620117185, 0.007195648193359375, 0.007271423816680909, 0.0071905279159545895, 0.007142399787902832, 0.007202816009521484, 0.007177216053009033, 0.007013376235961914, 0.007063551902770996, 0.0070594558715820314, 0.006924287796020508, 0.007104512214660645, 0.0070830078125, 0.006959104061126709, 0.007122975826263428, 0.007072735786437989, 0.007016448020935059, 0.007002111911773682, 0.007139328002929687, 0.0071526398658752445, 0.007106560230255127, 0.007176191806793213, 0.007162879943847656, 0.007126016139984131, 0.007175168037414551, 0.007164927959442138, 0.007137279987335205, 0.0071198720932006835, 0.007157760143280029, 0.007158815860748291, 0.0070665922164917, 0.007154687881469727, 0.007217152118682861, 0.007060480117797851, 0.00709119987487793, 0.007069695949554444, 0.016109567642211914, 0.007122943878173828, 0.007139328002929687, 0.0071485438346862796, 0.00692633581161499, 0.006904831886291504, 0.006908927917480469, 0.00693555212020874, 0.0069027838706970214, 0.00687718391418457, 0.0069816322326660156, 0.006924287796020508, 0.006907904148101806, 0.006894591808319092, 0.006905856132507324, 0.006933504104614257, 0.006914048194885254, 0.006876160144805908, 0.006894591808319092, 0.00688640022277832, 0.006990848064422607, 0.0069918718338012695, 0.006900735855102539, 0.006895616054534912, 0.006866943836212158, 0.006913023948669434, 0.006889472007751465, 0.006881279945373535, 0.006919167995452881, 0.0069745278358459475, 0.006863808155059815, 0.00688640022277832, 0.006916096210479736, 0.006899712085723877, 0.006957087993621826, 0.0068904638290405275, 0.006952960014343262, 0.006961152076721191, 0.00687718391418457, 0.006906879901885986, 0.006896639823913574, 0.006859776020050049, 0.006874112129211426, 0.006914048194885254, 0.006919167995452881, 0.006923264026641846, 0.006890495777130127, 0.006916096210479736, 0.0068853759765625, 0.006887423992156983, 0.006916128158569336, 0.006922207832336426, 0.00690176010131836, 0.00688640022277832, 0.006903808116912841, 0.006910975933074951, 0.006867968082427979, 0.006896639823913574, 0.006927360057830811, 0.006906879901885986, 0.00689462423324585, 0.006898655891418457, 0.006930431842803955, 0.015681535720825195, 0.006894591808319092, 0.00687820816040039, 0.006929408073425293, 0.006918144226074219, 0.006887423992156983, 0.00693555212020874, 0.006923264026641846, 0.006923264026641846, 0.006872064113616944, 0.006964223861694336, 0.006919167995452881, 0.006899712085723877, 0.006964223861694336, 0.006941696166992187, 0.0070522880554199216, 0.00692633581161499, 0.006903808116912841, 0.006884352207183838, 0.006908927917480469, 0.006971392154693603, 0.006914048194885254, 0.0069959678649902345, 0.006884352207183838, 0.006892543792724609, 0.006929408073425293, 0.006909952163696289, 0.006872064113616944, 0.006933504104614257, 0.006906879901885986, 0.006894591808319092, 0.00688640022277832, 0.006876160144805908, 0.006912000179290771, 0.006887423992156983, 0.006876160144805908, 0.006873087882995605, 0.006831103801727295, 0.00678604793548584, 0.0068351998329162595, 0.006820864200592041, 0.006846464157104492, 0.006809599876403808, 0.006830080032348633, 0.0068321280479431154, 0.006818816184997558, 0.006816768169403077, 0.0068351998329162595, 0.006780960083007812, 0.006862815856933594, 0.006908927917480469, 0.006791168212890625, 0.006788095951080322, 0.006913023948669434, 0.006813695907592773, 0.006780960083007812, 0.0067983040809631344, 0.00687718391418457, 0.0068280320167541505, 0.006800384044647217, 0.006953983783721924, 0.006894591808319092, 0.006880256175994873, 0.015646719932556154, 0.006913023948669434, 0.006916096210479736, 0.006903808116912841, 0.006907904148101806, 0.006892543792724609, 0.006876160144805908, 0.006894591808319092, 0.006906879901885986, 0.0068884482383728025, 0.0068884482383728025, 0.0068853759765625, 0.006920191764831543, 0.0069304962158203125, 0.00685971212387085, 0.006913023948669434, 0.006924287796020508, 0.006914048194885254, 0.006912000179290771, 0.0069550080299377445, 0.006978559970855713, 0.006874176025390625, 0.006881216049194336, 0.006914048194885254, 0.00687820816040039, 0.006899712085723877, 0.00694374418258667, 0.0069283838272094726, 0.00687820816040039, 0.006864895820617676, 0.006900735855102539, 0.006905856132507324, 0.006879231929779053, 0.00688640022277832, 0.006858751773834228, 0.0069283838272094726, 0.006958079814910889, 0.007062528133392334, 0.007972864151000977, 0.007472127914428711, 0.00719974422454834, 0.007162879943847656, 0.007126016139984131, 0.007173120021820068, 0.007172160148620606, 0.007097280025482178, 0.0071526398658752445, 0.007130112171173096, 0.007117824077606201, 0.0071823358535766605, 0.006915071964263916, 0.0068689918518066405, 0.006914048194885254, 0.0069027838706970214, 0.006875135898590088, 0.00689356803894043, 0.00690176010131836, 0.006897664070129395, 0.006905856132507324, 0.0068853759765625, 0.006897664070129395, 0.0068915200233459475, 0.0069560642242431645, 0.015705056190490722, 0.006912000179290771, 0.006875135898590088, 0.006905856132507324, 0.006883327960968018, 0.0069324798583984375, 0.006924287796020508, 0.006884352207183838, 0.00692633581161499, 0.006993919849395752, 0.007068672180175781, 0.007006207942962647, 0.006982656002044678, 0.006843391895294189, 0.006830080032348633, 0.006894591808319092, 0.006913087844848633, 0.0068791680335998535, 0.0068853759765625, 0.006915071964263916, 0.006952960014343262, 0.006903808116912841, 0.006905856132507324, 0.006905856132507324, 0.006912000179290771, 0.006892608165740967, 0.006883296012878418, 0.006904799938201905, 0.006906879901885986, 0.006857791900634766, 0.006956992149353028, 0.006998015880584717, 0.006892543792724609, 0.006880256175994873, 0.0068915200233459475, 0.006896639823913574, 0.006880256175994873, 0.006887423992156983, 0.006913023948669434, 0.006913023948669434, 0.006845439910888672, 0.006904831886291504, 0.006889472007751465, 0.006900735855102539, 0.006872064113616944, 0.0068884482383728025, 0.006879231929779053, 0.0069027838706970214, 0.006859807968139648, 0.006940639972686767, 0.006892543792724609, 0.006884352207183838, 0.006897664070129395, 0.006914048194885254, 0.006908927917480469, 0.006882304191589355, 0.006924287796020508, 0.00689356803894043, 0.006871039867401123, 0.006892543792724609, 0.0068986878395080565, 0.00693555212020874, 0.0068986878395080565, 0.01566412830352783, 0.006918144226074219, 0.006895616054534912, 0.006900735855102539, 0.006927360057830811, 0.006920191764831543, 0.006902847766876221, 0.006888383865356445, 0.006924287796020508, 0.006883327960968018, 0.006905856132507324, 0.006889472007751465, 0.006970367908477783, 0.006940671920776367, 0.006825984001159668, 0.006896671772003174, 0.006927328109741211, 0.0069253120422363285, 0.006863872051239014, 0.00687718391418457, 0.006916096210479736, 0.006882304191589355, 0.006922239780426025, 0.006907904148101806, 0.006889472007751465, 0.006904831886291504, 0.0068884482383728025, 0.006917119979858399, 0.006855679988861084, 0.006825984001159668, 0.0068280320167541505, 0.006852608203887939, 0.006814720153808594, 0.006865920066833496, 0.007031807899475098, 0.0069324798583984375, 0.006907904148101806, 0.006944767951965332, 0.006892543792724609, 0.006937600135803222, 0.006886464118957519, 0.006919104099273682, 0.006922304153442383, 0.006875072002410889, 0.0069027838706970214, 0.006915071964263916, 0.0069283838272094726, 0.006894656181335449, 0.006899648189544678, 0.006941696166992187, 0.006895679950714112, 0.006926271915435791, 0.006994944095611572, 0.006916096210479736, 0.006887423992156983, 0.006887423992156983, 0.006896639823913574, 0.006885407924652099, 0.006882271766662598, 0.006906879901885986, 0.006889472007751465, 0.006882304191589355, 0.006875135898590088, 0.01575014400482178, 0.006917119979858399, 0.006881279945373535, 0.006897664070129395, 0.006905856132507324, 0.0069027838706970214, 0.006956031799316406, 0.006873087882995605, 0.00693452787399292, 0.006905856132507324, 0.006876160144805908, 0.006951935768127441, 0.006915071964263916, 0.006900735855102539, 0.006890528202056885, 0.006919136047363281, 0.0069304962158203125, 0.007155648231506348, 0.007073791980743408, 0.006931456089019775, 0.00693452787399292, 0.006882304191589355, 0.0068915200233459475, 0.006904831886291504, 0.007117856025695801, 0.006944736003875732, 0.006883327960968018, 0.0069816322326660156, 0.00692633581161499, 0.006873087882995605, 0.006879231929779053, 0.006915071964263916, 0.0069027838706970214, 0.006890495777130127, 0.006908927917480469, 0.006907904148101806, 0.006884352207183838, 0.00688640022277832, 0.006900735855102539, 0.006913023948669434, 0.006890495777130127, 0.0068915200233459475, 0.006895616054534912, 0.006830080032348633, 0.006799359798431396, 0.006829055786132812, 0.006797311782836914, 0.006854656219482422, 0.007012351989746094, 0.00698367977142334, 0.006917119979858399, 0.006876192092895508, 0.006906847953796386, 0.006892543792724609, 0.006897664070129395, 0.0068915200233459475, 0.0068915200233459475, 0.006899712085723877, 0.006833151817321777, 0.006899712085723877, 0.006913023948669434, 0.006860799789428711, 0.0068884482383728025, 0.0156364803314209, 0.0069253120422363285, 0.006875135898590088, 0.006897664070129395, 0.006920191764831543, 0.006918208122253418, 0.006869952201843262, 0.006882304191589355, 0.006907904148101806, 0.006898719787597657, 0.0068853440284729, 0.006919167995452881, 0.006919167995452881, 0.006913023948669434, 0.006904831886291504, 0.006912000179290771, 0.0069621758460998535, 0.006881279945373535, 0.00687820816040039, 0.006884352207183838, 0.006933504104614257, 0.006909952163696289, 0.006989823818206787, 0.006903808116912841, 0.006881279945373535, 0.006883327960968018, 0.006907904148101806, 0.0069027838706970214, 0.006892543792724609, 0.00690176010131836, 0.006908927917480469, 0.006946815967559815, 0.006872064113616944, 0.006907904148101806, 0.006912000179290771, 0.006913023948669434, 0.006856704235076904, 0.006820864200592041, 0.006788095951080322, 0.006807551860809326, 0.006872064113616944, 0.006845439910888672, 0.0067870721817016606, 0.006803455829620361, 0.00682700777053833, 0.006825984001159668, 0.006782976150512696, 0.006820864200592041, 0.006806528091430664, 0.006830080032348633, 0.006871039867401123, 0.006975488185882568, 0.006975488185882568, 0.006872064113616944, 0.006919167995452881, 0.006912000179290771, 0.006896639823913574, 0.006941696166992187, 0.006896639823913574, 0.006904831886291504, 0.006890495777130127, 0.006873087882995605, 0.00693452787399292, 0.015656959533691405, 0.006898719787597657, 0.0068576960563659665, 0.006899712085723877, 0.00690176010131836, 0.006974463939666748, 0.006918144226074219, 0.0068915200233459475, 0.006905856132507324, 0.006852608203887939, 0.006823935985565186, 0.006845439910888672, 0.006845439910888672, 0.006819839954376221, 0.0068321280479431154, 0.006809599876403808, 0.006833151817321777, 0.006799359798431396, 0.006853631973266602, 0.006830080032348633, 0.00690176010131836, 0.006956031799316406, 0.006916096210479736, 0.00688640022277832, 0.006920191764831543, 0.006899712085723877, 0.006874112129211426, 0.006889472007751465, 0.006922239780426025, 0.006920191764831543, 0.0069027838706970214, 0.006905856132507324, 0.0069212160110473635, 0.006914048194885254, 0.006900735855102539, 0.006924287796020508, 0.0069028158187866215, 0.006882271766662598, 0.006872064113616944, 0.006904831886291504, 0.006889472007751465, 0.0068772478103637695, 0.006905792236328125, 0.006917119979858399, 0.006894591808319092, 0.006899712085723877, 0.006914048194885254, 0.0068986878395080565, 0.006909952163696289, 0.006876160144805908, 0.0068986878395080565, 0.006899712085723877, 0.006851583957672119, 0.0068321280479431154, 0.006808576107025147, 0.006883327960968018, 0.006794271945953369, 0.006814752101898193, 0.006809535980224609, 0.006796288013458252, 0.006829055786132812, 0.006816768169403077, 0.006796288013458252, 0.015989760398864745, 0.007045119762420654, 0.006917119979858399, 0.006904831886291504, 0.00689356803894043, 0.007004159927368164, 0.0070860800743103025, 0.007123968124389648, 0.00710041618347168, 0.007063551902770996, 0.007134208202362061, 0.007085055828094483, 0.007085055828094483, 0.007068672180175781, 0.007178239822387696, 0.007120895862579346, 0.007120895862579346, 0.0070522880554199216, 0.00709119987487793, 0.0078919677734375, 0.007162879943847656, 0.007117824077606201, 0.0070891518592834475, 0.007077888011932373, 0.0071188478469848635, 0.007159808158874512, 0.007130112171173096, 0.00709222412109375, 0.007082015991210938, 0.007109663963317871, 0.007156735897064209, 0.007131072044372558, 0.007120895862579346, 0.00719974422454834, 0.00714035177230835, 0.00709119987487793, 0.0070522880554199216, 0.007108607769012451, 0.007112703800201416, 0.007097343921661377, 0.006978559970855713, 0.006870016098022461, 0.006909952163696289, 0.006882304191589355, 0.006817791938781738, 0.006879231929779053, 0.006905888080596924, 0.00688431978225708, 0.00687718391418457, 0.006919167995452881, 0.006904831886291504, 0.006804480075836182, 0.006833151817321777, 0.0068321280479431154, 0.006779935836791992, 0.0068802242279052735, 0.006973440170288086, 0.006904831886291504, 0.006871039867401123, 0.006823935985565186, 0.006837247848510742, 0.006778880119323731, 0.00682700777053833]",tokens/s,141.02086808064445,,,main,False,False,,,,,,,,,,,,,,,,,,,,,,, float16-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,databricks/dbrx-base,databricks/dbrx-base,cuda,0,42,,,True,,,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",dbrx,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-350m,,cuda,0,42,,,,,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,d35829e539df8480b726c647eeabf91e41eae047,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,,MB,1188.442112,1455.947776,0.0,809.500672,723.533824,s,13,0.19395561599731448,0.014919662769024188,0.0006263890092848103,0.014573216438293457,0.015982310676574706,0.016239801406860353,0.01632139320373535,"[0.016341791152954102, 0.014573216438293457, 0.014939104080200195, 0.014454079627990722, 0.01617180824279785, 0.014889568328857421, 0.015062560081481933, 0.014445055961608886, 0.015224320411682129, 0.014482432365417481, 0.014442303657531737, 0.014444767951965332, 0.014484607696533203]",tokens/s,17158.56477208724,kWh,1.7314691897115033e-07,9.487634185087318e-08,4.882458761990672e-07,7.562691370210907e-07,tokens/kWh,338503830.80337274,MB,1188.442112,1455.947776,0.0,809.500672,749.21984,s,13,10.502904846191404,0.8079157573993387,0.009505664839521193,0.80391015625,0.823531689453125,0.8280160888671875,0.8301153271484375,"[0.8262667236328125, 0.812591552734375, 0.7997078247070313, 0.8014157104492188, 0.8047113037109375, 0.8048790283203126, 0.8026774291992187, 0.80391015625, 0.83064013671875, 0.799624755859375, 0.8023453369140625, 0.8034739990234375, 0.810660888671875]",tokens/s,77.97842711076149,kWh,9.53687434708987e-06,5.225731663599511e-06,1.7475379605797372e-05,3.223798561648675e-05,tokens/kWh,1954216.3939604627,,s,818,10.666156065940855,0.013039310594059727,0.0018623556299793502,0.01268838405609131,0.013345097541809082,0.013433292531967162,0.02762204216003418,"[0.014240768432617188, 0.014169088363647461, 0.01459404754638672, 0.013119487762451172, 0.012732416152954102, 0.012678144454956054, 0.01267404842376709, 0.012804096221923827, 0.012634112358093261, 0.012650495529174804, 0.012620800018310547, 0.01265664005279541, 0.012689408302307128, 0.012697600364685058, 0.012653568267822265, 0.012718079566955566, 0.012681216239929198, 0.01273036766052246, 0.01273855972290039, 0.01277132797241211, 0.01287065601348877, 0.012635135650634765, 0.012709888458251953, 0.012792832374572754, 0.012743680000305176, 0.012625920295715331, 0.012667903900146485, 0.013364224433898926, 0.013499423980712891, 0.013488096237182617, 0.01343283176422119, 0.013444095611572266, 0.013674495697021484, 0.013442048072814941, 0.013431808471679688, 0.013444095611572266, 0.01338265609741211, 0.013436927795410156, 0.013435903549194337, 0.013356032371520997, 0.013319168090820312, 0.01337446403503418, 0.013344767570495606, 0.013794303894042969, 0.013689855575561523, 0.013816831588745117, 0.013475839614868163, 0.013178879737854005, 0.013119487762451172, 0.01305292797088623, 0.0130764799118042, 0.013071359634399414, 0.012957695960998536, 0.01304371166229248, 0.012981247901916505, 0.012682239532470703, 0.012643327713012695, 0.012678144454956054, 0.01263923168182373, 0.012892160415649414, 0.012887040138244628, 0.012646431922912598, 0.027743200302124023, 0.012618816375732422, 0.012636096000671386, 0.012667936325073242, 0.012675040245056153, 0.01268019199371338, 0.012786687850952149, 0.013175807952880859, 0.013033472061157226, 0.013027327537536621, 0.01287168025970459, 0.012644351959228516, 0.012681216239929198, 0.012724224090576173, 0.012677120208740235, 0.012673055648803711, 0.013089759826660156, 0.013005824089050292, 0.01303756809234619, 0.013096960067749023, 0.01306828784942627, 0.01285529613494873, 0.013398015975952148, 0.013198335647583008, 0.013189120292663574, 0.013041664123535156, 0.012875776290893554, 0.012718079566955566, 0.012957695960998536, 0.012887040138244628, 0.013853759765625, 0.013312959671020507, 0.01313587188720703, 0.012958720207214355, 0.012652544021606446, 0.012717056274414062, 0.012744704246520995, 0.012703743934631348, 0.012854304313659668, 0.012755935668945312, 0.012660736083984376, 0.012723199844360352, 0.012751872062683106, 0.013008895874023438, 0.013055999755859376, 0.013071359634399414, 0.013191167831420898, 0.012981247901916505, 0.0129617919921875, 0.012751872062683106, 0.012683327674865724, 0.01267807960510254, 0.012695551872253418, 0.012652544021606446, 0.012698623657226562, 0.012679167747497559, 0.01267199993133545, 0.012724224090576173, 0.01276211166381836, 0.01316864013671875, 0.013163552284240723, 0.013011936187744141, 0.013031423568725586, 0.02811494445800781, 0.012653568267822265, 0.012625920295715331, 0.012712960243225097, 0.01267199993133545, 0.012660736083984376, 0.012636159896850586, 0.012699647903442383, 0.012702719688415527, 0.012731424331665039, 0.012691424369812012, 0.012652544021606446, 0.012620800018310547, 0.012661760330200195, 0.012686335563659668, 0.012689408302307128, 0.01266483211517334, 0.012604415893554688, 0.01264742374420166, 0.012640255928039551, 0.012633088111877442, 0.013162495613098145, 0.012964863777160645, 0.01267404842376709, 0.012651519775390625, 0.012662783622741699, 0.012654591560363769, 0.012649472236633302, 0.012652544021606446, 0.01264844799041748, 0.012696576118469239, 0.012709888458251953, 0.012782591819763184, 0.01273036766052246, 0.012650495529174804, 0.012634112358093261, 0.012666879653930664, 0.012620800018310547, 0.012634112358093261, 0.012636159896850586, 0.012592127799987793, 0.012606464385986327, 0.012710911750793457, 0.012608511924743653, 0.012685312271118163, 0.012636159896850586, 0.01267404842376709, 0.012951552391052246, 0.013063167572021485, 0.012712960243225097, 0.012618751525878906, 0.012733440399169921, 0.012619775772094726, 0.012630016326904296, 0.012682239532470703, 0.012653568267822265, 0.012666879653930664, 0.012626943588256835, 0.012678144454956054, 0.012660736083984376, 0.012755968093872071, 0.012668928146362305, 0.012577856063842773, 0.027557823181152345, 0.01268019199371338, 0.012681247711181641, 0.012670944213867188, 0.012637184143066407, 0.012636159896850586, 0.012627967834472656, 0.012699647903442383, 0.012670975685119629, 0.012638208389282226, 0.01275699234008789, 0.012728320121765137, 0.012866559982299805, 0.012702719688415527, 0.012619775772094726, 0.01268838405609131, 0.012700672149658204, 0.012640255928039551, 0.012675071716308594, 0.01264742374420166, 0.012584959983825684, 0.012597248077392579, 0.012613632202148438, 0.012653568267822265, 0.012667936325073242, 0.012670944213867188, 0.012726271629333496, 0.01266585636138916, 0.012860416412353515, 0.012805120468139648, 0.01266483211517334, 0.012630016326904296, 0.01263923168182373, 0.012644351959228516, 0.012660736083984376, 0.012700703620910645, 0.012637151718139648, 0.012637184143066407, 0.012658687591552734, 0.012681216239929198, 0.012719103813171387, 0.01268838405609131, 0.01268838405609131, 0.012645376205444337, 0.012675071716308594, 0.012662783622741699, 0.0126310396194458, 0.012697600364685058, 0.012628000259399414, 0.012675104141235351, 0.012633024215698242, 0.012652544021606446, 0.01267199993133545, 0.012668928146362305, 0.012744704246520995, 0.012694527626037597, 0.012626943588256835, 0.01277337646484375, 0.01297920036315918, 0.013008895874023438, 0.013016063690185547, 0.013484031677246093, 0.013275135993957519, 0.029944831848144532, 0.013188096046447753, 0.013150208473205567, 0.013015040397644043, 0.012989439964294434, 0.012986368179321289, 0.013121536254882812, 0.013156352043151855, 0.013131775856018067, 0.013018112182617187, 0.013028351783752442, 0.01276313591003418, 0.012759039878845215, 0.012679200172424316, 0.01259830379486084, 0.012623807907104493, 0.012736512184143066, 0.01276313591003418, 0.013088768005371093, 0.012916735649108887, 0.012654591560363769, 0.012597248077392579, 0.012613632202148438, 0.012823552131652831, 0.013117440223693847, 0.012744704246520995, 0.012619775772094726, 0.012611583709716797, 0.012609536170959473, 0.012619775772094726, 0.012646400451660156, 0.012701696395874023, 0.012603391647338867, 0.012619775772094726, 0.012678144454956054, 0.012645376205444337, 0.01266380786895752, 0.012632063865661621, 0.01264742374420166, 0.012626943588256835, 0.012624896049499512, 0.012678144454956054, 0.012712960243225097, 0.012628992080688477, 0.012992544174194336, 0.012634079933166504, 0.012583935737609863, 0.012695551872253418, 0.012606528282165528, 0.012789759635925293, 0.012874688148498535, 0.01263315200805664, 0.012698559761047364, 0.01266380786895752, 0.01265664005279541, 0.012699647903442383, 0.012698623657226562, 0.012641280174255372, 0.012625920295715331, 0.012706815719604492, 0.012696576118469239, 0.01258188819885254, 0.012677120208740235, 0.02831667137145996, 0.012606464385986327, 0.01265664005279541, 0.012679167747497559, 0.012706815719604492, 0.01266585636138916, 0.012889087677001953, 0.012800000190734863, 0.012645376205444337, 0.012679167747497559, 0.012709888458251953, 0.012957695960998536, 0.013225983619689942, 0.012997695922851563, 0.013011903762817382, 0.012888064384460449, 0.012697600364685058, 0.012643327713012695, 0.012949503898620606, 0.013002816200256348, 0.012976063728332519, 0.012841983795166016, 0.012986368179321289, 0.012749823570251465, 0.013048831939697265, 0.012907520294189453, 0.01277337646484375, 0.012610560417175292, 0.012690431594848632, 0.01277132797241211, 0.01277952003479004, 0.012600383758544921, 0.012631999969482421, 0.012697600364685058, 0.012607487678527832, 0.012606464385986327, 0.012785663604736328, 0.012637184143066407, 0.012717056274414062, 0.012916735649108887, 0.012967935562133789, 0.012820480346679687, 0.01265664005279541, 0.012654623985290527, 0.012633055686950683, 0.012536831855773926, 0.012596223831176758, 0.012686335563659668, 0.012594176292419433, 0.012640255928039551, 0.01259008026123047, 0.012613632202148438, 0.012716032028198243, 0.012677120208740235, 0.013054976463317871, 0.01303756809234619, 0.013161503791809082, 0.01261257553100586, 0.012543999671936035, 0.012634112358093261, 0.01265664005279541, 0.01267302417755127, 0.012973055839538575, 0.02857369613647461, 0.012920831680297852, 0.01265664005279541, 0.01267199993133545, 0.012651519775390625, 0.012661760330200195, 0.012716032028198243, 0.012653568267822265, 0.012934176445007325, 0.01292899227142334, 0.012678144454956054, 0.012637184143066407, 0.01266585636138916, 0.012812288284301757, 0.012645376205444337, 0.012654591560363769, 0.012593152046203614, 0.01264844799041748, 0.01265766429901123, 0.012711935997009278, 0.012685312271118163, 0.012659711837768555, 0.01264134407043457, 0.01267091178894043, 0.012679167747497559, 0.01275494384765625, 0.012920831680297852, 0.012644351959228516, 0.01265766429901123, 0.012677120208740235, 0.012643327713012695, 0.012636159896850586, 0.012632063865661621, 0.012723199844360352, 0.012698623657226562, 0.012840959548950195, 0.013023232460021973, 0.012902400016784669, 0.012978176116943359, 0.012928000450134277, 0.01265766429901123, 0.012686335563659668, 0.013231103897094726, 0.013021183967590331, 0.01307033634185791, 0.01287065601348877, 0.01267302417755127, 0.012659711837768555, 0.012668928146362305, 0.012627967834472656, 0.012677120208740235, 0.012650495529174804, 0.01285324764251709, 0.0126310396194458, 0.012684288024902344, 0.012651583671569825, 0.012615615844726563, 0.012626943588256835, 0.012667903900146485, 0.012568575859069824, 0.012666879653930664, 0.012611583709716797, 0.012628992080688477, 0.02757529640197754, 0.012661760330200195, 0.0126310396194458, 0.012653568267822265, 0.012684288024902344, 0.01265664005279541, 0.012895232200622558, 0.012717056274414062, 0.012638208389282226, 0.012760064125061036, 0.012695551872253418, 0.012587008476257324, 0.012679167747497559, 0.012686335563659668, 0.013087743759155274, 0.012921855926513673, 0.012700672149658204, 0.012634112358093261, 0.01266585636138916, 0.012699647903442383, 0.012696576118469239, 0.012741632461547851, 0.012981247901916505, 0.012956671714782715, 0.012623871803283691, 0.012635135650634765, 0.012576767921447754, 0.012645376205444337, 0.012579839706420898, 0.01263923168182373, 0.012615679740905762, 0.012609536170959473, 0.012608511924743653, 0.012613632202148438, 0.012637184143066407, 0.012741632461547851, 0.012616703987121582, 0.012633088111877442, 0.012609536170959473, 0.012736512184143066, 0.01306828784942627, 0.01283993625640869, 0.012678144454956054, 0.012614656448364257, 0.012616703987121582, 0.01255731201171875, 0.012610560417175292, 0.012644351959228516, 0.01273036766052246, 0.012637184143066407, 0.01265664005279541, 0.012678144454956054, 0.012640255928039551, 0.012692480087280274, 0.012614656448364257, 0.01265664005279541, 0.012595199584960937, 0.013418496131896973, 0.013426688194274903, 0.013331487655639648, 0.013219807624816894, 0.0133570556640625, 0.013368320465087891, 0.029088768005371093, 0.013415424346923829, 0.013373439788818359, 0.013381631851196289, 0.013415424346923829, 0.01336627197265625, 0.013334527969360351, 0.013522944450378417, 0.013331456184387207, 0.01347379207611084, 0.01335910415649414, 0.013396991729736327, 0.013362175941467285, 0.013454336166381836, 0.013419520378112794, 0.013431808471679688, 0.013356032371520997, 0.013361151695251466, 0.013379584312438965, 0.013342720031738281, 0.013344799995422364, 0.01334169578552246, 0.013369312286376953, 0.013345791816711425, 0.013394944190979004, 0.01335807991027832, 0.01334988784790039, 0.01346457576751709, 0.013472767829895019, 0.013459456443786622, 0.013431808471679688, 0.013305855751037597, 0.013364224433898926, 0.013399040222167969, 0.013304832458496094, 0.013323264122009277, 0.013394944190979004, 0.013298687934875488, 0.01339187240600586, 0.013467647552490235, 0.013634559631347656, 0.012940287590026855, 0.013379584312438965, 0.013088768005371093, 0.013044735908508302, 0.012978176116943359, 0.013116415977478027, 0.01293824005126953, 0.012650495529174804, 0.012600319862365723, 0.012641280174255372, 0.01263923168182373, 0.012625920295715331, 0.012577823638916016, 0.012571616172790527, 0.012713983535766601, 0.012654591560363769, 0.012658687591552734, 0.012650495529174804, 0.012661760330200195, 0.01267302417755127, 0.012633088111877442, 0.012692480087280274, 0.02755379295349121, 0.012614656448364257, 0.012627967834472656, 0.01264742374420166, 0.012670975685119629, 0.012587039947509766, 0.012592096328735352, 0.012668928146362305, 0.012605440139770508, 0.012703743934631348, 0.012628992080688477, 0.012661760330200195, 0.01264742374420166, 0.012667903900146485, 0.012621824264526366, 0.012642304420471191, 0.012634112358093261, 0.012627967834472656, 0.012626943588256835, 0.012633119583129883, 0.012674015998840332, 0.012614656448364257, 0.012769344329833985, 0.012698559761047364, 0.012705792427062988, 0.012649472236633302, 0.012717056274414062, 0.01264844799041748, 0.012660736083984376, 0.012628992080688477, 0.012652607917785644, 0.012702655792236329, 0.012614656448364257, 0.012883968353271484, 0.012732416152954102, 0.01266585636138916, 0.01266585636138916, 0.01264742374420166, 0.012682239532470703, 0.01264844799041748, 0.012615679740905762, 0.012621824264526366, 0.012618751525878906, 0.012709888458251953, 0.012693504333496093, 0.012662783622741699, 0.012726271629333496, 0.012676159858703613, 0.012638143539428711, 0.012684288024902344, 0.012635135650634765, 0.012579839706420898, 0.012618751525878906, 0.012637184143066407, 0.012933119773864746, 0.012983296394348144, 0.01298739242553711, 0.012948479652404785, 0.012741632461547851, 0.01287782382965088, 0.0127293758392334, 0.01274672031402588, 0.01276416015625, 0.02775551986694336, 0.013016063690185547, 0.013085696220397949, 0.013007871627807617, 0.012955648422241211, 0.012884991645812988, 0.013025279998779296, 0.013073408126831054, 0.012821503639221191, 0.012677120208740235, 0.012650495529174804, 0.01266380786895752, 0.012933152198791504, 0.013019136428833008, 0.012878815650939941, 0.0127457275390625, 0.012684288024902344, 0.012620800018310547, 0.012676095962524414, 0.012650495529174804, 0.012615679740905762, 0.012611583709716797, 0.012623871803283691, 0.012766207695007324, 0.012654591560363769, 0.012570624351501464, 0.012652544021606446, 0.01262284755706787, 0.012633088111877442, 0.012670975685119629, 0.012620800018310547, 0.012747776031494141, 0.012667903900146485, 0.0126310396194458, 0.01264742374420166, 0.012791808128356934, 0.01293619155883789, 0.01277132797241211, 0.012660799980163574, 0.012624832153320312, 0.012646400451660156, 0.012686335563659668, 0.012658687591552734, 0.012612607955932617, 0.012638208389282226, 0.012646400451660156, 0.012668928146362305, 0.012695551872253418, 0.012690431594848632, 0.01266585636138916, 0.012695551872253418, 0.012638208389282226, 0.012676159858703613, 0.01266374397277832, 0.012600319862365723, 0.012750847816467285, 0.012659711837768555, 0.01269148826599121, 0.012680159568786622, 0.012645376205444337, 0.01285529613494873, 0.012615679740905762, 0.012691455841064453, 0.027667455673217774, 0.012662783622741699, 0.012801024436950683, 0.012718079566955566, 0.012702719688415527, 0.012652544021606446, 0.012624896049499512, 0.012679167747497559, 0.01263923168182373, 0.01264844799041748, 0.012722175598144531, 0.012602368354797363, 0.012630016326904296, 0.01266585636138916, 0.012653568267822265, 0.012707839965820313, 0.012662783622741699, 0.012679167747497559, 0.012661760330200195, 0.012652544021606446, 0.012644351959228516, 0.01265766429901123, 0.012623871803283691, 0.012640255928039551, 0.012630016326904296, 0.01267404842376709, 0.012741632461547851, 0.012731391906738282, 0.013028351783752442, 0.012931072235107421, 0.013342720031738281, 0.01307852840423584, 0.012836864471435547, 0.012995583534240723, 0.012844032287597656, 0.012902400016784669, 0.012878848075866698, 0.01296895980834961, 0.012806143760681152, 0.012687359809875488, 0.012687359809875488, 0.012642304420471191, 0.012614656448364257, 0.012822527885437012, 0.012659711837768555, 0.012650495529174804, 0.01287168025970459, 0.012810239791870117, 0.012850175857543946, 0.012685312271118163, 0.012826623916625977, 0.012894207954406739, 0.01266380786895752, 0.012678144454956054, 0.012677120208740235, 0.012659711837768555, 0.01275494384765625, 0.012603391647338867, 0.012624896049499512, 0.012687359809875488, 0.012644351959228516, 0.01284505558013916, 0.013010944366455078, 0.027631616592407225, 0.012638208389282226, 0.01266585636138916, 0.012693504333496093, 0.01266483211517334, 0.012918784141540527, 0.013204480171203613, 0.013760512351989745, 0.013348863601684571, 0.013119487762451172, 0.01306220817565918, 0.01278048038482666, 0.012691455841064453, 0.012751872062683106, 0.012723199844360352, 0.012685312271118163, 0.012743680000305176, 0.012720128059387208, 0.01276211166381836, 0.012649472236633302, 0.012650495529174804, 0.012619839668273927, 0.012644288063049316, 0.012620832443237305, 0.012706784248352051, 0.012622912406921387, 0.012762047767639161, 0.012717056274414062, 0.01264844799041748, 0.012726271629333496, 0.012751872062683106, 0.012689408302307128, 0.012744704246520995, 0.012652544021606446, 0.012651519775390625, 0.012691455841064453, 0.012668928146362305, 0.01267199993133545, 0.012662783622741699, 0.012638208389282226, 0.012719103813171387, 0.01287782382965088, 0.013023232460021973, 0.012935168266296386, 0.013083647727966309, 0.013044735908508302, 0.012875776290893554, 0.012675104141235351, 0.012684255599975587, 0.012780544281005859, 0.01268838405609131, 0.01266380786895752, 0.012666879653930664, 0.01264844799041748, 0.01266380786895752, 0.013351936340332032, 0.01343283176422119, 0.013469696044921875, 0.0134717435836792, 0.013395968437194825, 0.013404159545898438, 0.013455360412597657, 0.013386783599853515]",tokens/s,76.69117111571578,,,main,False,False,,,,,,,,,,,,,,,,,,,,,,, float16-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,-,-,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 106, in _inner_fn validate_repo_id(arg_value) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 160, in validate_repo_id raise HFValidationError( huggingface_hub.errors.HFValidationError: Repo id must use alphanumeric chars or '-', '_', '.', '--' and '..' are forbidden, '-' and '.' cannot start or end the name, max length is 96: '-'. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 466, in cached_file raise EnvironmentError( OSError: Incorrect path_or_model_id: '-'. Please provide either the path to a local folder or the repo_id of a model on the Hub. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,stabilityai/stablelm-2-12b,stabilityai/stablelm-2-12b,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ self.load_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 250, in load_model_with_no_weights self.pretrained_model = self.automodel_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3710, in from_pretrained model = cls(config, *model_args, **model_kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/stablelm/modeling_stablelm.py"", line 1188, in __init__ self.model = StableLmModel(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/stablelm/modeling_stablelm.py"", line 966, in __init__ [StableLmDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/stablelm/modeling_stablelm.py"", line 966, in [StableLmDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/stablelm/modeling_stablelm.py"", line 748, in __init__ self.mlp = StableLmMLP(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/stablelm/modeling_stablelm.py"", line 206, in __init__ self.down_proj = nn.Linear(self.intermediate_size, self.hidden_size, bias=False) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/linear.py"", line 98, in __init__ self.weight = Parameter(torch.empty((out_features, in_features), **factory_kwargs)) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_device.py"", line 78, in __torch_function__ return func(*args, **kwargs) torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 136.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,v,v,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/v/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6693255b-4bf4892d48fb990d234324d2;735d79cd-8441-4d90-826b-64383e1bc912) Repository Not Found for url: https://huggingface.co/v/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: v is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,openai-community/gpt2,openai-community/gpt2,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.1,,,,MB,1255.809024,1032.323072,0.0,385.875968,318.907904,s,20,0.18153276729583742,0.009076638364791872,0.0003141694243506662,0.00895747184753418,0.009315273666381836,0.009725623798370362,0.010021233310699462,"[0.010095135688781738, 0.009706175804138184, 0.009121312141418456, 0.008833888053894044, 0.008868639945983886, 0.00916540813446045, 0.008800352096557617, 0.009146112442016602, 0.008865407943725586, 0.008980511665344238, 0.009082176208496094, 0.008888511657714844, 0.008934432029724122, 0.009149375915527344, 0.008866016387939452, 0.008854751586914062, 0.008865119934082032, 0.009175711631774903, 0.00886188793182373, 0.009271840095520019]",tokens/s,28204.274502443517,kWh,1.0640743616464977e-07,5.830441011287778e-08,2.6364986056935747e-07,4.2836170684688503e-07,tokens/kWh,597625781.9224384,MB,1255.809024,1032.323072,0.0,385.875968,328.80384,s,20,10.166184967041014,0.5083092483520507,0.010734162763515121,0.5057615814208984,0.5166134033203125,0.5200022735595703,0.542347915649414,"[0.547934326171875, 0.5185321655273437, 0.5053041076660156, 0.49898333740234374, 0.5047973327636719, 0.5029664306640625, 0.5064540710449219, 0.5104552612304688, 0.5024306335449219, 0.5062190551757813, 0.5002470397949219, 0.4974671630859375, 0.4992158203125, 0.5164002075195312, 0.5033699645996094, 0.5123206176757813, 0.5014171142578125, 0.5091772155761719, 0.50909814453125, 0.5133949584960937]",tokens/s,123.94029855692638,kWh,5.848512832556401e-06,3.2047421455629507e-06,1.0156483475532398e-05,1.9209738453651745e-05,tokens/kWh,3279586.557203947,,s,1259,10.32876235866545,0.008203941508074238,0.001181372539690216,0.007948287963867188,0.008281497573852539,0.00864504632949829,0.017236337471008303,"[0.009099264144897461, 0.009018367767333984, 0.00902451229095459, 0.008921088218688965, 0.00872652816772461, 0.008663040161132812, 0.008729599952697753, 0.008604672431945801, 0.00859545612335205, 0.0086364164352417, 0.008860671997070312, 0.008765439987182617, 0.008774656295776367, 0.008648703575134278, 0.008652799606323243, 0.008827903747558593, 0.00873574447631836, 0.008738880157470702, 0.008805312156677247, 0.008754176139831543, 0.008760319709777833, 0.008846336364746094, 0.008641535758972169, 0.008644607543945313, 0.008755200386047364, 0.008639488220214844, 0.008819711685180665, 0.00869478416442871, 0.008803359985351563, 0.008815584182739257, 0.008763392448425293, 0.008699904441833496, 0.008644607543945313, 0.008833024024963379, 0.008669183731079102, 0.00871833610534668, 0.008622079849243165, 0.008683520317077637, 0.00868556785583496, 0.008761343955993652, 0.00862003231048584, 0.00830668830871582, 0.00829030418395996, 0.00829036808013916, 0.008532928466796876, 0.00871628761291504, 0.008802304267883301, 0.008753151893615722, 0.008752127647399903, 0.008654848098754882, 0.008884223937988281, 0.008652799606323243, 0.008581119537353516, 0.008530943870544434, 0.00841318416595459, 0.008622079849243165, 0.008498175621032715, 0.008257535934448243, 0.008384511947631837, 0.008523776054382324, 0.00859340763092041, 0.00864463996887207, 0.019183584213256836, 0.008521727561950684, 0.008385536193847656, 0.008358912467956543, 0.008514623641967773, 0.008490943908691406, 0.008364031791687012, 0.00818892765045166, 0.008162303924560547, 0.008186944007873535, 0.00818169593811035, 0.008201215744018555, 0.008531968116760253, 0.00827187156677246, 0.008317952156066894, 0.008223744392395019, 0.008216575622558593, 0.00820633602142334, 0.008196096420288086, 0.008170495986938477, 0.008268799781799317, 0.00818073558807373, 0.008162303924560547, 0.008212479591369629, 0.008173567771911621, 0.008135680198669434, 0.008182784080505372, 0.008175616264343261, 0.008145919799804687, 0.008164352416992187, 0.008160256385803222, 0.00818995189666748, 0.008218688011169433, 0.008130496025085449, 0.008138751983642578, 0.008152064323425292, 0.008161279678344726, 0.008153087615966797, 0.008167424201965333, 0.008202239990234375, 0.008166399955749512, 0.008175616264343261, 0.008168448448181152, 0.008178688049316407, 0.008253439903259278, 0.008184831619262695, 0.008218624114990235, 0.008167424201965333, 0.008147968292236327, 0.00809881591796875, 0.008159232139587403, 0.008144895553588867, 0.008196096420288086, 0.008171520233154296, 0.0081725435256958, 0.0081725435256958, 0.008155136108398438, 0.008146944046020508, 0.008144895553588867, 0.008152064323425292, 0.008246272087097169, 0.00818892765045166, 0.00820633602142334, 0.017756160736083985, 0.008161279678344726, 0.00819711971282959, 0.00819916820526123, 0.00821452808380127, 0.008217599868774414, 0.008182784080505372, 0.008183839797973633, 0.008160223960876465, 0.008208383560180664, 0.008182784080505372, 0.008232959747314453, 0.008248319625854492, 0.008203264236450195, 0.00820531177520752, 0.008327168464660644, 0.008169471740722656, 0.008141823768615723, 0.008276991844177246, 0.008146944046020508, 0.00818899154663086, 0.008183744430541992, 0.00819711971282959, 0.008224767684936523, 0.00819200038909912, 0.008167424201965333, 0.008166399955749512, 0.00800051212310791, 0.00779366397857666, 0.007783455848693848, 0.007816160202026367, 0.007880703926086426, 0.00788479995727539, 0.007886847972869874, 0.007890944004058837, 0.007907328128814697, 0.007920639991760254, 0.007906303882598878, 0.007896063804626464, 0.00787660789489746, 0.007787519931793213, 0.007804927825927735, 0.007823359966278077, 0.007795711994171142, 0.007813119888305664, 0.007845888137817383, 0.007881728172302246, 0.00790015983581543, 0.0078919677734375, 0.007890944004058837, 0.007923711776733398, 0.007927807807922363, 0.007887872219085693, 0.007857151985168457, 0.007902207851409913, 0.007882751941680909, 0.007894015789031983, 0.007897088050842285, 0.00790937614440918, 0.007885824203491211, 0.007881760120391846, 0.007926752090454101, 0.00787660789489746, 0.01720832061767578, 0.0078919677734375, 0.007903232097625732, 0.007897088050842285, 0.00788585615158081, 0.007939040184020996, 0.007898111820220948, 0.007946239948272706, 0.007913472175598145, 0.008083456039428711, 0.008185855865478516, 0.008216575622558593, 0.008212479591369629, 0.008209407806396484, 0.008053759574890136, 0.007925759792327881, 0.007935999870300293, 0.007925759792327881, 0.007913472175598145, 0.007906303882598878, 0.007914495944976807, 0.00786636781692505, 0.007880703926086426, 0.007903232097625732, 0.007961599826812745, 0.007911488056182862, 0.007871424198150635, 0.007821311950683594, 0.007815167903900147, 0.007855103969573975, 0.007812096118927002, 0.007820288181304931, 0.007916543960571289, 0.007880703926086426, 0.00786636781692505, 0.0078919677734375, 0.008148991584777832, 0.007803904056549072, 0.007786496162414551, 0.007812096118927002, 0.00780083179473877, 0.007873536109924317, 0.007898111820220948, 0.007906303882598878, 0.007950335979461669, 0.00790015983581543, 0.007948287963867188, 0.007911424160003662, 0.00791756820678711, 0.007919616222381591, 0.007916543960571289, 0.007890944004058837, 0.007797760009765625, 0.007795711994171142, 0.007799808025360107, 0.007833600044250488, 0.007790592193603516, 0.007904255867004394, 0.007924736022949219, 0.007878655910491944, 0.007888895988464355, 0.007879680156707763, 0.007928832054138184, 0.0172410888671875, 0.007907328128814697, 0.007896063804626464, 0.007915520191192627, 0.00790118408203125, 0.00790937614440918, 0.007914495944976807, 0.0079267840385437, 0.007911424160003662, 0.00790937614440918, 0.007933951854705811, 0.007922688007354736, 0.008061951637268066, 0.007914527893066407, 0.00790012788772583, 0.007907328128814697, 0.007904255867004394, 0.007920639991760254, 0.007902207851409913, 0.007924736022949219, 0.00790118408203125, 0.007902207851409913, 0.007888895988464355, 0.007878655910491944, 0.0078919677734375, 0.007897088050842285, 0.007868415832519531, 0.007885824203491211, 0.007894015789031983, 0.007887872219085693, 0.007890944004058837, 0.00791756820678711, 0.007890944004058837, 0.007879680156707763, 0.007865344047546387, 0.007877632141113282, 0.007871488094329833, 0.007950335979461669, 0.007863296031951903, 0.007905280113220215, 0.007861248016357422, 0.00787660789489746, 0.007882751941680909, 0.007885824203491211, 0.007904255867004394, 0.00788479995727539, 0.00790118408203125, 0.007897088050842285, 0.007953407764434815, 0.008878080368041993, 0.008540160179138183, 0.008606719970703124, 0.008441856384277344, 0.008605695724487305, 0.008258560180664062, 0.008253439903259278, 0.00823196792602539, 0.008240096092224121, 0.008230912208557128, 0.00820736026763916, 0.008218624114990235, 0.008193023681640625, 0.00818995189666748, 0.01782476806640625, 0.008174592018127442, 0.008221728324890137, 0.008191967964172363, 0.008193023681640625, 0.008229887962341309, 0.008227840423583984, 0.008202239990234375, 0.008233983993530274, 0.00821555233001709, 0.00820633602142334, 0.008216575622558593, 0.008195072174072266, 0.008235008239746093, 0.008251392364501953, 0.008165375709533691, 0.008159232139587403, 0.00818073558807373, 0.008204287528991699, 0.008169471740722656, 0.008005632400512695, 0.008017919540405273, 0.00807423973083496, 0.008037376403808593, 0.007845888137817383, 0.007797760009765625, 0.007871488094329833, 0.007904255867004394, 0.007889920234680176, 0.007863296031951903, 0.007873536109924317, 0.007870463848114014, 0.00790015983581543, 0.007889920234680176, 0.007880703926086426, 0.00780185604095459, 0.007803904056549072, 0.007815167903900147, 0.007857151985168457, 0.0077916159629821775, 0.007796735763549805, 0.007829504013061523, 0.00780083179473877, 0.007805952072143555, 0.007804927825927735, 0.007779327869415284, 0.007938047885894776, 0.007906303882598878, 0.007871488094329833, 0.007882751941680909, 0.007903232097625732, 0.007869440078735352, 0.007931903839111328, 0.007869440078735352, 0.007874559879302979, 0.007857151985168457, 0.007883776187896728, 0.007840767860412597, 0.007763967990875244, 0.007799808025360107, 0.007831552028656007, 0.007782400131225586, 0.008065024375915527, 0.01717043113708496, 0.00790015983581543, 0.007921664237976075, 0.007902207851409913, 0.007853055953979492, 0.008322048187255859, 0.008127488136291505, 0.007894015789031983, 0.008001567840576172, 0.007922656059265136, 0.007880703926086426, 0.007911424160003662, 0.007864319801330566, 0.007890944004058837, 0.007867392063140868, 0.007897088050842285, 0.007886847972869874, 0.007886847972869874, 0.007905280113220215, 0.007904255867004394, 0.007899136066436767, 0.007889920234680176, 0.007954432010650634, 0.007912447929382324, 0.007865375995635986, 0.007875552177429199, 0.007886847972869874, 0.007906303882598878, 0.007895040035247802, 0.007861248016357422, 0.007877632141113282, 0.007871520042419434, 0.007874527931213378, 0.007887872219085693, 0.007888895988464355, 0.00789299201965332, 0.007906303882598878, 0.007916543960571289, 0.00791756820678711, 0.007889920234680176, 0.007885824203491211, 0.00785920000076294, 0.007981056213378907, 0.008501248359680176, 0.008315903663635254, 0.008260607719421387, 0.008532992362976074, 0.008777728080749512, 0.008256511688232422, 0.008236031532287597, 0.008268799781799317, 0.008210432052612305, 0.008203264236450195, 0.008254464149475099, 0.008151040077209473, 0.008193023681640625, 0.008193023681640625, 0.008326175689697266, 0.008201184272766114, 0.008251392364501953, 0.008194047927856446, 0.008208383560180664, 0.00822169589996338, 0.01780121612548828, 0.008263680458068847, 0.008257535934448243, 0.008294400215148925, 0.008150015830993652, 0.00821350383758545, 0.008178688049316407, 0.00818995189666748, 0.00827187156677246, 0.008283136367797851, 0.008255488395690918, 0.008225791931152344, 0.008157183647155761, 0.00821350383758545, 0.00820633602142334, 0.008196096420288086, 0.008194080352783203, 0.007993311882019042, 0.007887872219085693, 0.007881728172302246, 0.007865344047546387, 0.007927807807922363, 0.008216575622558593, 0.008180800437927247, 0.008211392402648926, 0.008177663803100586, 0.008217663764953613, 0.008179648399353028, 0.008140800476074218, 0.008185855865478516, 0.008164352416992187, 0.008286208152770995, 0.008202239990234375, 0.008169471740722656, 0.008226816177368163, 0.008267775535583496, 0.008186880111694337, 0.008175616264343261, 0.008173567771911621, 0.008147968292236327, 0.008184831619262695, 0.008159232139587403, 0.008184831619262695, 0.008184831619262695, 0.008158207893371582, 0.007902207851409913, 0.007878655910491944, 0.007915520191192627, 0.007875648021697998, 0.007902143955230713, 0.007882751941680909, 0.007889920234680176, 0.007897088050842285, 0.007896063804626464, 0.007914495944976807, 0.007906303882598878, 0.007935008049011231, 0.007913440227508546, 0.007902207851409913, 0.00789299201965332, 0.007982079982757568, 0.007964672088623047, 0.007902207851409913, 0.01724723243713379, 0.007904255867004394, 0.007905280113220215, 0.007911424160003662, 0.007861248016357422, 0.008273920059204102, 0.00832102394104004, 0.00819916820526123, 0.008141823768615723, 0.008167424201965333, 0.008162303924560547, 0.00819711971282959, 0.008161279678344726, 0.008163328170776368, 0.008228863716125488, 0.00821555233001709, 0.008186880111694337, 0.008212479591369629, 0.008190976142883302, 0.00818995189666748, 0.008160256385803222, 0.00802406406402588, 0.007877632141113282, 0.00785203218460083, 0.007870463848114014, 0.007853055953979492, 0.007986176013946533, 0.007881728172302246, 0.007860223770141601, 0.00786636781692505, 0.007886847972869874, 0.007888895988464355, 0.007903232097625732, 0.007822336196899414, 0.007865407943725585, 0.0078755202293396, 0.007857151985168457, 0.007855103969573975, 0.007881728172302246, 0.007868415832519531, 0.0079267840385437, 0.007854080200195313, 0.007869440078735352, 0.007904255867004394, 0.007918591976165772, 0.007896063804626464, 0.007915520191192627, 0.007882751941680909, 0.007872511863708496, 0.00786636781692505, 0.007939072132110595, 0.007885824203491211, 0.007896096229553222, 0.0078919358253479, 0.007875584125518798, 0.007882751941680909, 0.00790015983581543, 0.007874559879302979, 0.007887872219085693, 0.007868415832519531, 0.007878655910491944, 0.007858208179473877, 0.007984096050262452, 0.017343488693237305, 0.007899136066436767, 0.007886847972869874, 0.007896063804626464, 0.00786636781692505, 0.0079267840385437, 0.007888895988464355, 0.007879680156707763, 0.007890944004058837, 0.007886847972869874, 0.007927807807922363, 0.007918591976165772, 0.00790015983581543, 0.007912511825561524, 0.007911359786987305, 0.00790835189819336, 0.007883776187896728, 0.007897088050842285, 0.00790835189819336, 0.007870463848114014, 0.007861248016357422, 0.00793497610092163, 0.007870463848114014, 0.007845888137817383, 0.007880703926086426, 0.007881728172302246, 0.007918591976165772, 0.007889920234680176, 0.00787660789489746, 0.007869440078735352, 0.007882751941680909, 0.007874591827392578, 0.007881696224212647, 0.008164352416992187, 0.008173567771911621, 0.008168448448181152, 0.00821555233001709, 0.00818073558807373, 0.00820736026763916, 0.008218624114990235, 0.008194047927856446, 0.008147968292236327, 0.008137727737426758, 0.00812339210510254, 0.00818380832672119, 0.008168448448181152, 0.008186880111694337, 0.008179712295532226, 0.008196096420288086, 0.008160256385803222, 0.008182784080505372, 0.008173567771911621, 0.008145919799804687, 0.008156160354614257, 0.008138751983642578, 0.008176639556884765, 0.008255488395690918, 0.008156160354614257, 0.008184831619262695, 0.008239104270935058, 0.008164352416992187, 0.008223744392395019, 0.008173600196838379, 0.01773257637023926, 0.008209407806396484, 0.00827187156677246, 0.008174592018127442, 0.00819916820526123, 0.008186880111694337, 0.008146944046020508, 0.008160256385803222, 0.008067071914672852, 0.008022015571594238, 0.007989247798919678, 0.007802879810333252, 0.007813119888305664, 0.007858176231384278, 0.007910399913787843, 0.007923711776733398, 0.007919616222381591, 0.007906303882598878, 0.007931903839111328, 0.007875584125518798, 0.007882751941680909, 0.007890944004058837, 0.007902207851409913, 0.007840767860412597, 0.007860223770141601, 0.007906303882598878, 0.007880703926086426, 0.007862271785736084, 0.007863296031951903, 0.007865344047546387, 0.008020992279052735, 0.007911424160003662, 0.007895040035247802, 0.007886847972869874, 0.007870463848114014, 0.007912447929382324, 0.007878655910491944, 0.007879680156707763, 0.007877632141113282, 0.007897088050842285, 0.007912447929382324, 0.007873536109924317, 0.007869440078735352, 0.007937024116516114, 0.007863296031951903, 0.007895040035247802, 0.007899136066436767, 0.0078919677734375, 0.007888895988464355, 0.007856128215789794, 0.007881728172302246, 0.007871488094329833, 0.007879680156707763, 0.0078919677734375, 0.00787660789489746, 0.007867392063140868, 0.007869440078735352, 0.007871488094329833, 0.007871488094329833, 0.007898111820220948, 0.008027135848999023, 0.007895040035247802, 0.007894015789031983, 0.01723289680480957, 0.007872511863708496, 0.007886847972869874, 0.00791964817047119, 0.007878623962402343, 0.00789299201965332, 0.007907328128814697, 0.00790835189819336, 0.007914495944976807, 0.007910399913787843, 0.007860223770141601, 0.00790118408203125, 0.007869440078735352, 0.007890944004058837, 0.007910399913787843, 0.0078919677734375, 0.007905280113220215, 0.007902207851409913, 0.007888895988464355, 0.007877632141113282, 0.007871488094329833, 0.007929855823516846, 0.007889920234680176, 0.007877632141113282, 0.007838719844818116, 0.007967743873596191, 0.0078919677734375, 0.007904255867004394, 0.00784281587600708, 0.007886847972869874, 0.00790118408203125, 0.007854080200195313, 0.0078919677734375, 0.007887872219085693, 0.00789299201965332, 0.007879680156707763, 0.00786636781692505, 0.00787660789489746, 0.007903232097625732, 0.007895040035247802, 0.00785920000076294, 0.007869440078735352, 0.00786636781692505, 0.007870463848114014, 0.00789299201965332, 0.007877632141113282, 0.007898111820220948, 0.007910399913787843, 0.007886847972869874, 0.007945216178894043, 0.00790118408203125, 0.00789299201965332, 0.00788479995727539, 0.007875584125518798, 0.007883776187896728, 0.007870463848114014, 0.007871488094329833, 0.007894015789031983, 0.007879680156707763, 0.007877632141113282, 0.00790835189819336, 0.007924736022949219, 0.007874559879302979, 0.01730361557006836, 0.007900095939636231, 0.00791756820678711, 0.007912447929382324, 0.007887872219085693, 0.007942143917083741, 0.007913472175598145, 0.007890944004058837, 0.007881728172302246, 0.007892032146453858, 0.007878592014312744, 0.00790835189819336, 0.007919616222381591, 0.007904255867004394, 0.007903232097625732, 0.007875584125518798, 0.00791756820678711, 0.007914495944976807, 0.007897088050842285, 0.007918591976165772, 0.00790118408203125, 0.00790937614440918, 0.007888927936553955, 0.00786736011505127, 0.007796735763549805, 0.00782643222808838, 0.007814144134521485, 0.007860223770141601, 0.007886847972869874, 0.0079267840385437, 0.007931903839111328, 0.007899136066436767, 0.007899136066436767, 0.007881728172302246, 0.007912447929382324, 0.0079267840385437, 0.007932928085327149, 0.007911456108093263, 0.007890944004058837, 0.008017888069152832, 0.007903232097625732, 0.007905280113220215, 0.007906303882598878, 0.007899136066436767, 0.007916543960571289, 0.007924736022949219, 0.007902207851409913, 0.007890944004058837, 0.00789299201965332, 0.007905280113220215, 0.007894015789031983, 0.007886847972869874, 0.007914495944976807, 0.00787766408920288, 0.0078438081741333, 0.007877632141113282, 0.00785203218460083, 0.007885824203491211, 0.00790835189819336, 0.008228863716125488, 0.00820531177520752, 0.00820019245147705, 0.008182784080505372, 0.017821695327758787, 0.008186880111694337, 0.008231936454772949, 0.00822169589996338, 0.008170495986938477, 0.008228863716125488, 0.008225791931152344, 0.008195072174072266, 0.008164352416992187, 0.008204287528991699, 0.008203264236450195, 0.008186880111694337, 0.00815007972717285, 0.008200127601623535, 0.00818892765045166, 0.008195072174072266, 0.008237055778503418, 0.008138751983642578, 0.008160256385803222, 0.008162303924560547, 0.008135680198669434, 0.008171520233154296, 0.008137727737426758, 0.008157183647155761, 0.008146944046020508, 0.00818995189666748, 0.008181759834289551, 0.008156160354614257, 0.008162303924560547, 0.00840601634979248, 0.008299519538879394, 0.008226816177368163, 0.008169471740722656, 0.0081725435256958, 0.00821350383758545, 0.008179712295532226, 0.008161279678344726, 0.008163328170776368, 0.008194047927856446, 0.0081725435256958, 0.008156160354614257, 0.008208383560180664, 0.00819711971282959, 0.0081725435256958, 0.008167424201965333, 0.00818892765045166, 0.008160287857055665, 0.00817353630065918, 0.008201215744018555, 0.008173567771911621, 0.008160256385803222, 0.008153087615966797, 0.008144895553588867, 0.008170495986938477, 0.00818073558807373, 0.008195072174072266, 0.00819200038909912, 0.00818380832672119, 0.008269824028015137, 0.00818892765045166, 0.008237055778503418, 0.008267775535583496, 0.008158207893371582, 0.017276927947998046, 0.007921664237976075, 0.007931903839111328, 0.007938047885894776, 0.007931903839111328, 0.007923711776733398, 0.007915520191192627, 0.007874559879302979, 0.007900191783905029, 0.007911392211914062, 0.007922751903533936, 0.007906239986419678, 0.007880703926086426, 0.007888895988464355, 0.007903232097625732, 0.007906303882598878, 0.007923711776733398, 0.007904255867004394, 0.007928832054138184, 0.007916543960571289, 0.00788479995727539, 0.008005632400512695, 0.008069120407104492, 0.008210432052612305, 0.00820736026763916, 0.008160256385803222, 0.008229887962341309, 0.008371199607849121, 0.008203264236450195, 0.008147968292236327, 0.008193023681640625, 0.00818073558807373, 0.008166399955749512, 0.008162303924560547, 0.008184831619262695, 0.00819711971282959, 0.008153087615966797, 0.008143872261047362, 0.008269887924194336, 0.008174528121948242, 0.008178688049316407, 0.008060928344726562, 0.007870463848114014, 0.007886847972869874, 0.007877632141113282, 0.007862271785736084, 0.007861248016357422, 0.007875584125518798, 0.007878655910491944, 0.007840767860412597, 0.007856192111968995, 0.007905216217041015, 0.007863296031951903, 0.007905280113220215, 0.00795036792755127, 0.007897056102752686, 0.007879680156707763, 0.007808000087738037, 0.007794688224792481, 0.00778547191619873, 0.00779366397857666, 0.007828479766845703, 0.007899136066436767, 0.01720319938659668, 0.007896063804626464, 0.007902207851409913, 0.007902207851409913, 0.007933951854705811, 0.007899168014526368, 0.007875552177429199, 0.007885824203491211, 0.00790118408203125, 0.007880703926086426, 0.007872543811798095, 0.007870431900024413, 0.007860223770141601, 0.007912447929382324, 0.007863296031951903, 0.007904255867004394, 0.007878655910491944, 0.00786636781692505, 0.007889952182769775, 0.008605664253234863, 0.008276991844177246, 0.008180768013000488, 0.008194016456604005, 0.008203264236450195, 0.008166399955749512, 0.008138751983642578, 0.008159232139587403, 0.008185855865478516, 0.00828825569152832, 0.008177663803100586, 0.008231936454772949, 0.009232383728027344, 0.009074687957763672, 0.008319999694824219, 0.008264703750610352, 0.008244223594665527, 0.008226816177368163, 0.00821452808380127, 0.00821350383758545, 0.008294400215148925, 0.008179712295532226, 0.008155136108398438, 0.008151040077209473, 0.008144895553588867, 0.008149056434631348, 0.008166336059570312, 0.008150015830993652, 0.008157183647155761, 0.008175616264343261, 0.008162303924560547, 0.008129535675048828, 0.008153087615966797, 0.008181759834289551, 0.008162303924560547, 0.008159232139587403, 0.008168448448181152, 0.008137727737426758, 0.008223744392395019, 0.008177663803100586, 0.008250368118286134, 0.007933951854705811, 0.007960576057434082, 0.007880703926086426, 0.017229824066162108, 0.007864319801330566, 0.007895040035247802, 0.007912447929382324, 0.007916543960571289, 0.007874559879302979, 0.007874559879302979, 0.007890944004058837, 0.007875616073608399, 0.007880671977996826, 0.007912447929382324, 0.007871488094329833, 0.007870463848114014, 0.007885824203491211, 0.007863296031951903, 0.007890944004058837, 0.007896063804626464, 0.007872511863708496, 0.007821311950683594, 0.00780185604095459, 0.007768064022064209, 0.007796735763549805, 0.00775270414352417, 0.007776256084442139, 0.00779366397857666, 0.00783462381362915, 0.007882751941680909, 0.007856128215789794, 0.007874559879302979, 0.007960576057434082, 0.007889920234680176, 0.00786844778060913, 0.007910367965698242, 0.007907328128814697, 0.00790937614440918, 0.007875584125518798, 0.007925759792327881, 0.00787660789489746, 0.007921664237976075, 0.007858176231384278, 0.007813119888305664, 0.007776256084442139, 0.007779359817504883, 0.007778272151947021, 0.007781375885009765, 0.007783423900604248, 0.007788544178009033, 0.007868415832519531, 0.008491007804870606, 0.00829644775390625, 0.008425472259521484, 0.008185855865478516, 0.008160256385803222, 0.00818892765045166, 0.00818892765045166, 0.008158207893371582, 0.008307711601257324, 0.008154111862182617, 0.008186880111694337, 0.008160256385803222, 0.008375295639038086, 0.00827187156677246, 0.008228863716125488, 0.017871871948242187, 0.00821350383758545, 0.008118271827697754, 0.008087615966796875, 0.008040384292602539, 0.00809267234802246, 0.008062975883483887, 0.008042495727539062, 0.00788479995727539, 0.007778304100036621, 0.007779327869415284, 0.007795711994171142, 0.00785100793838501, 0.007890944004058837, 0.008228863716125488, 0.008224767684936523, 0.008241151809692383, 0.008231936454772949, 0.00818892765045166, 0.008136704444885253, 0.008165375709533691, 0.008133631706237793, 0.008110079765319824, 0.008011775970458984, 0.008096768379211425, 0.008028160095214844, 0.008018943786621094, 0.00783462381362915, 0.007794688224792481, 0.007845888137817383, 0.007902207851409913, 0.0078919677734375, 0.00788479995727539, 0.008270848274230956, 0.008190976142883302, 0.00819814395904541, 0.008216575622558593, 0.008133631706237793, 0.008177663803100586, 0.008167424201965333, 0.008155136108398438, 0.008165375709533691, 0.008184831619262695, 0.008193056106567382, 0.008167391777038574, 0.008281087875366211, 0.00818995189666748, 0.008165375709533691, 0.00816646385192871, 0.00813766384124756, 0.008171520233154296, 0.008169471740722656, 0.008249343872070313, 0.008170528411865234, 0.008159199714660644, 0.008159232139587403, 0.00816646385192871, 0.008050623893737794, 0.007875584125518798, 0.007879680156707763, 0.007878655910491944, 0.007945216178894043, 0.007898111820220948, 0.01722470474243164, 0.00790835189819336, 0.00790015983581543, 0.007928832054138184, 0.00791756820678711, 0.007931903839111328, 0.007912447929382324, 0.007899136066436767, 0.007907392024993897, 0.007789504051208496, 0.007820288181304931, 0.007802879810333252, 0.007817215919494629, 0.00782643222808838, 0.007779327869415284, 0.007822336196899414, 0.007856128215789794, 0.007896096229553222, 0.007898079872131347, 0.007916543960571289, 0.007932928085327149, 0.007896063804626464, 0.007888895988464355, 0.007906303882598878, 0.00787660789489746, 0.00785203218460083, 0.00790835189819336, 0.007948287963867188, 0.007944191932678223, 0.008835071563720704, 0.008396800041198731, 0.008232959747314453, 0.008202239990234375, 0.008181759834289551, 0.008441856384277344, 0.008204287528991699, 0.00819916820526123, 0.00821555233001709, 0.008211456298828124, 0.008182784080505372, 0.008186880111694337, 0.008201215744018555, 0.008148991584777832, 0.008391679763793946, 0.008201215744018555, 0.008293375968933106, 0.008458239555358887, 0.008228863716125488, 0.00819814395904541, 0.008233983993530274, 0.008269824028015137, 0.00821555233001709, 0.008225791931152344, 0.008218624114990235, 0.008053759574890136, 0.008059904098510743, 0.008062975883483887, 0.008028160095214844, 0.008202239990234375, 0.008179712295532226, 0.008186880111694337, 0.008274944305419921, 0.008225791931152344, 0.017954816818237306, 0.00827187156677246, 0.0081725435256958, 0.008227904319763184, 0.008158143997192382, 0.00819916820526123, 0.008041472434997558, 0.008038399696350097, 0.007941120147705078, 0.007833600044250488, 0.007811071872711181, 0.007898111820220948, 0.007895040035247802, 0.007914495944976807, 0.007872511863708496, 0.007804927825927735, 0.008203264236450195, 0.008086527824401855, 0.00819916820526123, 0.008169471740722656, 0.008151040077209473, 0.008156160354614257, 0.00818380832672119, 0.008162303924560547, 0.00820019245147705, 0.00818995189666748, 0.008207391738891602, 0.008174559593200684, 0.008146944046020508, 0.008261631965637208, 0.008174592018127442, 0.008193023681640625, 0.008158207893371582, 0.008162303924560547, 0.008375295639038086, 0.008216575622558593, 0.00821350383758545, 0.00819200038909912, 0.008142848014831543, 0.008150015830993652, 0.008171520233154296, 0.008144895553588867, 0.008160256385803222, 0.008120320320129394, 0.008193023681640625, 0.008174592018127442, 0.008244223594665527, 0.0081725435256958, 0.008279040336608886, 0.00818892765045166, 0.008140800476074218, 0.00790015983581543, 0.008181759834289551, 0.008391679763793946, 0.008224767684936523, 0.008196096420288086, 0.008192031860351563, 0.008178655624389648, 0.008195072174072266, 0.00820531177520752, 0.008159232139587403, 0.008219648361206054, 0.008171520233154296]",tokens/s,121.89262917291767,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Salesforce/codegen-6B-nl,,cuda,0,42,,,,,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,d35829e539df8480b726c647eeabf91e41eae047,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,,MB,1293.193216,15339.094016,0.0,14692.646912,14512.889344,s,10,0.991220542907715,0.09912205429077149,0.0017180429556808861,0.09854598236083985,0.09924651260375977,0.10175880775451661,0.10376864387512208,"[0.10427110290527344, 0.09868822479248048, 0.09859490966796874, 0.0985947494506836, 0.09848131561279297, 0.09845728302001953, 0.09865718078613281, 0.0984832992553711, 0.0984972152709961, 0.09849526214599609]",tokens/s,2582.6744797785555,kWh,1.1663765267089561e-06,6.391110982483095e-07,6.150544136117673e-06,7.956031761074939e-06,tokens/kWh,32176844.900555283,MB,1293.193216,15339.094016,0.0,14692.646912,14645.240832,s,10,25.872080810546876,2.587208081054688,0.005879826434083232,2.5851978759765624,2.5924181640625,2.5974760498046874,2.6015223583984373,"[2.602533935546875, 2.589749755859375, 2.591294189453125, 2.58463671875, 2.58274462890625, 2.585759033203125, 2.5870078125, 2.58321630859375, 2.58270849609375, 2.582429931640625]",tokens/s,24.350573292240856,kWh,3.0502337609176272e-05,1.6716556825863603e-05,0.0001228022534704814,0.0001700211479055213,tokens/kWh,370542.1400578259,,s,629,26.717135948181152,0.04247557384448514,0.011740700365855691,0.041030654907226564,0.04134707260131836,0.04154818649291992,0.13978116455078124,"[0.04153343963623047, 0.04160921478271484, 0.04155801773071289, 0.04172185516357422, 0.041798656463623046, 0.04166764831542969, 0.0418037109375, 0.041837566375732424, 0.04157952117919922, 0.04144537734985351, 0.04150579071044922, 0.04227481460571289, 0.04138700866699219, 0.04106752014160156, 0.041060352325439455, 0.04107161712646484, 0.04100815963745117, 0.041103328704833984, 0.041169921875, 0.04095078277587891, 0.041109504699707033, 0.041076736450195314, 0.041073665618896485, 0.04106649780273437, 0.04105420684814453, 0.0411146240234375, 0.04103577423095703, 0.041076736450195314, 0.04113715362548828, 0.041169921875, 0.04112588882446289, 0.041299968719482424, 0.04132761764526367, 0.041212928771972655, 0.041166847229003906, 0.04117708969116211, 0.04112998580932617, 0.041163776397705076, 0.04112588882446289, 0.041188350677490236, 0.041115646362304685, 0.041145343780517575, 0.04116284942626953, 0.041113502502441404, 0.04120576095581055, 0.04131840133666992, 0.04127948760986328, 0.041286655426025394, 0.04140851211547852, 0.041390079498291016, 0.04129177474975586, 0.041319423675537106, 0.041332736968994144, 0.04134195327758789, 0.041289726257324216, 0.04135424041748047, 0.04134400177001953, 0.04135628890991211, 0.04137267303466797, 0.04125491333007812, 0.04132556915283203, 0.041316352844238284, 0.14016307067871095, 0.04084121704101563, 0.04090060806274414, 0.040995838165283204, 0.0411596794128418, 0.041218048095703126, 0.041003009796142575, 0.04120883178710937, 0.041286655426025394, 0.0410695686340332, 0.04097945785522461, 0.04101222229003906, 0.04100096130371094, 0.04084121704101563, 0.041073665618896485, 0.04097536087036133, 0.040970241546630856, 0.0411412467956543, 0.041106433868408204, 0.04097740936279297, 0.04073164749145508, 0.04108697509765625, 0.04118527984619141, 0.04117401504516602, 0.041016319274902346, 0.04104294586181641, 0.041180160522460936, 0.04106649780273437, 0.04105318450927734, 0.0413040657043457, 0.04124671936035156, 0.041095169067382815, 0.04132352066040039, 0.041361438751220704, 0.041285598754882816, 0.04127436828613281, 0.04134502410888672, 0.04131327819824219, 0.041253887176513675, 0.04115868759155274, 0.04095894241333008, 0.04094566345214844, 0.040978431701660156, 0.040904705047607424, 0.040883201599121094, 0.040910846710205076, 0.04095180892944336, 0.041014270782470705, 0.04101324844360352, 0.04100096130371094, 0.04101222229003906, 0.04104499053955078, 0.041081951141357424, 0.041060256958007815, 0.04104297637939453, 0.041067489624023436, 0.04099993515014649, 0.04108595275878906, 0.041224193572998044, 0.04142182540893555, 0.0413757438659668, 0.041605121612548826, 0.04143820953369141, 0.1398835144042969, 0.040653823852539066, 0.04069580841064453, 0.040784896850585936, 0.040714241027832034, 0.04073577499389648, 0.040729568481445315, 0.04091187286376953, 0.04090163040161133, 0.04077875137329102, 0.04075929641723633, 0.04124467086791992, 0.04134195327758789, 0.04207923126220703, 0.04134912109375, 0.041181182861328124, 0.0411514892578125, 0.041095169067382815, 0.04113510513305664, 0.04119244766235351, 0.04118527984619141, 0.041183231353759765, 0.04132761764526367, 0.04120883178710937, 0.04124364852905273, 0.04115353775024414, 0.041164798736572264, 0.04118732833862305, 0.04130918502807617, 0.0412303352355957, 0.04090675354003906, 0.04114432144165039, 0.04124879837036133, 0.041250816345214845, 0.04127945709228516, 0.04128870391845703, 0.041583614349365236, 0.041420833587646484, 0.041315296173095706, 0.04130815887451172, 0.041062400817871096, 0.04104806518554688, 0.04101939010620117, 0.04098457717895508, 0.040997886657714845, 0.041062400817871096, 0.041024513244628906, 0.04101939010620117, 0.040997920989990236, 0.04108284759521484, 0.04101529693603516, 0.04105215835571289, 0.041073665618896485, 0.0409692497253418, 0.041034721374511716, 0.041059326171875, 0.04127129745483398, 0.041371646881103515, 0.04106137466430664, 0.04108697509765625, 0.041523200988769535, 0.041434112548828124, 0.04140031814575195, 0.13996543884277343, 0.04087807846069336, 0.0409804801940918, 0.04107059097290039, 0.0410777587890625, 0.04115456008911133, 0.04108697509765625, 0.041133056640625, 0.04113919830322266, 0.04088422393798828, 0.040784896850585936, 0.04085971069335938, 0.04085958480834961, 0.040904705047607424, 0.04083200073242187, 0.04087807846069336, 0.04083097457885742, 0.04081151962280274, 0.04086783981323242, 0.040941566467285154, 0.040892414093017575, 0.04082995223999023, 0.0408985595703125, 0.04091494369506836, 0.040880126953125, 0.040834049224853515, 0.040891391754150394, 0.040954879760742184, 0.04089651107788086, 0.04092620849609375, 0.04093132781982422, 0.040970241546630856, 0.04093132781982422, 0.04091187286376953, 0.04095078277587891, 0.04101836776733398, 0.0410777587890625, 0.041063423156738284, 0.040959999084472655, 0.041082878112792966, 0.041752574920654296, 0.04126105499267578, 0.04124467086791992, 0.04131123352050781, 0.04134912109375, 0.04122214508056641, 0.041024513244628906, 0.04111360168457031, 0.04109724807739258, 0.04099580764770508, 0.04101836776733398, 0.041060352325439455, 0.04106547164916992, 0.04112179183959961, 0.04112691116333008, 0.04114739227294922, 0.04104806518554688, 0.04107468795776367, 0.04110335922241211, 0.04113100814819336, 0.041099262237548825, 0.041095169067382815, 0.0410777587890625, 0.13973811340332032, 0.040738815307617186, 0.040768512725830076, 0.04083609771728516, 0.04079206466674805, 0.04076236724853516, 0.04074905776977539, 0.04086272048950195, 0.04079411315917969, 0.04083097457885742, 0.040809471130371096, 0.04087807846069336, 0.04083302307128906, 0.040793087005615236, 0.04088217544555664, 0.040834049224853515, 0.04080332946777344, 0.040910846710205076, 0.040842239379882815, 0.04085657501220703, 0.04079718399047851, 0.040891391754150394, 0.04085964965820312, 0.040850433349609375, 0.040855552673339846, 0.040871936798095705, 0.04091904067993164, 0.04110233688354492, 0.04109209442138672, 0.04092825698852539, 0.041196544647216796, 0.041090049743652345, 0.04090367889404297, 0.04093648147583008, 0.04095177459716797, 0.04090777587890625, 0.04094771194458008, 0.04095795059204101, 0.04100505447387695, 0.04098051071166992, 0.04091183853149414, 0.04097644805908203, 0.041048000335693356, 0.041078784942626956, 0.04096819305419922, 0.04101529693603516, 0.04101529693603516, 0.04107980728149414, 0.040992801666259765, 0.04099887847900391, 0.04106649780273437, 0.04110131072998047, 0.0415283203125, 0.04128460693359375, 0.04113817596435547, 0.04181507110595703, 0.04162249755859375, 0.04131123352050781, 0.04111360168457031, 0.04111769485473633, 0.041093120574951174, 0.04134707260131836, 0.041275390625, 0.13978317260742187, 0.04070912170410156, 0.04072447967529297, 0.04085145568847656, 0.041008129119873046, 0.04102143859863281, 0.04084121704101563, 0.04084531021118164, 0.04074905776977539, 0.040799232482910154, 0.04087295913696289, 0.04110233688354492, 0.04108595275878906, 0.040837120056152344, 0.041132030487060545, 0.04108697509765625, 0.04080230331420898, 0.04073779296875, 0.040796161651611325, 0.040891391754150394, 0.04082486343383789, 0.04083299255371094, 0.04087295913696289, 0.04090572738647461, 0.040799232482910154, 0.04078899383544922, 0.04084940719604492, 0.04095900726318359, 0.041207775115966794, 0.04115251159667969, 0.04118527984619141, 0.04106444931030274, 0.04088115310668945, 0.04093337631225586, 0.040965118408203126, 0.04106752014160156, 0.04098252868652344, 0.04094259262084961, 0.040948734283447266, 0.04101939010620117, 0.0409804801940918, 0.040995838165283204, 0.040973312377929685, 0.04124671936035156, 0.04134707260131836, 0.041011199951171876, 0.041022464752197264, 0.041090049743652345, 0.04118425750732422, 0.04102860641479492, 0.04114432144165039, 0.041247745513916016, 0.041017345428466793, 0.041038848876953124, 0.04130201721191406, 0.041204734802246096, 0.041078784942626956, 0.04111257553100586, 0.04127743911743164, 0.041940990447998046, 0.041409534454345705, 0.041996318817138674, 0.04162966537475586, 0.14006886291503906, 0.0406978874206543, 0.040834049224853515, 0.04084323120117188, 0.04074393463134766, 0.04080537414550781, 0.04097334289550781, 0.04080944061279297, 0.04091494369506836, 0.04117913436889648, 0.04101222229003906, 0.04081049728393555, 0.04076134490966797, 0.04101945495605469, 0.04082886505126953, 0.040883201599121094, 0.04096102523803711, 0.04096921539306641, 0.04172390365600586, 0.04139724731445313, 0.04121702575683594, 0.04117299270629883, 0.04093644714355469, 0.04104499053955078, 0.04106649780273437, 0.040927230834960936, 0.04098252868652344, 0.041164798736572264, 0.041093120574951174, 0.04099174499511719, 0.040899585723876954, 0.04094259262084961, 0.04098355102539063, 0.040970241546630856, 0.04093644714355469, 0.04097740936279297, 0.04093030548095703, 0.040940544128417966, 0.040948734283447266, 0.0409169921875, 0.04096819305419922, 0.04097433471679687, 0.04105420684814453, 0.041306110382080076, 0.04133990478515625, 0.04127641677856445, 0.04133990478515625, 0.041376766204833985, 0.041312286376953125, 0.041029598236083986, 0.04104191970825195, 0.04117708969116211, 0.0412149772644043, 0.04101324844360352, 0.04109107208251953, 0.04136243057250977, 0.04111667251586914, 0.04135116958618164, 0.04128153610229492, 0.041188350677490236, 0.041350143432617184, 0.04123648071289063, 0.04113715362548828, 0.13997669982910158, 0.04078079986572265, 0.04114227294921875, 0.04093030548095703, 0.04080537414550781, 0.04074393463134766, 0.040871936798095705, 0.041097278594970706, 0.040892353057861326, 0.04077260971069336, 0.04086272048950195, 0.04116889572143555, 0.04096102523803711, 0.04081769561767578, 0.04090262222290039, 0.04084531021118164, 0.0410511360168457, 0.041006080627441405, 0.041011199951171876, 0.040869888305664064, 0.040812545776367185, 0.04151295852661133, 0.04152012634277344, 0.04115456008911133, 0.04090572738647461, 0.04088934326171875, 0.040892414093017575, 0.04098252868652344, 0.04086067199707031, 0.040981502532958985, 0.04093030548095703, 0.04097743988037109, 0.040901599884033205, 0.0408985595703125, 0.04095590209960937, 0.04096307373046875, 0.040994815826416016, 0.04093952178955078, 0.04098355102539063, 0.04103168106079102, 0.040981502532958985, 0.04094057464599609, 0.040964065551757814, 0.040992767333984374, 0.040933441162109375, 0.04096198272705078, 0.041011199951171876, 0.041111553192138675, 0.041027584075927735, 0.04098252868652344, 0.0410184326171875, 0.04112377548217774, 0.04106547164916992, 0.04111974334716797, 0.041073665618896485, 0.04115865707397461, 0.04110028839111328, 0.04108492660522461, 0.041060352325439455, 0.041145343780517575, 0.04109619140625, 0.041073665618896485, 0.04110745620727539, 0.1397760009765625, 0.040683521270751956, 0.04065792083740234, 0.040648704528808595, 0.04064051055908203, 0.04068864059448242, 0.0407562255859375, 0.04078185653686523, 0.04077769470214844, 0.04079206466674805, 0.040880126953125, 0.04092211151123047, 0.04090572738647461, 0.040826911926269534, 0.040865760803222656, 0.04090163040161133, 0.040908798217773434, 0.04086374282836914, 0.040839168548583986, 0.04087807846069336, 0.040866817474365234, 0.04088524627685547, 0.04095078277587891, 0.04091596984863281, 0.04092006301879883, 0.04104806518554688, 0.041583614349365236, 0.04112793731689453, 0.040943614959716795, 0.04095078277587891, 0.04096307373046875, 0.040970241546630856, 0.041008129119873046, 0.040966209411621095, 0.041135040283203125, 0.041016319274902346, 0.04100710296630859, 0.04174848175048828, 0.04137472152709961, 0.04109619140625, 0.041011199951171876, 0.0409989128112793, 0.04099379348754883, 0.04112384033203125, 0.041016319274902346, 0.041047039031982424, 0.04111775970458984, 0.04111254501342773, 0.04104393768310547, 0.04102656173706055, 0.041105438232421875, 0.041083873748779295, 0.041025535583496094, 0.041025535583496094, 0.04108902359008789, 0.04110438537597656, 0.041057281494140625, 0.041093120574951174, 0.04111769485473633, 0.04114739227294922, 0.0410880012512207, 0.04115251159667969, 0.04118937683105469, 0.13979136657714844, 0.04074803161621094, 0.04074700927734375, 0.040793087005615236, 0.04079206466674805, 0.0408350715637207, 0.04074905776977539, 0.04080646514892578, 0.04078176116943359, 0.040828929901123044, 0.04083609771728516, 0.04089651107788086, 0.040812545776367185, 0.04080847930908203, 0.04085449600219727, 0.04094259262084961, 0.04082995223999023, 0.04083510589599609, 0.04089955139160156, 0.0408985595703125, 0.040880126953125, 0.040828929901123044, 0.040871936798095705, 0.04096614456176758, 0.04090675354003906, 0.040891391754150394, 0.040904705047607424, 0.041014270782470705, 0.0408985595703125, 0.04087705612182617, 0.040930335998535156, 0.04104291152954102, 0.04095078277587891, 0.04094464111328125, 0.04095795059204101, 0.04097945785522461, 0.040948734283447266, 0.04095180892944336, 0.040995838165283204, 0.04096716690063477, 0.040965152740478517, 0.04125078582763672, 0.041059326171875, 0.041093120574951174, 0.041761791229248044, 0.04213145446777344, 0.04130713653564453, 0.041040897369384766, 0.040995838165283204, 0.04108083343505859, 0.04113622283935547, 0.04113711929321289, 0.04105516815185547, 0.04109209442138672, 0.04102656173706055, 0.041030654907226564, 0.04096921539306641, 0.04108595275878906, 0.04110131072998047, 0.041099262237548825, 0.04108083343505859, 0.041148414611816404, 0.041115680694580076]",tokens/s,23.542942672446934,,,main,False,False,,,,,,,,,,,,,,,,,,,,,,, float16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-410m,,cuda,0,42,,,,,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,217063f5c507ed7cc255df7e1f64c4333a0b4dfe,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,,MB,1295.101952,1711.80032,0.0,1065.353216,1018.226688,s,10,0.2013744029998779,0.02013744029998779,0.0008730184027724601,0.01980724811553955,0.020394617843627928,0.021564012908935542,0.02249952896118164,"[0.022733407974243162, 0.01984217643737793, 0.02013475227355957, 0.01979795265197754, 0.01981654357910156, 0.019989408493041993, 0.01977084732055664, 0.019749568939208983, 0.01976016044616699, 0.019779584884643556]",tokens/s,12712.638557153423,kWh,2.3687972911772287e-07,1.297966301216819e-07,6.353857231823793e-07,1.0020620824217839e-06,tokens/kWh,255473193.21902606,MB,1295.101952,1711.80032,0.0,1065.353216,1058.88,s,10,11.73953161621094,1.1739531616210939,0.011730416071497548,1.1694104003906252,1.1823773559570314,1.1941164123535157,1.203507657470703,"[1.20585546875, 1.1772835693359376, 1.1797686767578126, 1.1686829833984376, 1.169909912109375, 1.164593017578125, 1.165302490234375, 1.1649390869140626, 1.168910888671875, 1.1742855224609374]",tokens/s,53.66483268634352,kWh,1.3735118035257752e-05,7.526501696206713e-06,2.3843189415217027e-05,4.5104809146681495e-05,tokens/kWh,1396746.8478831844,,s,629,11.897618431091306,0.018915132640844688,0.002428042337562234,0.01850160026550293,0.018991923141479492,0.019597497177124022,0.038714585723876954,"[0.019733535766601563, 0.019600351333618163, 0.01957689666748047, 0.01968841552734375, 0.01974064064025879, 0.019740671157836915, 0.020624383926391602, 0.01965260887145996, 0.01963417625427246, 0.01960038375854492, 0.019558399200439454, 0.019720191955566405, 0.01971004867553711, 0.01987164878845215, 0.01964339256286621, 0.01983180809020996, 0.01967616081237793, 0.0196495361328125, 0.019817472457885742, 0.02000588798522949, 0.019535871505737306, 0.01959321594238281, 0.01948057556152344, 0.019458047866821288, 0.019532800674438477, 0.01983897590637207, 0.02004377555847168, 0.020068351745605468, 0.018985984802246093, 0.018689023971557618, 0.01847091293334961, 0.01883955192565918, 0.018696191787719727, 0.018479103088378905, 0.01863580894470215, 0.018444255828857423, 0.018516992568969725, 0.01856716728210449, 0.018562047958374024, 0.01842278480529785, 0.018962432861328125, 0.01863065528869629, 0.019122175216674805, 0.01839923286437988, 0.018315263748168945, 0.01861222457885742, 0.018464767456054687, 0.018489343643188477, 0.018540544509887694, 0.018570240020751954, 0.01864499282836914, 0.018524160385131837, 0.01840640068054199, 0.018461696624755858, 0.018490367889404297, 0.018449407577514648, 0.01869004821777344, 0.018420736312866212, 0.01848320007324219, 0.019054592132568358, 0.01873209571838379, 0.01857632064819336, 0.038790145874023435, 0.018527231216430663, 0.01846784019470215, 0.018348031997680665, 0.018406463623046876, 0.01850060844421387, 0.018520000457763672, 0.01841152000427246, 0.018647039413452148, 0.01882521629333496, 0.01848320007324219, 0.018485248565673826, 0.018715648651123046, 0.018749439239501953, 0.01869004821777344, 0.01842790412902832, 0.018449407577514648, 0.01839923286437988, 0.01840742492675781, 0.018485248565673826, 0.01887539291381836, 0.01845043182373047, 0.018482175827026368, 0.018689023971557618, 0.0187064323425293, 0.01866547203063965, 0.01880268859863281, 0.018959360122680666, 0.018705408096313478, 0.018324480056762696, 0.018431999206542968, 0.018770944595336913, 0.018502656936645507, 0.018391040802001952, 0.018770944595336913, 0.018840576171875, 0.018780160903930664, 0.01887027168273926, 0.01884876823425293, 0.01878118324279785, 0.018873344421386717, 0.01902284812927246, 0.01881395149230957, 0.018721792221069337, 0.018758655548095703, 0.018882560729980468, 0.018717695236206054, 0.018635839462280274, 0.018702272415161134, 0.019133440017700197, 0.018856960296630858, 0.018766847610473633, 0.018860031127929687, 0.018561023712158203, 0.01861734390258789, 0.01884160041809082, 0.019008575439453126, 0.018829248428344728, 0.01879347229003906, 0.018949119567871094, 0.018926591873168946, 0.018877439498901367, 0.018648128509521484, 0.03926111984252929, 0.01864396858215332, 0.01859584045410156, 0.0188221435546875, 0.01885593605041504, 0.018736127853393555, 0.018884607315063476, 0.018957311630249024, 0.01899622344970703, 0.019158016204833983, 0.01900441551208496, 0.018893823623657227, 0.018694143295288086, 0.018671615600585938, 0.018991104125976564, 0.01882624053955078, 0.018612287521362306, 0.018602943420410155, 0.018579456329345705, 0.01902796745300293, 0.018968576431274413, 0.018755584716796874, 0.018506752014160157, 0.01843814468383789, 0.018508800506591795, 0.018542591094970702, 0.01847603225708008, 0.018506784439086914, 0.018529247283935547, 0.018777088165283205, 0.01879347229003906, 0.018507776260375978, 0.018521087646484375, 0.018545663833618165, 0.018545663833618165, 0.018413568496704103, 0.018512895584106445, 0.018485248565673826, 0.019091455459594727, 0.019005439758300782, 0.01860095977783203, 0.018436159133911133, 0.01873094367980957, 0.01905971145629883, 0.019960832595825196, 0.019219455718994142, 0.01899519920349121, 0.01901875114440918, 0.018957311630249024, 0.018531328201293946, 0.018790399551391602, 0.01903923225402832, 0.018548736572265623, 0.01843302345275879, 0.01847603225708008, 0.01843612861633301, 0.018780128479003906, 0.0186429443359375, 0.018315263748168945, 0.018357248306274415, 0.01841766357421875, 0.018472959518432617, 0.01846784019470215, 0.038719486236572266, 0.018576383590698242, 0.018502656936645507, 0.018482175827026368, 0.018485248565673826, 0.018332735061645507, 0.018309055328369142, 0.018417728424072265, 0.01851897621154785, 0.018501632690429686, 0.018529279708862305, 0.018502656936645507, 0.0184453125, 0.018511871337890624, 0.018479103088378905, 0.018515968322753908, 0.01845452880859375, 0.01848422431945801, 0.018456575393676757, 0.018542591094970702, 0.018295808792114256, 0.018457599639892578, 0.018528255462646484, 0.018367488861083983, 0.01828556823730469, 0.018336767196655272, 0.01845248031616211, 0.018487295150756835, 0.018921472549438476, 0.018703359603881836, 0.018774015426635742, 0.018947071075439453, 0.018489343643188477, 0.018540544509887694, 0.01845145606994629, 0.018495487213134765, 0.018534400939941405, 0.01864089584350586, 0.01840947151184082, 0.01845248031616211, 0.018779136657714843, 0.019087360382080077, 0.018893823623657227, 0.01939967918395996, 0.019009536743164062, 0.018570240020751954, 0.01887129592895508, 0.018463743209838866, 0.01846272087097168, 0.018461696624755858, 0.01838591957092285, 0.01847500801086426, 0.018489343643188477, 0.018489343643188477, 0.018463743209838866, 0.018916351318359375, 0.018341888427734376, 0.01844326400756836, 0.018463743209838866, 0.018440191268920898, 0.01843097686767578, 0.018356224060058594, 0.01848320007324219, 0.03881881713867188, 0.018539520263671876, 0.018508800506591795, 0.018479103088378905, 0.018525184631347655, 0.018542591094970702, 0.01855897521972656, 0.018515968322753908, 0.01880575942993164, 0.018932735443115235, 0.018707456588745116, 0.018548736572265623, 0.018448383331298827, 0.018548736572265623, 0.01862553596496582, 0.018714624404907225, 0.01848422431945801, 0.018563072204589845, 0.018626560211181642, 0.018496511459350586, 0.018405376434326173, 0.018447359085083007, 0.01843302345275879, 0.018391040802001952, 0.018488319396972656, 0.018447359085083007, 0.018513919830322266, 0.018576448440551757, 0.018470848083496094, 0.018478080749511717, 0.018491392135620118, 0.018525184631347655, 0.018518016815185546, 0.018533376693725585, 0.018585599899291993, 0.018440191268920898, 0.018487295150756835, 0.018603008270263673, 0.01844326400756836, 0.018392127990722658, 0.018435039520263673, 0.018387935638427735, 0.01847603225708008, 0.01841459274291992, 0.018441247940063476, 0.01850160026550293, 0.01901158332824707, 0.018525215148925783, 0.01865212821960449, 0.018457599639892578, 0.01840233612060547, 0.01844528007507324, 0.018463775634765624, 0.01849545669555664, 0.018371583938598633, 0.018448383331298827, 0.018485248565673826, 0.020388864517211915, 0.019298303604125978, 0.018562047958374024, 0.01846380805969238, 0.01849337577819824, 0.0184770565032959, 0.039277568817138675, 0.018464767456054687, 0.01844633674621582, 0.018480127334594726, 0.018456607818603515, 0.01846985626220703, 0.018428928375244142, 0.018431999206542968, 0.018909183502197266, 0.01845452880859375, 0.01840025520324707, 0.01847500801086426, 0.01843302345275879, 0.018447359085083007, 0.01845248031616211, 0.018421760559082033, 0.01845248031616211, 0.01845248031616211, 0.01841049575805664, 0.018455551147460936, 0.018391103744506837, 0.018453439712524413, 0.018481151580810547, 0.01843507194519043, 0.01846272087097168, 0.018471935272216796, 0.01852012825012207, 0.018537408828735353, 0.01840332794189453, 0.01842995262145996, 0.01842585563659668, 0.018507776260375978, 0.018486272811889647, 0.018449407577514648, 0.018455551147460936, 0.018579456329345705, 0.01843507194519043, 0.01846067237854004, 0.01843507194519043, 0.018449407577514648, 0.018431999206542968, 0.018437120437622072, 0.01843609619140625, 0.018428928375244142, 0.018663423538208008, 0.018478080749511717, 0.018453504562377928, 0.018449407577514648, 0.018479103088378905, 0.01848320007324219, 0.018426944732666015, 0.01841862487792969, 0.0184586238861084, 0.018548736572265623, 0.01842483139038086, 0.018523136138916017, 0.01847500801086426, 0.018473983764648438, 0.018465887069702147, 0.018414495468139648, 0.018473983764648438, 0.01845452880859375, 0.018778112411499022, 0.03877996826171875, 0.01850054359436035, 0.018472959518432617, 0.01847091293334961, 0.01841049575805664, 0.018469951629638673, 0.018467775344848635, 0.018440191268920898, 0.01843404769897461, 0.01843404769897461, 0.01859993553161621, 0.01864499282836914, 0.01846886444091797, 0.01906380844116211, 0.01860095977783203, 0.0184770565032959, 0.018586624145507814, 0.018533376693725585, 0.01843302345275879, 0.01843814468383789, 0.0182794246673584, 0.01837065505981445, 0.018466720581054686, 0.01844326400756836, 0.018535423278808593, 0.018521087646484375, 0.01843097686767578, 0.018496511459350586, 0.0184770565032959, 0.018731008529663085, 0.018544639587402344, 0.018488319396972656, 0.018480127334594726, 0.018737152099609376, 0.018786304473876952, 0.01843507194519043, 0.01843916893005371, 0.018413568496704103, 0.018478080749511717, 0.01840127944946289, 0.01842995262145996, 0.01839923286437988, 0.01841263961791992, 0.018468767166137694, 0.018487295150756835, 0.018437120437622072, 0.01849241638183594, 0.01841766357421875, 0.01844223976135254, 0.018423807144165038, 0.018428928375244142, 0.01841971206665039, 0.01887948799133301, 0.01847603225708008, 0.018481151580810547, 0.018397184371948243, 0.018431999206542968, 0.01839411163330078, 0.01839820861816406, 0.01843302345275879, 0.01843814468383789, 0.018465791702270508, 0.0184116153717041, 0.03870198440551758, 0.018505727767944336, 0.018463743209838866, 0.01842483139038086, 0.01846067237854004, 0.01840742492675781, 0.018502656936645507, 0.018465791702270508, 0.018471935272216796, 0.018499584197998048, 0.018421760559082033, 0.018455551147460936, 0.018415615081787108, 0.01843507194519043, 0.01841971206665039, 0.01850060844421387, 0.018417728424072265, 0.018436031341552736, 0.0184453125, 0.018503679275512695, 0.01841663932800293, 0.01841971206665039, 0.018510847091674804, 0.018480127334594726, 0.01864396858215332, 0.018501632690429686, 0.01845043182373047, 0.018503679275512695, 0.018226175308227538, 0.018330623626708984, 0.01848320007324219, 0.01846067237854004, 0.018522111892700196, 0.018927616119384767, 0.018724863052368163, 0.018456575393676757, 0.01838591957092285, 0.018397216796875, 0.01844220733642578, 0.01840025520324707, 0.018312192916870116, 0.018486272811889647, 0.018696256637573242, 0.01859987258911133, 0.018705440521240235, 0.018421728134155272, 0.018325504302978517, 0.01825382423400879, 0.018249727249145507, 0.018318336486816408, 0.01820467185974121, 0.018321407318115233, 0.018272256851196288, 0.01844428825378418, 0.018939903259277344, 0.018840576171875, 0.018480127334594726, 0.018486272811889647, 0.018461696624755858, 0.018874368667602538, 0.018914304733276367, 0.01856716728210449, 0.018502656936645507, 0.03866316986083984, 0.018490367889404297, 0.018498559951782227, 0.01862246322631836, 0.018948095321655273, 0.01875663948059082, 0.018512863159179688, 0.018510847091674804, 0.01840640068054199, 0.018463743209838866, 0.018428928375244142, 0.018463743209838866, 0.018885696411132812, 0.018785215377807616, 0.0184453125, 0.018440191268920898, 0.01844326400756836, 0.01843916893005371, 0.018464767456054687, 0.01841049575805664, 0.018899967193603515, 0.01897881507873535, 0.018525184631347655, 0.01840640068054199, 0.018463743209838866, 0.01841663932800293, 0.018395135879516602, 0.01843404769897461, 0.018739200592041014, 0.01866854476928711, 0.018449407577514648, 0.018440191268920898, 0.01884671974182129, 0.01884671974182129, 0.0186296329498291, 0.01845452880859375, 0.018465791702270508, 0.01845145606994629, 0.018478080749511717, 0.018554880142211915, 0.018903039932250978, 0.018525184631347655, 0.018480127334594726, 0.018397184371948243, 0.018494464874267577, 0.0188242244720459, 0.018550752639770508, 0.018479103088378905, 0.018447359085083007, 0.018521087646484375, 0.018530303955078126, 0.018420736312866212, 0.01846272087097168, 0.018373632431030275, 0.018495487213134765, 0.018415615081787108, 0.018606111526489257, 0.018444255828857423, 0.018486272811889647, 0.018553855895996094, 0.01847500801086426, 0.01865011215209961, 0.018542591094970702, 0.039008255004882815, 0.01843916893005371, 0.018572320938110353, 0.018645984649658203, 0.018721792221069337, 0.018486272811889647, 0.018408512115478514, 0.018450368881225587, 0.01859993553161621, 0.018506752014160157, 0.01842995262145996, 0.01848320007324219, 0.018528255462646484, 0.01882931137084961, 0.018415615081787108, 0.018464767456054687, 0.01842995262145996, 0.018480127334594726, 0.018689023971557618, 0.01867366409301758, 0.01841766357421875, 0.018509824752807616, 0.01886207962036133, 0.018561023712158203, 0.01927065658569336, 0.018735103607177735, 0.018537471771240235, 0.018486272811889647, 0.018457599639892578, 0.018547712326049806, 0.01882316780090332, 0.018619392395019533, 0.018562047958374024, 0.01861529541015625, 0.018619392395019533, 0.0187064323425293, 0.018515968322753908, 0.018487295150756835, 0.018381824493408205, 0.01860403251647949, 0.018646015167236327, 0.018466880798339844, 0.018506687164306642, 0.01840947151184082, 0.018453504562377928, 0.01860915184020996, 0.01865216064453125, 0.018463743209838866, 0.018577407836914063, 0.01884262466430664, 0.018671615600585938, 0.018929664611816405, 0.018790399551391602, 0.01878118324279785, 0.018528255462646484, 0.018726911544799805, 0.018951168060302736, 0.018900991439819336, 0.01901875114440918, 0.018754560470581053, 0.018983936309814452, 0.018982912063598634, 0.01886617660522461]",tokens/s,52.867723371954284,,,main,False,False,,,,,,,,,,,,,,,,,,,,,,, float16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,.,.,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 373, in cached_file raise EnvironmentError( OSError: . does not appear to have a file named config.json. Checkout 'https://huggingface.co/./tree/None' for available files. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,a,a,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/a/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6694748b-31bbd64524e54ab57045f92d;75ceb8d1-167e-4921-bef0-545c959d72a9) Repository Not Found for url: https://huggingface.co/a/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: a is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-7B,Qwen/Qwen-7B,cuda,0,42,,,True,,,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",qwen,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-66b,facebook/opt-66b,cuda,0,42,,,True,,,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",opt,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,stabilityai/stablelm-base-alpha-7b,stabilityai/stablelm-base-alpha-7b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.42.3,,0.31.0,,,,1.20.0,,,,0.11.1,,,,MB,1275.428864,17176.199168,0.0,16529.752064,16322.781696,s,10,0.7599501113891602,0.07599501113891602,0.00033945009770586053,0.07586393356323243,0.07629016418457031,0.0766181541442871,0.07688054611206055,"[0.07694614410400391, 0.07587407684326172, 0.07571731567382813, 0.07584198760986328, 0.07585379028320313, 0.07583280181884766, 0.07591193389892578, 0.07591075134277343, 0.07584403228759766, 0.07621727752685546]",tokens/s,3368.6421801036604,kWh,8.959694041146174e-07,4.9095040940262e-07,5.400141104621145e-06,6.787060918138382e-06,tokens/kWh,37718830.4463043,MB,1275.428864,17176.199168,0.0,16529.752064,16442.131456,s,10,22.000514892578124,2.2000514892578122,0.0007280570989062907,2.19981005859375,2.2006001220703126,2.2012802856445313,2.2018244165039063,"[2.20196044921875, 2.20020849609375, 2.19935595703125, 2.199746826171875, 2.20027099609375, 2.199352294921875, 2.19960302734375, 2.199694580078125, 2.199873291015625, 2.200448974609375]",tokens/s,28.635693440635364,kWh,2.5978705237309144e-05,1.423470402100106e-05,0.0001313500516287797,0.0001715634608870899,tokens/kWh,367211.05807874695,,s,629,22.648429618835433,0.03600704231929325,0.009026360330505433,0.03491020965576172,0.035014860534667964,0.03503636627197266,0.11084595184326172,"[0.034800640106201174, 0.03482419204711914, 0.03482726287841797, 0.0348221435546875, 0.034802688598632815, 0.03482726287841797, 0.034819072723388675, 0.03482931137084961, 0.03479859161376953, 0.03483852767944336, 0.03500851058959961, 0.035297279357910154, 0.03489382553100586, 0.0349306869506836, 0.0351539192199707, 0.0350300178527832, 0.034871295928955076, 0.0348590087890625, 0.034840576171875, 0.03485696029663086, 0.03486105728149414, 0.03484467315673828, 0.03488870239257812, 0.03488665771484375, 0.034887680053710936, 0.0349224967956543, 0.034939903259277344, 0.034907135009765625, 0.03487948989868164, 0.034904064178466795, 0.03489996719360351, 0.03493273544311523, 0.034890750885009765, 0.03490816116333008, 0.03490611267089844, 0.03540172958374024, 0.03497574234008789, 0.034947071075439456, 0.034955265045166016, 0.03495116806030273, 0.0349378547668457, 0.034956287384033204, 0.0350382080078125, 0.03496857452392578, 0.034971649169921876, 0.03497062301635742, 0.034969600677490234, 0.03499008178710938, 0.034985984802246094, 0.03497574234008789, 0.03496448135375976, 0.03503308868408203, 0.0350115852355957, 0.03500339126586914, 0.03499724960327148, 0.035009536743164066, 0.03508019256591797, 0.03500543975830078, 0.034994174957275394, 0.035023872375488284, 0.035046398162841795, 0.035026943206787106, 0.11091251373291015, 0.03481190490722656, 0.03490611267089844, 0.03481702423095703, 0.03481497573852539, 0.03482112121582031, 0.034830337524414064, 0.034835456848144535, 0.034841598510742186, 0.03482624053955078, 0.03484774398803711, 0.03483750534057617, 0.034857982635498046, 0.03486310577392578, 0.03483955383300781, 0.03488358306884766, 0.0348487663269043, 0.03486310577392578, 0.03484672164916992, 0.03486412811279297, 0.03485388946533203, 0.034872318267822264, 0.034874366760253905, 0.03486003112792969, 0.03484979248046875, 0.03488051223754883, 0.034895870208740236, 0.034887680053710936, 0.03489382553100586, 0.03488665771484375, 0.03488972854614258, 0.034890750885009765, 0.034907135009765625, 0.03493580627441406, 0.03492659378051758, 0.03492454528808594, 0.03490611267089844, 0.03494400024414063, 0.034938880920410156, 0.034938880920410156, 0.03493580627441406, 0.03497574234008789, 0.0349224967956543, 0.03494297790527344, 0.03494400024414063, 0.03498188781738281, 0.034953216552734374, 0.035310592651367184, 0.03497369766235352, 0.034994174957275394, 0.03495116806030273, 0.03499724960327148, 0.03497983932495117, 0.035007488250732424, 0.034985984802246094, 0.034988033294677735, 0.03498086547851562, 0.03514572906494141, 0.035019775390625, 0.03499212646484375, 0.03499929428100586, 0.035171329498291014, 0.03503411102294922, 0.11072716522216797, 0.03479654312133789, 0.03482624053955078, 0.03482316970825195, 0.034830337524414064, 0.03483750534057617, 0.03482316970825195, 0.03481497573852539, 0.03482316970825195, 0.034830337524414064, 0.03482931137084961, 0.03484467315673828, 0.03488665771484375, 0.03485388946533203, 0.03483852767944336, 0.0348671989440918, 0.03484672164916992, 0.03483238220214844, 0.034852863311767575, 0.0348671989440918, 0.034872318267822264, 0.034840576171875, 0.03485696029663086, 0.034872318267822264, 0.034890750885009765, 0.03486310577392578, 0.0349224967956543, 0.03489382553100586, 0.034885631561279294, 0.03487744140625, 0.034909183502197266, 0.034900993347167966, 0.034900993347167966, 0.03490611267089844, 0.03490816116333008, 0.03490508651733398, 0.03492659378051758, 0.03494604873657227, 0.03494911956787109, 0.03491839981079101, 0.03492966461181641, 0.03495116806030273, 0.034928638458251955, 0.0349378547668457, 0.03497267150878906, 0.03497267150878906, 0.034948097229003904, 0.034956287384033204, 0.03496550369262695, 0.03495116806030273, 0.03498291015625, 0.03496448135375976, 0.03498495864868164, 0.03499622344970703, 0.034993152618408206, 0.034991104125976565, 0.03502489471435547, 0.03500543975830078, 0.03500851058959961, 0.03499724960327148, 0.03501772689819336, 0.03500236892700195, 0.03502284622192383, 0.11085311889648437, 0.03481190490722656, 0.03480985641479492, 0.03482624053955078, 0.03480166244506836, 0.03483852767944336, 0.03482112121582031, 0.03481087875366211, 0.03484774398803711, 0.03483852767944336, 0.034840576171875, 0.03483647918701172, 0.03484262466430664, 0.03484569549560547, 0.034854911804199216, 0.034874366760253905, 0.03484774398803711, 0.0348487663269043, 0.03484569549560547, 0.034871295928955076, 0.03484262466430664, 0.034871295928955076, 0.034874366760253905, 0.03486105728149414, 0.03488153457641602, 0.034857982635498046, 0.034885631561279294, 0.03488972854614258, 0.034885631561279294, 0.034900993347167966, 0.03487334442138672, 0.03489382553100586, 0.03491635131835937, 0.03491430282592774, 0.03491123199462891, 0.03492147064208984, 0.03488972854614258, 0.03492966461181641, 0.034955265045166016, 0.034958335876464845, 0.03509964752197266, 0.03496038436889649, 0.03493580627441406, 0.034958335876464845, 0.034950145721435545, 0.03497369766235352, 0.034947071075439456, 0.034945022583007815, 0.03499212646484375, 0.034988033294677735, 0.034956287384033204, 0.03497574234008789, 0.03497881698608398, 0.0350013427734375, 0.034988033294677735, 0.03500339126586914, 0.035163135528564454, 0.035026943206787106, 0.03504844665527344, 0.034994174957275394, 0.03502592086791992, 0.035019775390625, 0.03500646209716797, 0.11086643218994141, 0.03480473709106445, 0.034813953399658204, 0.034816001892089846, 0.03483852767944336, 0.03484467315673828, 0.0348221435546875, 0.03483135986328125, 0.034841598510742186, 0.03484672164916992, 0.0348487663269043, 0.03488665771484375, 0.03486412811279297, 0.03485081481933594, 0.03485184097290039, 0.03486617660522461, 0.0348671989440918, 0.03488460922241211, 0.034855934143066404, 0.03485388946533203, 0.03485388946533203, 0.034974720001220705, 0.03506380844116211, 0.03487027359008789, 0.03486515045166016, 0.03487948989868164, 0.03488256072998047, 0.03487948989868164, 0.034902015686035154, 0.03489894485473633, 0.03487334442138672, 0.03489382553100586, 0.03489791870117188, 0.03492761611938477, 0.03489996719360351, 0.034904064178466795, 0.03489894485473633, 0.0349194221496582, 0.034977790832519534, 0.034956287384033204, 0.03495219039916992, 0.03501875305175781, 0.03506995010375977, 0.03497369766235352, 0.03498495864868164, 0.03497062301635742, 0.0350115852355957, 0.03496345520019531, 0.03496755218505859, 0.034969600677490234, 0.03498393630981445, 0.034994174957275394, 0.03500032043457031, 0.035009536743164066, 0.03498393630981445, 0.03499622344970703, 0.03501465606689453, 0.03503513717651367, 0.03499827194213867, 0.03500339126586914, 0.03501670455932617, 0.035037185668945314, 0.03503308868408203, 0.11082752227783203, 0.03482419204711914, 0.03477811050415039, 0.03483340835571289, 0.03482726287841797, 0.03482624053955078, 0.03482624053955078, 0.03482316970825195, 0.034830337524414064, 0.03484569549560547, 0.03482828903198242, 0.034830337524414064, 0.034852863311767575, 0.03488153457641602, 0.03485184097290039, 0.034855934143066404, 0.03483647918701172, 0.034857982635498046, 0.03485388946533203, 0.03486822509765625, 0.034840576171875, 0.03488153457641602, 0.03486310577392578, 0.03486412811279297, 0.03483750534057617, 0.034950145721435545, 0.034885631561279294, 0.03488256072998047, 0.03488665771484375, 0.03490816116333008, 0.034900993347167966, 0.034895870208740236, 0.03490508651733398, 0.034920448303222655, 0.034909183502197266, 0.034895870208740236, 0.03490304183959961, 0.034923519134521484, 0.03490611267089844, 0.034938880920410156, 0.03495935821533203, 0.034936832427978515, 0.0349409294128418, 0.03495116806030273, 0.034947071075439456, 0.03496755218505859, 0.03496448135375976, 0.034956287384033204, 0.03495219039916992, 0.034976768493652347, 0.03498188781738281, 0.03498905563354492, 0.034976768493652347, 0.03496857452392578, 0.03501465606689453, 0.034994174957275394, 0.03499008178710938, 0.03499520111083984, 0.035004417419433595, 0.03500851058959961, 0.0350115852355957, 0.035046398162841795, 0.03501875305175781, 0.11091149139404297, 0.03480575942993164, 0.03479347229003906, 0.03480476760864258, 0.03480368041992187, 0.03482624053955078, 0.03480985641479492, 0.034825214385986326, 0.03504742431640625, 0.03487744140625, 0.03481190490722656, 0.03484467315673828, 0.03483647918701172, 0.034854911804199216, 0.03485388946533203, 0.034887680053710936, 0.03486003112792969, 0.034841598510742186, 0.03484672164916992, 0.03486105728149414, 0.034852863311767575, 0.0348590087890625, 0.034857982635498046, 0.034869247436523435, 0.03486310577392578, 0.034917377471923826, 0.03488153457641602, 0.03489382553100586, 0.03487846374511719, 0.034907135009765625, 0.034895870208740236, 0.034907135009765625, 0.03489177703857422, 0.03491328048706055, 0.03493580627441406, 0.034931713104248044, 0.034907135009765625, 0.0349306869506836, 0.034923519134521484, 0.034933761596679686, 0.0349306869506836, 0.03492659378051758, 0.034923519134521484, 0.03494297790527344, 0.03495219039916992, 0.03498393630981445, 0.034993152618408206, 0.03498291015625, 0.034976768493652347, 0.03496448135375976, 0.03494400024414063, 0.034993152618408206, 0.03499929428100586, 0.035009536743164066, 0.03497369766235352, 0.03503104019165039, 0.034988033294677735, 0.03500032043457031, 0.03500339126586914, 0.03501567840576172, 0.03501055908203125, 0.034994174957275394, 0.03503923034667969, 0.1109381103515625, 0.03480780792236328, 0.03480678558349609, 0.034800640106201174, 0.03480166244506836, 0.03484774398803711, 0.034819072723388675, 0.0348221435546875, 0.03481804656982422, 0.03485696029663086, 0.034835456848144535, 0.034855934143066404, 0.034830337524414064, 0.0348590087890625, 0.03484467315673828, 0.03487846374511719, 0.03483750534057617, 0.034857982635498046, 0.034885631561279294, 0.034874366760253905, 0.03485184097290039, 0.034885631561279294, 0.035007488250732424, 0.03485388946533203, 0.03486822509765625, 0.03488460922241211, 0.03487027359008789, 0.03491839981079101, 0.034907135009765625, 0.034874366760253905, 0.03490611267089844, 0.034909183502197266, 0.03487641525268555, 0.03490611267089844, 0.03491123199462891, 0.03491839981079101, 0.03490508651733398, 0.03492966461181641, 0.03506073760986328, 0.03497881698608398, 0.034931713104248044, 0.03494297790527344, 0.034947071075439456, 0.034953216552734374, 0.0349378547668457, 0.03501567840576172, 0.03497062301635742, 0.034945022583007815, 0.03496448135375976, 0.03497881698608398, 0.034955265045166016, 0.03498905563354492, 0.03497983932495117, 0.03500032043457031, 0.03496755218505859, 0.035019775390625, 0.03499724960327148, 0.03500851058959961, 0.03498291015625, 0.0350115852355957, 0.0350300178527832, 0.03502592086791992, 0.03499929428100586, 0.11090943908691406, 0.0347955207824707, 0.03482726287841797, 0.034816001892089846, 0.034816001892089846, 0.03483340835571289, 0.034819072723388675, 0.03481497573852539, 0.03482009506225586, 0.034920448303222655, 0.03485388946533203, 0.03483750534057617, 0.03483955383300781, 0.03487948989868164, 0.034840576171875, 0.034854911804199216, 0.03492966461181641, 0.03489689636230469, 0.0348487663269043, 0.03484467315673828, 0.03485696029663086, 0.03486105728149414, 0.03484979248046875, 0.034855934143066404, 0.03487334442138672, 0.03488358306884766, 0.03487846374511719, 0.03491020965576172, 0.034887680053710936, 0.03486822509765625, 0.03489996719360351, 0.03487846374511719, 0.03488870239257812, 0.0349306869506836, 0.03495423889160156, 0.03495116806030273, 0.034912254333496096, 0.03498700714111328, 0.034938880920410156, 0.03493580627441406, 0.034925567626953126, 0.034931713104248044, 0.03494400024414063, 0.03494400024414063, 0.034941951751708986, 0.034961406707763674, 0.03494400024414063, 0.03505254364013672, 0.03501875305175781, 0.03498700714111328, 0.03497369766235352, 0.03498905563354492, 0.034991104125976565, 0.03499212646484375, 0.03498188781738281, 0.03499827194213867, 0.0350013427734375, 0.035004417419433595, 0.035007488250732424, 0.03501670455932617, 0.03501055908203125, 0.03501670455932617, 0.035020801544189455, 0.11135590362548828, 0.034920448303222655, 0.0350013427734375, 0.03488153457641602, 0.03482112121582031, 0.03483135986328125, 0.03482316970825195, 0.03484467315673828, 0.03481087875366211, 0.03486207962036133, 0.03481292724609375, 0.03487539291381836, 0.034841598510742186, 0.03488460922241211, 0.03489279937744141, 0.03486207962036133, 0.03483955383300781, 0.03486412811279297, 0.034840576171875, 0.03486207962036133, 0.0348590087890625, 0.03489484786987305, 0.03489996719360351, 0.03486515045166016, 0.03487334442138672, 0.03489382553100586, 0.03488460922241211, 0.03489177703857422, 0.03491328048706055, 0.034923519134521484, 0.034900993347167966, 0.03489177703857422, 0.034928638458251955, 0.03492966461181641, 0.03490816116333008, 0.03491328048706055, 0.03491635131835937, 0.034941951751708986, 0.03492966461181641, 0.034961406707763674, 0.034958335876464845, 0.034953216552734374, 0.03494297790527344, 0.03498291015625, 0.03497369766235352, 0.03498188781738281, 0.03496243286132812, 0.03498393630981445, 0.03496345520019531, 0.034969600677490234, 0.03512319946289062, 0.0350300178527832, 0.03498905563354492, 0.03498291015625, 0.034988033294677735, 0.0350300178527832, 0.03499622344970703, 0.03500032043457031, 0.03501875305175781, 0.035043327331542966, 0.034991104125976565, 0.0350382080078125, 0.03503104019165039]",tokens/s,27.772344952203454,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,1,1,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/1/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66932609-0ba5d77764bfd258149b2e47;2aefc574-02b3-4f42-820d-4be40e365ad5) Repository Not Found for url: https://huggingface.co/1/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: 1 is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-7b,tiiuae/falcon-7b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ self.load_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 250, in load_model_with_no_weights self.pretrained_model = self.automodel_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 559, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3704, in from_pretrained config = cls._autoset_attn_implementation( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1490, in _autoset_attn_implementation config = cls._check_and_enable_sdpa( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1656, in _check_and_enable_sdpa raise ValueError( ValueError: FalconForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,meta-llama/Llama-2-7b-hf,meta-llama/Llama-2-7b-hf,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.1,,,,MB,1220.804608,14563.147776,0.0,13916.700672,13738.740736,s,10,0.9359868392944336,0.09359868392944336,0.0002275031437491741,0.09355839920043946,0.09371135940551757,0.09397617454528809,0.09418802665710449,"[0.09424098968505859, 0.09356479644775391, 0.093552001953125, 0.09348719787597656, 0.09365251159667969, 0.09336857604980468, 0.09358521270751953, 0.09344294738769532, 0.09350450897216797, 0.09358809661865235]",tokens/s,2735.08119187849,kWh,1.1069706744617887e-06,6.065673582729388e-07,6.1301970121495445e-06,7.843735044884272e-06,tokens/kWh,32637512.42680802,MB,1220.804608,14563.147776,0.0,13916.700672,13889.779712,s,10,22.558502685546873,2.2558502685546875,0.0009753264580620002,2.25562646484375,2.2571563720703125,2.2574865112304687,2.2577506225585937,"[2.255815673828125, 2.25638037109375, 2.254172119140625, 2.255611083984375, 2.255641845703125, 2.257816650390625, 2.25538134765625, 2.2570830078125, 2.255073974609375, 2.255526611328125]",tokens/s,27.927385464445653,kWh,2.6576527158419297e-05,1.4564238501452137e-05,0.00011811940238744975,0.00015926016804732117,tokens/kWh,395579.13803833694,,s,629,23.363786796569805,0.03714433512968176,0.011185130091063316,0.035789825439453124,0.035883416748046876,0.035952844238281254,0.12990066833496094,"[0.035757057189941405, 0.03570687866210937, 0.03571200180053711, 0.03569356918334961, 0.03569356918334961, 0.03574272155761719, 0.0357130241394043, 0.03570995330810547, 0.0357386245727539, 0.035724288940429685, 0.035729408264160156, 0.03568646240234375, 0.03574163055419922, 0.035901439666748046, 0.03585945510864258, 0.035781631469726564, 0.03575913619995117, 0.03576009750366211, 0.03576934432983398, 0.03573766326904297, 0.03574470520019531, 0.035730430603027344, 0.03575193786621094, 0.0357386245727539, 0.03576115036010742, 0.0357498893737793, 0.03577241516113281, 0.03578572845458984, 0.03581340789794922, 0.03576316833496094, 0.03582361602783203, 0.03576422500610352, 0.03580108642578125, 0.035789825439453124, 0.03581235122680664, 0.035784736633300784, 0.03580617523193359, 0.03578675079345703, 0.035789825439453124, 0.03577139282226562, 0.035814399719238284, 0.03578060913085938, 0.035827713012695314, 0.0358021125793457, 0.035814399719238284, 0.03583180618286133, 0.03583692932128906, 0.03582156753540039, 0.03584204864501953, 0.035799072265625, 0.035862495422363284, 0.03582566452026367, 0.03585433578491211, 0.03582668685913086, 0.03588915252685547, 0.0359659538269043, 0.0359024658203125, 0.03585433578491211, 0.03587071990966797, 0.03585433578491211, 0.03586764907836914, 0.03583488082885742, 0.12998042297363283, 0.03574169540405273, 0.03569561767578125, 0.03573555374145508, 0.035746814727783204, 0.03574275207519531, 0.035724254608154295, 0.03573555374145508, 0.035716094970703126, 0.035730430603027344, 0.03573350524902344, 0.0357283821105957, 0.035740673065185545, 0.0357386245727539, 0.03573657608032227, 0.03573350524902344, 0.03574169540405273, 0.035716094970703126, 0.03574272155761719, 0.035776512145996094, 0.03575193786621094, 0.035760128021240234, 0.03576627349853516, 0.035745792388916016, 0.03574784088134766, 0.03580723190307617, 0.035767295837402346, 0.03578265762329102, 0.03577139282226562, 0.035783679962158206, 0.03577139282226562, 0.03579391860961914, 0.035778560638427735, 0.035806209564208984, 0.035768318176269534, 0.035814399719238284, 0.03594854354858398, 0.03617587280273438, 0.035980289459228515, 0.03594342422485351, 0.035871742248535156, 0.03581542587280274, 0.03582156753540039, 0.035829761505126956, 0.03583078384399414, 0.03585126495361328, 0.035885055541992186, 0.03587788772583008, 0.035855358123779296, 0.03585638427734375, 0.03582668685913086, 0.03582259368896484, 0.035850238800048825, 0.03586150360107422, 0.03583692932128906, 0.03586150360107422, 0.03584105682373047, 0.03586659240722656, 0.035860481262207033, 0.03587686538696289, 0.035844097137451174, 0.03586457443237305, 0.03586150360107422, 0.12993536376953124, 0.03565366363525391, 0.03567919921875, 0.035675136566162106, 0.03566284942626953, 0.035688449859619144, 0.0356864013671875, 0.035722240447998044, 0.035681278228759765, 0.03570687866210937, 0.035716094970703126, 0.03574784088134766, 0.035699710845947266, 0.0357283821105957, 0.035694591522216795, 0.03570687866210937, 0.0357130241394043, 0.03575603103637695, 0.035862529754638675, 0.03582566452026367, 0.035724288940429685, 0.0357386245727539, 0.035727359771728515, 0.03575296020507813, 0.03573555374145508, 0.03575398254394531, 0.03575398254394531, 0.03576319885253906, 0.03576627349853516, 0.03578265762329102, 0.03574476623535156, 0.03577139282226562, 0.035762176513671876, 0.03577139282226562, 0.035783679962158206, 0.035783679962158206, 0.03577241516113281, 0.035794944763183595, 0.035773441314697264, 0.03580108642578125, 0.03575500869750976, 0.03580825424194336, 0.03579904174804688, 0.03579391860961914, 0.03582463836669922, 0.03581849670410156, 0.03581542587280274, 0.03586969757080078, 0.03581542587280274, 0.0358287353515625, 0.03581542587280274, 0.03582463836669922, 0.03580416107177734, 0.03581951904296875, 0.03583488082885742, 0.03585945510864258, 0.03581951904296875, 0.035833854675292966, 0.03583692932128906, 0.03585228729248047, 0.03582566452026367, 0.03585843276977539, 0.03584921646118164, 0.1298851776123047, 0.03567103958129883, 0.035676158905029294, 0.03569868850708008, 0.03569664001464844, 0.035683326721191407, 0.03572019195556641, 0.035697662353515625, 0.03572019195556641, 0.035729408264160156, 0.035705856323242184, 0.03573350524902344, 0.035732479095458985, 0.035722240447998044, 0.0357130241394043, 0.03573350524902344, 0.035714046478271484, 0.03578879928588867, 0.03571916961669922, 0.03574272155761719, 0.03572019195556641, 0.035757057189941405, 0.035730430603027344, 0.035768318176269534, 0.035757057189941405, 0.035783679962158206, 0.03578777694702148, 0.035811328887939455, 0.03618406295776367, 0.03581951904296875, 0.035794944763183595, 0.03578060913085938, 0.03578265762329102, 0.0357498893737793, 0.035773441314697264, 0.03589427185058594, 0.036021247863769534, 0.035860481262207033, 0.035806209564208984, 0.03577446365356445, 0.03579289627075195, 0.035794944763183595, 0.035784702301025394, 0.03579596710205078, 0.03580313491821289, 0.035840030670166015, 0.035810272216796876, 0.03583590316772461, 0.03583180618286133, 0.035846145629882815, 0.035844097137451174, 0.03584102249145508, 0.03583795166015625, 0.03587071990966797, 0.03587891387939453, 0.035866622924804685, 0.03585433578491211, 0.035860481262207033, 0.03584921646118164, 0.03583078384399414, 0.03582566452026367, 0.03586969757080078, 0.035871742248535156, 0.13025587463378907, 0.035714046478271484, 0.0359741439819336, 0.03581849670410156, 0.035783679962158206, 0.035729408264160156, 0.03572019195556641, 0.0357130241394043, 0.035768318176269534, 0.03570380783081055, 0.035697662353515625, 0.0357130241394043, 0.0356864013671875, 0.035759105682373046, 0.03570073699951172, 0.035759105682373046, 0.035737598419189456, 0.03579084777832031, 0.0357498893737793, 0.03582156753540039, 0.03585331344604492, 0.035748863220214845, 0.035691520690917966, 0.03575398254394531, 0.03582668685913086, 0.03580928039550781, 0.035722240447998044, 0.035743743896484374, 0.03575603103637695, 0.035796993255615236, 0.03580313491821289, 0.0357918701171875, 0.03574272155761719, 0.03579084777832031, 0.03575193786621094, 0.035796993255615236, 0.035778560638427735, 0.03578265762329102, 0.03576115036010742, 0.03579084777832031, 0.035768318176269534, 0.03578060913085938, 0.035765247344970705, 0.03580416107177734, 0.03589017486572266, 0.035922943115234376, 0.035860481262207033, 0.03584000015258789, 0.03579289627075195, 0.03583488082885742, 0.03584000015258789, 0.03584511947631836, 0.03580825424194336, 0.035844097137451174, 0.03590348815917969, 0.035850238800048825, 0.03581542587280274, 0.03587481689453125, 0.035833854675292966, 0.035871742248535156, 0.03582463836669922, 0.03586150360107422, 0.03585638427734375, 0.12979814147949217, 0.03571916961669922, 0.035672065734863284, 0.03577040100097656, 0.0358614730834961, 0.03574784088134766, 0.03577139282226562, 0.035745792388916016, 0.035765247344970705, 0.03584000015258789, 0.035757057189941405, 0.03570483016967774, 0.03572127914428711, 0.03572934341430664, 0.03574169540405273, 0.03570175933837891, 0.03571814346313477, 0.03575603103637695, 0.03571507263183594, 0.03575807952880859, 0.035746814727783204, 0.035776512145996094, 0.035776512145996094, 0.035813377380371096, 0.035813377380371096, 0.03579391860961914, 0.035776512145996094, 0.035811328887939455, 0.035999744415283204, 0.03596492767333984, 0.03579391860961914, 0.03579084777832031, 0.035767295837402346, 0.03579391860961914, 0.03580928039550781, 0.03579904174804688, 0.035781631469726564, 0.03584716796875, 0.03585638427734375, 0.03586457443237305, 0.035879936218261715, 0.03583795166015625, 0.03589324951171875, 0.03589427185058594, 0.035885055541992186, 0.03584102249145508, 0.03584102249145508, 0.03587481689453125, 0.03589734268188476, 0.03596799850463867, 0.03591167831420899, 0.03584307098388672, 0.03590348815917969, 0.035862529754638675, 0.035945472717285154, 0.03609600067138672, 0.035896320343017575, 0.03596905517578125, 0.0358614730834961, 0.03597721481323242, 0.035844097137451174, 0.035931137084960936, 0.03596083068847656, 0.13005824279785155, 0.03567923355102539, 0.035683326721191407, 0.035675136566162106, 0.03569868850708008, 0.03569356918334961, 0.03568537521362305, 0.03570892715454101, 0.035705856323242184, 0.03575500869750976, 0.035814399719238284, 0.03571814346313477, 0.03571200180053711, 0.03573561477661133, 0.035684288024902346, 0.03572531127929687, 0.035737598419189456, 0.03572531127929687, 0.035714046478271484, 0.0357386245727539, 0.035705856323242184, 0.0357283821105957, 0.035729408264160156, 0.035768318176269534, 0.03574272155761719, 0.035765247344970705, 0.03575091171264649, 0.0357386245727539, 0.03576934432983398, 0.035746814727783204, 0.035746814727783204, 0.03577241516113281, 0.03575193786621094, 0.035776512145996094, 0.03587686538696289, 0.036395008087158204, 0.03588608169555664, 0.03579391860961914, 0.03578777694702148, 0.03580928039550781, 0.03578060913085938, 0.035794944763183595, 0.035783679962158206, 0.0357918701171875, 0.035796993255615236, 0.03579084777832031, 0.035810302734375, 0.03580313491821289, 0.03582259368896484, 0.0358021125793457, 0.03581542587280274, 0.035813377380371096, 0.03595775985717774, 0.035855358123779296, 0.03582672119140625, 0.03584201431274414, 0.035846145629882815, 0.03587891387939453, 0.03589734268188476, 0.03599359893798828, 0.03586457443237305, 0.03587071990966797, 0.035883007049560545, 0.12990669250488282, 0.03572326278686523, 0.03573555374145508, 0.035676158905029294, 0.03573350524902344, 0.03585433578491211, 0.03578675079345703, 0.03577446365356445, 0.03573964691162109, 0.0357498893737793, 0.03579289627075195, 0.03585945510864258, 0.03571916961669922, 0.035745792388916016, 0.03578675079345703, 0.03581849670410156, 0.03585331344604492, 0.03575193786621094, 0.035743743896484374, 0.0357498893737793, 0.03575398254394531, 0.035746814727783204, 0.035748863220214845, 0.035765247344970705, 0.03576934432983398, 0.03580928039550781, 0.03577753448486328, 0.03578265762329102, 0.035778560638427735, 0.035759105682373046, 0.035776512145996094, 0.03576422500610352, 0.03578777694702148, 0.035796993255615236, 0.0357918701171875, 0.03580108642578125, 0.035781631469726564, 0.03581951904296875, 0.03582668685913086, 0.03579596710205078, 0.03580518341064453, 0.03583180618286133, 0.03582156753540039, 0.035866622924804685, 0.03598441696166992, 0.03613180923461914, 0.035862529754638675, 0.035866622924804685, 0.03585433578491211, 0.03588608169555664, 0.035888126373291016, 0.03588915252685547, 0.03599155044555664, 0.036087806701660154, 0.0358809585571289, 0.03584511947631836, 0.03583283233642578, 0.0358737907409668, 0.03584511947631836, 0.03584819030761719, 0.03585126495361328, 0.03585331344604492, 0.03586150360107422, 0.1299251251220703, 0.03570483016967774, 0.03566592025756836, 0.03569359970092773, 0.035685344696044924, 0.03573452758789063, 0.03574169540405273, 0.035722240447998044, 0.03569356918334961, 0.03570073699951172, 0.03572019195556641, 0.03575603103637695, 0.03566592025756836, 0.035773441314697264, 0.035888126373291016, 0.03590758514404297, 0.035783679962158206, 0.035972095489501955, 0.03571916961669922, 0.03571814346313477, 0.03571200180053711, 0.0357314567565918, 0.03570892715454101, 0.035732479095458985, 0.03571814346313477, 0.035748863220214845, 0.035732479095458985, 0.03573657608032227, 0.03577446365356445, 0.035746814727783204, 0.035781631469726564, 0.03580825424194336, 0.0358287353515625, 0.03582361602783203, 0.035776512145996094, 0.0358021125793457, 0.03577241516113281, 0.035783679962158206, 0.035757057189941405, 0.035762176513671876, 0.035781631469726564, 0.03579596710205078, 0.03576422500610352, 0.035775489807128906, 0.03583283233642578, 0.03581951904296875, 0.03577958297729492, 0.03583488082885742, 0.03582262420654297, 0.03582870483398438, 0.0358205451965332, 0.035813377380371096, 0.03585948944091797, 0.03584815979003906, 0.035866622924804685, 0.035862529754638675, 0.035833854675292966, 0.03584511947631836, 0.035833854675292966, 0.0358389778137207, 0.03585433578491211, 0.035945472717285154, 0.035932159423828124, 0.13005619812011718, 0.03570175933837891, 0.03573555374145508, 0.03573350524902344, 0.03570380783081055, 0.03572326278686523, 0.0357283821105957, 0.035702816009521486, 0.03569865417480469, 0.03573555374145508, 0.03570278549194336, 0.03571200180053711, 0.035714046478271484, 0.035710975646972655, 0.03572531127929687, 0.03572531127929687, 0.03571200180053711, 0.035757057189941405, 0.0357212142944336, 0.03575500869750976, 0.035730464935302735, 0.03573859024047851, 0.03571507263183594, 0.03587891387939453, 0.035768318176269534, 0.035810302734375, 0.03584102249145508, 0.03576319885253906, 0.03575296020507813, 0.03577958297729492, 0.035760128021240234, 0.03578265762329102, 0.03575296020507813, 0.035789825439453124, 0.03578265762329102, 0.03581235122680664, 0.03580825424194336, 0.03582668685913086, 0.035794944763183595, 0.03581951904296875, 0.03577753448486328, 0.03578476715087891, 0.03584915161132812, 0.035955711364746096, 0.03580928039550781, 0.03581951904296875, 0.03580723190307617, 0.03586969757080078, 0.035811328887939455, 0.03584819030761719, 0.035922943115234376, 0.03597619247436523, 0.035846145629882815, 0.0358389778137207, 0.03583180618286133, 0.03587071990966797, 0.03580416107177734, 0.035846145629882815, 0.03584921646118164, 0.03587071990966797, 0.0358809585571289, 0.03589120101928711, 0.035934207916259765]",tokens/s,26.922005643894472,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-rw-1b,tiiuae/falcon-rw-1b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ self.load_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 250, in load_model_with_no_weights self.pretrained_model = self.automodel_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 559, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3704, in from_pretrained config = cls._autoset_attn_implementation( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1490, in _autoset_attn_implementation config = cls._check_and_enable_sdpa( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1656, in _check_and_enable_sdpa raise ValueError( ValueError: FalconForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,8,8,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/8/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66932439-0a15b8d819aa74f0227c326a;d06b5489-81f0-4d6a-a969-a3cfb2e8ce4c) Repository Not Found for url: https://huggingface.co/8/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: 8 is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-4B,,cuda,0,42,,,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,217063f5c507ed7cc255df7e1f64c4333a0b4dfe,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,,MB,1290.592256,10129.768448,0.0,9483.321344,8972.087296,s,10,0.46866477203369145,0.046866477203369145,0.0002547663708246897,0.04677824020385742,0.04692714157104492,0.04727507514953613,0.0475534220123291,"[0.04762300872802734, 0.046763809204101565, 0.04673324966430664, 0.04680486297607422, 0.046758113861083986, 0.04676835250854492, 0.046788127899169925, 0.04683481597900391, 0.046849822998046874, 0.04674060821533203]",tokens/s,5462.326491686827,kWh,5.548380871085246e-07,3.0397010657292366e-07,3.183625395098492e-06,4.04243358877994e-06,tokens/kWh,63328189.41306694,MB,1290.919936,10129.768448,0.0,9483.321344,9040.66816,s,10,20.485314697265622,2.048531469726562,0.006087315327469203,2.0474648437500003,2.0553745849609375,2.058837805175781,2.0616083813476562,"[2.062301025390625, 2.043177978515625, 2.048565673828125, 2.052588623046875, 2.0448070068359376, 2.0465029296875, 2.0421121826171875, 2.0484267578125, 2.05460498046875, 2.0422275390625]",tokens/s,30.753737948877706,kWh,2.417273671499876e-05,1.3247342717938193e-05,7.716763333030445e-05,0.0001145877127632414,tokens/kWh,549797.1683069476,,s,629,20.872635410308874,0.033183840079982256,0.005624682921060703,0.03241984176635742,0.032869989776611326,0.03340449752807617,0.07962308776855469,"[0.0333383674621582, 0.034283519744873044, 0.03558195114135742, 0.033296382904052735, 0.03338854217529297, 0.03346944046020508, 0.03356979370117188, 0.03341622543334961, 0.03341513442993164, 0.03335782241821289, 0.03353497695922852, 0.03325337600708008, 0.03318374252319336, 0.03321855926513672, 0.03407257461547852, 0.03328204727172852, 0.032985088348388675, 0.032299007415771484, 0.03222323226928711, 0.03246492767333985, 0.032371681213378904, 0.03242700958251953, 0.03224576187133789, 0.0323583984375, 0.03244851303100586, 0.03236556625366211, 0.032215038299560544, 0.03218124771118164, 0.03253350448608398, 0.0331960334777832, 0.03276902389526367, 0.03197542381286621, 0.03207680130004883, 0.03289395141601562, 0.03273830413818359, 0.032553985595703126, 0.03234815979003906, 0.032118785858154295, 0.032118785858154295, 0.03220588684082031, 0.03214227294921875, 0.03210649490356445, 0.03238092803955078, 0.032333824157714845, 0.032266239166259765, 0.03230310440063477, 0.03220275115966797, 0.032315391540527344, 0.03228057479858398, 0.03232972717285156, 0.03274956893920898, 0.032389118194580076, 0.03233078384399414, 0.03222524642944336, 0.03222323226928711, 0.032328704833984374, 0.03222220611572266, 0.03236556625366211, 0.032912384033203124, 0.03271987152099609, 0.032571392059326174, 0.032797695159912106, 0.07957810974121093, 0.03236761474609375, 0.032535552978515625, 0.032279552459716795, 0.032231422424316404, 0.03231027221679687, 0.032287742614746096, 0.03228876876831055, 0.032323585510253904, 0.03234406280517578, 0.03243724822998047, 0.03216588973999023, 0.032574462890625, 0.032345088958740234, 0.032146430969238284, 0.03233894348144531, 0.032497665405273435, 0.032385025024414066, 0.03246284866333008, 0.03274956893920898, 0.032568321228027344, 0.03286732864379883, 0.03219558334350586, 0.032535552978515625, 0.03282944107055664, 0.03243212890625, 0.03235635375976562, 0.03268915176391601, 0.0325294075012207, 0.03245881652832031, 0.03243001556396485, 0.03238604736328125, 0.032402431488037106, 0.03253247833251953, 0.03241779327392578, 0.03251200103759765, 0.03222528076171875, 0.03233484649658203, 0.032301055908203126, 0.03217203140258789, 0.032301055908203126, 0.03222630310058594, 0.03244646453857422, 0.03285404968261719, 0.03260412979125977, 0.032561183929443356, 0.03243004989624024, 0.03241676712036133, 0.032421886444091795, 0.03244543838500977, 0.03272294235229492, 0.0324505615234375, 0.03240857696533203, 0.03257958221435547, 0.032546817779541014, 0.032198654174804685, 0.032361473083496094, 0.03236249542236328, 0.03244339370727539, 0.03211673736572265, 0.03232767868041992, 0.03237376022338867, 0.03250483322143555, 0.07981772613525391, 0.0324659194946289, 0.03243724822998047, 0.03250175857543945, 0.03206655883789063, 0.03212492752075195, 0.03239219284057617, 0.03246182250976563, 0.03241062545776367, 0.03265945434570312, 0.03256012725830078, 0.0325591049194336, 0.03245363235473633, 0.03241273498535156, 0.032452545166015624, 0.032353279113769534, 0.03290521621704102, 0.032568321228027344, 0.03247513580322266, 0.03366502380371094, 0.03251302337646484, 0.032484352111816404, 0.032435199737548825, 0.03250688171386719, 0.032876544952392575, 0.033375232696533204, 0.032500736236572264, 0.03241984176635742, 0.03244441604614258, 0.03240140914916992, 0.03256524658203125, 0.03252838516235351, 0.032519168853759765, 0.032492542266845705, 0.03242086410522461, 0.03244441604614258, 0.032194561004638675, 0.03220172882080078, 0.03246080017089844, 0.03240959930419922, 0.03253145599365234, 0.03236044692993164, 0.03243315124511719, 0.032402431488037106, 0.032489471435546875, 0.0324136962890625, 0.032395263671875, 0.032570369720458986, 0.032502784729003906, 0.03244339370727539, 0.03213216018676758, 0.032384960174560544, 0.03204915237426758, 0.03245260620117187, 0.03225088119506836, 0.0321638412475586, 0.03306496047973633, 0.03240959930419922, 0.03258367919921875, 0.032727039337158204, 0.03227238464355469, 0.03225907135009766, 0.03400191879272461, 0.07999488067626953, 0.03244134521484375, 0.03239324951171875, 0.03209008026123047, 0.03279052734375, 0.03277414321899414, 0.03264614486694336, 0.03238604736328125, 0.03251302337646484, 0.03234099197387695, 0.032217086791992186, 0.03265945434570312, 0.03221401596069336, 0.03275571060180664, 0.03261747360229492, 0.034119678497314454, 0.03313356781005859, 0.033642494201660156, 0.03451289749145508, 0.03281305694580078, 0.03233792114257812, 0.03278950500488281, 0.03254988861083984, 0.03233792114257812, 0.0324956169128418, 0.03234201431274414, 0.03246080017089844, 0.03243622589111328, 0.03243008041381836, 0.032481281280517575, 0.032451583862304685, 0.03241984176635742, 0.03241164779663086, 0.0323881607055664, 0.03250579071044922, 0.03239833450317383, 0.032353279113769534, 0.03279974365234375, 0.03258265686035156, 0.032315391540527344, 0.03235737609863281, 0.03228160095214844, 0.03238604736328125, 0.0323768310546875, 0.03264102554321289, 0.032494590759277346, 0.03232972717285156, 0.03235123062133789, 0.03240652847290039, 0.03246182250976563, 0.03237785720825195, 0.032304126739501955, 0.03237068939208984, 0.033704959869384765, 0.033276927947998046, 0.03259801483154297, 0.032438270568847655, 0.03236761474609375, 0.03215769577026367, 0.03240652847290039, 0.032189441680908204, 0.03249049758911133, 0.03245363235473633, 0.07964057922363281, 0.03237580871582031, 0.032302078247070314, 0.03285094451904297, 0.0329543685913086, 0.03231948852539063, 0.03234918212890625, 0.031939584732055666, 0.032239681243896486, 0.032334785461425784, 0.032494590759277346, 0.03237068939208984, 0.03236454391479492, 0.031923200607299806, 0.03236556625366211, 0.03238092803955078, 0.032315391540527344, 0.03223244857788086, 0.03246080017089844, 0.032778240203857424, 0.03254476928710937, 0.032113662719726564, 0.03279564666748047, 0.03229695892333984, 0.03229695892333984, 0.03228876876831055, 0.03256934356689453, 0.03269836807250977, 0.032216064453125, 0.0323164176940918, 0.03225190353393555, 0.03275161743164062, 0.03244236755371094, 0.032271358489990236, 0.03243008041381836, 0.03289907073974609, 0.032258079528808596, 0.03233788681030274, 0.03240652847290039, 0.032314369201660156, 0.03226931381225586, 0.032470016479492186, 0.03240140914916992, 0.03245363235473633, 0.032421886444091795, 0.03240857696533203, 0.032382976531982424, 0.032301055908203126, 0.03239116668701172, 0.03237376022338867, 0.03295948791503906, 0.03229183959960937, 0.03225190353393555, 0.03247513580322266, 0.03239014434814453, 0.032333824157714845, 0.032314369201660156, 0.03355136108398438, 0.03369068908691406, 0.03253753662109375, 0.03248230361938476, 0.0326297607421875, 0.032330753326416016, 0.0797470703125, 0.03248537445068359, 0.03246899032592773, 0.03242905426025391, 0.03241062545776367, 0.03239219284057617, 0.03248230361938476, 0.03263180923461914, 0.03277312088012695, 0.032489471435546875, 0.03239424133300781, 0.0323583984375, 0.03248025512695313, 0.03244441604614258, 0.03250380706787109, 0.03249772644042969, 0.03344377517700195, 0.03268198394775391, 0.03237785720825195, 0.032247806549072264, 0.03227033615112305, 0.032396289825439455, 0.0324659194946289, 0.03232460784912109, 0.03241471862792969, 0.0323175048828125, 0.03242387390136719, 0.032366592407226565, 0.032292865753173826, 0.032271358489990236, 0.03244646453857422, 0.032732158660888674, 0.03283967971801758, 0.032378944396972656, 0.032987071990966794, 0.032277503967285154, 0.032282623291015625, 0.032285694122314454, 0.03260211181640625, 0.032524288177490236, 0.032484352111816404, 0.03236556625366211, 0.03241267013549805, 0.03242393493652344, 0.03242086410522461, 0.03237273788452148, 0.03253247833251953, 0.0324587516784668, 0.032314369201660156, 0.03229183959960937, 0.03315507125854492, 0.032524288177490236, 0.03235942459106445, 0.03236556625366211, 0.03246080017089844, 0.03240447998046875, 0.0323061752319336, 0.03219968032836914, 0.03254579162597656, 0.032317440032958986, 0.0323133430480957, 0.032966655731201173, 0.03265331268310547, 0.07973375701904296, 0.03226521682739258, 0.032421886444091795, 0.03232460784912109, 0.032492542266845705, 0.03257241439819336, 0.03259699249267578, 0.032538623809814454, 0.03243110275268555, 0.03233280181884766, 0.032464897155761716, 0.03234611129760742, 0.032347137451171876, 0.03274137496948242, 0.032530433654785154, 0.032395263671875, 0.0323389778137207, 0.032132064819335934, 0.032097278594970705, 0.03203481674194336, 0.03239321517944336, 0.032519168853759765, 0.03250688171386719, 0.03260211181640625, 0.032421886444091795, 0.03236454391479492, 0.03242393493652344, 0.03229388809204101, 0.03232767868041992, 0.03241471862792969, 0.032605247497558595, 0.03264710235595703, 0.03248025512695313, 0.03245267105102539, 0.03249663925170899, 0.0325098876953125, 0.03278643035888672, 0.032625663757324216, 0.03241164779663086, 0.0324136962890625, 0.03248230361938476, 0.03238604736328125, 0.03234611129760742, 0.032295936584472655, 0.03213312149047851, 0.03230003356933594, 0.032292865753173826, 0.03231846237182617, 0.03230003356933594, 0.03232460784912109, 0.032524288177490236, 0.032368640899658206, 0.032481281280517575, 0.032379905700683595, 0.032396289825439455, 0.032328704833984374, 0.032279552459716795, 0.03234304046630859, 0.03244236755371094, 0.03231235122680664, 0.03232969665527344, 0.032331775665283204, 0.03241471862792969, 0.07955865478515625, 0.03236761474609375, 0.03266867065429688, 0.0323583984375, 0.032376895904541014, 0.03229996871948242, 0.03242393493652344, 0.03236556625366211, 0.03232972717285156, 0.0322979850769043, 0.032683006286621095, 0.03241779327392578, 0.032522239685058595, 0.032543743133544925, 0.03240345764160156, 0.032350208282470705, 0.03241471862792969, 0.032314369201660156, 0.032460830688476563, 0.03237171173095703, 0.03238089752197266, 0.03239321517944336, 0.03242803192138672, 0.032363521575927735, 0.03244236755371094, 0.03234918212890625, 0.03236454391479492, 0.03241779327392578, 0.032323585510253904, 0.032553985595703126, 0.03265740966796875, 0.032658432006835936, 0.032435199737548825, 0.032655361175537106, 0.032494590759277346, 0.03232563018798828, 0.032399360656738284, 0.032402431488037106, 0.03240447998046875, 0.03237580871582031, 0.032500736236572264, 0.03239219284057617, 0.03239321517944336, 0.03255807876586914, 0.03234611129760742, 0.03236044692993164, 0.03243724822998047, 0.03324006271362305, 0.03453849411010742, 0.03263692855834961, 0.03255295944213867, 0.03234918212890625, 0.03222016143798828, 0.032023551940917966, 0.03465625762939453, 0.03306905746459961, 0.032486400604248046, 0.03241984176635742, 0.032444480895996095, 0.032419776916503905, 0.03217510223388672, 0.03234099197387695, 0.03238092803955078, 0.08013005065917969, 0.03303424072265625, 0.03433062362670898, 0.03309875106811523, 0.032868350982666016, 0.03253561782836914, 0.032457664489746095, 0.03244748687744141, 0.03246182250976563, 0.032314369201660156, 0.03236454391479492, 0.032555007934570314, 0.03235123062133789, 0.03248025512695313, 0.0324587516784668, 0.03236044692993164, 0.03232460784912109, 0.032244800567626956, 0.03242387390136719, 0.03237785720825195, 0.0323175048828125, 0.032392127990722656, 0.032379905700683595, 0.0323133430480957, 0.03250688171386719, 0.03240857696533203, 0.032389118194580076, 0.03232563018798828, 0.03235635375976562, 0.03233484649658203, 0.03251609420776367, 0.03245977783203125, 0.03241984176635742, 0.03240959930419922, 0.03250790405273438, 0.03239321517944336, 0.032497665405273435, 0.03232665634155273, 0.0324659194946289, 0.03449043273925781, 0.034412479400634764, 0.033173503875732424, 0.032702465057373044, 0.03244851303100586, 0.03242496109008789, 0.03256217575073242, 0.03251507186889648, 0.03254278564453125, 0.032806846618652345, 0.032551937103271485, 0.032573440551757815, 0.03265740966796875, 0.03246697616577148, 0.033802207946777345, 0.03256012725830078, 0.03237068939208984, 0.03239219284057617, 0.03239424133300781, 0.032456703186035156, 0.032499713897705076, 0.03240345764160156, 0.03249049758911133, 0.03242291259765625, 0.07967948913574219, 0.03235123062133789, 0.03273830413818359, 0.03244748687744141, 0.03233894348144531, 0.03239116668701172, 0.032729087829589845, 0.032514049530029294, 0.03234614562988281, 0.03282838439941406, 0.0326195182800293, 0.032470016479492186, 0.03230003356933594, 0.03234099197387695, 0.032502784729003906, 0.032301055908203126, 0.032363521575927735, 0.032331775665283204, 0.03241267013549805, 0.03239014434814453, 0.0323133430480957, 0.03244134521484375, 0.03239731216430664, 0.03237580871582031, 0.03240652847290039, 0.03243417739868164, 0.03298918533325195, 0.03274444961547852, 0.032405502319335935, 0.03236556625366211, 0.03244857788085938, 0.03231123352050781, 0.032309249877929686, 0.03239424133300781, 0.03243622589111328, 0.03234406280517578, 0.03242496109008789, 0.03229699325561523, 0.03242697525024414, 0.03231129455566406, 0.032399360656738284, 0.03197337532043457, 0.03244646453857422, 0.03317862319946289, 0.03269222259521484, 0.03249663925170899, 0.03245772933959961, 0.0320338249206543, 0.032499679565429686, 0.03235430526733398, 0.0324587516784668, 0.032314369201660156, 0.032097278594970705, 0.03225088119506836, 0.03213516616821289, 0.03201638412475586, 0.03209011077880859, 0.03218227386474609, 0.03256320190429687, 0.03243110275268555, 0.03243212890625, 0.032129024505615236, 0.03241984176635742]",tokens/s,30.135150048629775,,,main,False,False,,,,,,,,,,,,,,,,,,,,,,, float16-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,internlm/internlm2-20b,internlm/internlm2-20b,cuda,0,42,,,True,,,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",internlm2,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,i,i,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/i/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669474c0-47815cb04b15ea3c0e3d0c0b;0eb05746-6b32-43b7-b397-5b6f8c4eb946) Repository Not Found for url: https://huggingface.co/i/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: i is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,stabilityai/stablelm-3b-4e1t,stabilityai/stablelm-3b-4e1t,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1249.275904,6715.604992,0.0,6069.157888,5763.863552,s,10,0.3415832290649414,0.034158322906494135,0.0002809213528596382,0.03404644775390625,0.03435670700073242,0.034658257675170896,0.034899498214721686,"[0.03495980834960938, 0.0340579833984375, 0.034001216888427735, 0.034066753387451174, 0.033984542846679684, 0.03401929473876953, 0.03401123046875, 0.03428969573974609, 0.034034912109375, 0.03415779113769531]",tokens/s,7494.513144008298,kWh,4.0388607734706223e-07,2.213100763245153e-07,2.4461614184368217e-06,3.0713575721083992e-06,tokens/kWh,83350763.95037368,MB,1249.275904,6736.576512,0.0,6090.129408,5863.102976,s,10,14.847766601562498,1.4847766601562498,0.009434452525821986,1.4838017578125,1.4956170288085937,1.5011369445800782,1.5055528771972655,"[1.5066568603515624, 1.477275390625, 1.48590478515625, 1.475824462890625, 1.4729095458984376, 1.4877525634765625, 1.4845987548828126, 1.494390380859375, 1.4794490966796876, 1.4830047607421875]",tokens/s,42.43062387131693,kWh,1.7854550211751467e-05,9.782296886967015e-06,5.96573327087643e-05,8.729417980748278e-05,tokens/kWh,721697.5993008837,,s,629,15.12954572105409,0.024053331829974687,0.004114993424087828,0.02345779228210449,0.024201830291748046,0.02438041572570801,0.05799691192626953,"[0.0243189754486084, 0.024070144653320313, 0.024370176315307617, 0.024418304443359375, 0.024333311080932618, 0.024264736175537108, 0.0241376953125, 0.024184831619262694, 0.024371200561523438, 0.024382463455200197, 0.02431391906738281, 0.024259519577026368, 0.02434048080444336, 0.024217599868774413, 0.024327167510986326, 0.024425472259521484, 0.024183807373046876, 0.024028160095214843, 0.023979007720947267, 0.024398847579956053, 0.02503167915344238, 0.024534015655517577, 0.024319999694824217, 0.02429542350769043, 0.024276992797851563, 0.024151039123535157, 0.024175615310668946, 0.02412646484375, 0.023448575973510744, 0.023397375106811523, 0.02346598434448242, 0.023434240341186522, 0.023419904708862304, 0.023451648712158202, 0.023436351776123045, 0.02344441604614258, 0.023434240341186522, 0.02351103973388672, 0.023455743789672853, 0.023382015228271484, 0.02346188735961914, 0.023341056823730468, 0.02342911911010742, 0.023448575973510744, 0.023391231536865235, 0.02345471954345703, 0.02342092704772949, 0.023367712020874024, 0.023459808349609375, 0.023391231536865235, 0.023417856216430662, 0.02351820755004883, 0.02349465560913086, 0.023439359664916993, 0.023432191848754884, 0.023401472091674806, 0.023475263595581056, 0.023490495681762695, 0.024637439727783202, 0.024976383209228514, 0.024239103317260743, 0.024284160614013672, 0.058036224365234375, 0.023432191848754884, 0.023398399353027344, 0.02345881652832031, 0.023421951293945312, 0.02344550323486328, 0.023401472091674806, 0.02347315216064453, 0.023396352767944335, 0.023349248886108398, 0.023488512039184572, 0.02344550323486328, 0.0233503360748291, 0.023406528472900392, 0.023444480895996093, 0.023436288833618164, 0.023385087966918947, 0.023367679595947266, 0.023378944396972655, 0.023363584518432616, 0.023425024032592775, 0.02343017578125, 0.023492576599121094, 0.023427072525024413, 0.023327743530273438, 0.023358463287353515, 0.02346905517578125, 0.023407615661621094, 0.02335436820983887, 0.023379968643188476, 0.023378944396972655, 0.023447551727294923, 0.023557119369506836, 0.023408639907836915, 0.023478271484375, 0.023478271484375, 0.023413759231567383, 0.023411712646484374, 0.023513088226318358, 0.023421951293945312, 0.02346598434448242, 0.023443456649780273, 0.023406591415405274, 0.023407615661621094, 0.02349158477783203, 0.023398399353027344, 0.023504896163940428, 0.02350492858886719, 0.023454687118530274, 0.023400480270385743, 0.02356425666809082, 0.023557119369506836, 0.02353152084350586, 0.023476224899291992, 0.023463935852050782, 0.023407615661621094, 0.023554048538208007, 0.02351411247253418, 0.02346598434448242, 0.02349567985534668, 0.02350796890258789, 0.023426048278808592, 0.02348134422302246, 0.057998336791992185, 0.023553024291992186, 0.02350284767150879, 0.023454784393310547, 0.023477184295654298, 0.023446527481079102, 0.023380992889404296, 0.023224319458007812, 0.023182336807250976, 0.023102495193481447, 0.023236576080322265, 0.023434240341186522, 0.023447551727294923, 0.02347929573059082, 0.02348953628540039, 0.02345267105102539, 0.023468032836914062, 0.023532543182373047, 0.023406591415405274, 0.023533567428588868, 0.023533567428588868, 0.023416864395141603, 0.023478239059448243, 0.023430143356323242, 0.02467532730102539, 0.025385984420776365, 0.024627199172973634, 0.02410713577270508, 0.023924640655517578, 0.023842912673950195, 0.0240392951965332, 0.024017919540405275, 0.02349465560913086, 0.023573503494262696, 0.023488512039184572, 0.02348646354675293, 0.02355200004577637, 0.024377344131469726, 0.023631967544555665, 0.023377824783325195, 0.023392288208007813, 0.023446495056152344, 0.023433216094970705, 0.02344960021972656, 0.023433216094970705, 0.023441408157348635, 0.023432191848754884, 0.0235284481048584, 0.023421951293945312, 0.023459840774536132, 0.023425024032592775, 0.02346598434448242, 0.02349567985534668, 0.023480319976806642, 0.023436288833618164, 0.02345369529724121, 0.023657472610473632, 0.02342092704772949, 0.023434240341186522, 0.02351411247253418, 0.02346188735961914, 0.023545856475830077, 0.023443456649780273, 0.05831987380981445, 0.02352332878112793, 0.023571456909179687, 0.02346291160583496, 0.02350592041015625, 0.02347110366821289, 0.02345779228210449, 0.023504896163940428, 0.02355615997314453, 0.023502784729003905, 0.02347110366821289, 0.02345881652832031, 0.023371776580810546, 0.02355200004577637, 0.02346086311340332, 0.023430143356323242, 0.02351411247253418, 0.023558143615722657, 0.023398399353027344, 0.02351103973388672, 0.023814144134521483, 0.02351820755004883, 0.023433216094970705, 0.02355200004577637, 0.02341072082519531, 0.023532512664794922, 0.02351103973388672, 0.02349875259399414, 0.023417856216430662, 0.02355200004577637, 0.023370752334594725, 0.023549951553344727, 0.02352230453491211, 0.02348441505432129, 0.02344960021972656, 0.023508991241455078, 0.023362560272216795, 0.023473215103149415, 0.0235611515045166, 0.023428096771240234, 0.023414783477783203, 0.02347929573059082, 0.023085056304931642, 0.023146495819091797, 0.02325196838378906, 0.02349056053161621, 0.023432191848754884, 0.023431167602539063, 0.023365631103515624, 0.023464960098266603, 0.023370752334594725, 0.02347724723815918, 0.02344960021972656, 0.02348236846923828, 0.02331340789794922, 0.02307174491882324, 0.02308403205871582, 0.023051424026489256, 0.02309209632873535, 0.02310243225097656, 0.02307379150390625, 0.02310553550720215, 0.023136255264282226, 0.057630718231201174, 0.023143423080444335, 0.023375871658325196, 0.023198720932006835, 0.02310041618347168, 0.02312716865539551, 0.023042943954467772, 0.023043071746826172, 0.02309734344482422, 0.023185407638549805, 0.023227392196655275, 0.023128063201904296, 0.02305638313293457, 0.02309222412109375, 0.023240703582763672, 0.023085056304931642, 0.023349248886108398, 0.023472127914428712, 0.023419904708862304, 0.023394304275512694, 0.0235100154876709, 0.023411712646484374, 0.02345779228210449, 0.02348236846923828, 0.023430143356323242, 0.023447551727294923, 0.02370560073852539, 0.023410688400268553, 0.02344550323486328, 0.023390207290649414, 0.023396352767944335, 0.023369728088378908, 0.023415807723999024, 0.023402496337890624, 0.023443456649780273, 0.023372800827026367, 0.023384063720703126, 0.023401472091674806, 0.023415807723999024, 0.02344550323486328, 0.02348953628540039, 0.023421951293945312, 0.023414880752563476, 0.023358367919921876, 0.023433216094970705, 0.02346188735961914, 0.023426048278808592, 0.023547903060913086, 0.023410688400268553, 0.023407615661621094, 0.023569408416748046, 0.02348646354675293, 0.023606271743774415, 0.02345779228210449, 0.023526464462280273, 0.023427007675170898, 0.02345471954345703, 0.02347007942199707, 0.02351923179626465, 0.023554048538208007, 0.023448575973510744, 0.023406591415405274, 0.023472127914428712, 0.05804339218139649, 0.02350387191772461, 0.02408755111694336, 0.023702527999877928, 0.02348441505432129, 0.023425024032592775, 0.023435264587402343, 0.02332057571411133, 0.024137727737426756, 0.02434764862060547, 0.02429542350769043, 0.024327167510986326, 0.024239103317260743, 0.023469120025634765, 0.023427007675170898, 0.023392255783081056, 0.024261632919311524, 0.024201215744018553, 0.023198720932006835, 0.023152639389038086, 0.023141376495361327, 0.02299920082092285, 0.023109472274780274, 0.02393497657775879, 0.023546880722045898, 0.025165824890136718, 0.025044031143188476, 0.024078271865844728, 0.024030208587646484, 0.023972864151000976, 0.023961599349975587, 0.023937023162841797, 0.024005632400512695, 0.023842815399169923, 0.02342911911010742, 0.023193599700927735, 0.023129215240478517, 0.02310335922241211, 0.02308710479736328, 0.023026687622070312, 0.02333695983886719, 0.023129087448120117, 0.023152736663818358, 0.023136159896850587, 0.023407615661621094, 0.0233175048828125, 0.023348224639892577, 0.023567359924316408, 0.023390207290649414, 0.02351206398010254, 0.023517183303833008, 0.023425056457519532, 0.02352943992614746, 0.02353152084350586, 0.02348748779296875, 0.02352639961242676, 0.023421951293945312, 0.02349363136291504, 0.023443487167358397, 0.023484384536743164, 0.023524351119995117, 0.023362560272216795, 0.023558143615722657, 0.057993247985839845, 0.023377887725830077, 0.023391231536865235, 0.023476224899291992, 0.023373823165893554, 0.023549951553344727, 0.023427072525024413, 0.023417856216430662, 0.023607295989990236, 0.023508991241455078, 0.023641088485717773, 0.02352742385864258, 0.02346598434448242, 0.02350182342529297, 0.023366655349731445, 0.02346598434448242, 0.023436288833618164, 0.023386112213134767, 0.02347007942199707, 0.023447551727294923, 0.023601152420043944, 0.02347315216064453, 0.023392255783081056, 0.023533567428588868, 0.02332876777648926, 0.02353049659729004, 0.023421951293945312, 0.023367679595947266, 0.02367283248901367, 0.024135679244995118, 0.025536512374877928, 0.02455449676513672, 0.024352767944335937, 0.024139776229858398, 0.02342911911010742, 0.02350796890258789, 0.023508991241455078, 0.02348646354675293, 0.02349056053161621, 0.02344960021972656, 0.023498783111572264, 0.02340553665161133, 0.023409664154052736, 0.02355200004577637, 0.023378944396972655, 0.02348441505432129, 0.02351206398010254, 0.023472127914428712, 0.023480319976806642, 0.023382015228271484, 0.02349056053161621, 0.02346086311340332, 0.023463935852050782, 0.023527456283569337, 0.02338096046447754, 0.023472127914428712, 0.023382112503051757, 0.023459775924682617, 0.023515104293823242, 0.02346700859069824, 0.023778303146362305, 0.023564287185668945, 0.023370752334594725, 0.058297344207763675, 0.023408639907836915, 0.02350284767150879, 0.023422975540161133, 0.023415807723999024, 0.02331648063659668, 0.023433216094970705, 0.02326323127746582, 0.02426572799682617, 0.024196096420288086, 0.024018943786621092, 0.024142847061157227, 0.023967744827270508, 0.0241582088470459, 0.0241582088470459, 0.024054784774780274, 0.024240127563476564, 0.02411827278137207, 0.02346700859069824, 0.023397375106811523, 0.023372800827026367, 0.02345267105102539, 0.023414783477783203, 0.023548959732055664, 0.023414751052856446, 0.023402496337890624, 0.02346700859069824, 0.023444480895996093, 0.023579647064208984, 0.02350592041015625, 0.023342079162597656, 0.023468032836914062, 0.0233175048828125, 0.02351820755004883, 0.023673856735229492, 0.02555801582336426, 0.024400896072387695, 0.024129535675048826, 0.024209407806396483, 0.0241213436126709, 0.024258560180664062, 0.02428620719909668, 0.02412441635131836, 0.02430975914001465, 0.024558591842651366, 0.023789567947387694, 0.023568416595458986, 0.023423967361450197, 0.023504896163940428, 0.023516159057617187, 0.02349056053161621, 0.02344960021972656, 0.02350387191772461, 0.02350387191772461, 0.023388160705566406, 0.023540735244750977, 0.02349158477783203, 0.02345062446594238, 0.02349875259399414, 0.023426048278808592, 0.023361536026000978, 0.023392255783081056, 0.02329190444946289, 0.05805055999755859, 0.023436288833618164, 0.023431167602539063, 0.02343731117248535, 0.023513088226318358, 0.02348953628540039, 0.023392255783081056, 0.023423040390014648, 0.023398336410522462, 0.023305248260498047, 0.023516128540039063, 0.023426048278808592, 0.023421951293945312, 0.023410688400268553, 0.023421951293945312, 0.023380992889404296, 0.023533567428588868, 0.023380992889404296, 0.023412736892700195, 0.02344960021972656, 0.023372800827026367, 0.023404544830322265, 0.023431167602539063, 0.02349977684020996, 0.023407615661621094, 0.023350271224975586, 0.02345779228210449, 0.023380992889404296, 0.023430143356323242, 0.023441408157348635, 0.02351206398010254, 0.023569440841674803, 0.023420896530151367, 0.023351295471191406, 0.02365644836425781, 0.023561216354370116, 0.02345779228210449, 0.02342911911010742, 0.023348224639892577, 0.023472127914428712, 0.023468032836914062, 0.02343731117248535, 0.023356416702270507, 0.02345471954345703, 0.023435264587402343, 0.02344960021972656, 0.023558143615722657, 0.02345779228210449, 0.02347007942199707, 0.023525375366210938, 0.023257087707519532, 0.023436288833618164, 0.023582719802856447, 0.023546880722045898, 0.02353459167480469, 0.0234833927154541, 0.023538688659667968, 0.02346086311340332, 0.023504896163940428, 0.02353971290588379, 0.023565311431884766, 0.02348953628540039, 0.024611839294433592, 0.05835366439819336, 0.023222272872924804, 0.024204288482666016, 0.024321023941040038, 0.024071168899536134, 0.024460351943969727, 0.02442848014831543, 0.024451072692871095, 0.02448588752746582, 0.024358911514282225, 0.023616512298583983, 0.023593984603881835, 0.023472127914428712, 0.023440383911132814, 0.023450687408447267, 0.02343622398376465, 0.02344550323486328, 0.02350182342529297, 0.023405567169189453, 0.023488512039184572, 0.023557119369506836, 0.023412736892700195, 0.023435295104980467, 0.023483360290527344, 0.023394304275512694, 0.023403520584106444, 0.023451648712158202, 0.02332159996032715, 0.02330624008178711, 0.023224319458007812, 0.023030784606933592, 0.02307379150390625, 0.02333798408508301, 0.023425024032592775, 0.023407615661621094, 0.023379968643188476, 0.02333286476135254, 0.02345471954345703, 0.023391231536865235, 0.023387136459350585, 0.023348224639892577, 0.02345062446594238, 0.023352319717407227, 0.023387136459350585, 0.023423999786376954, 0.023390207290649414, 0.023405567169189453, 0.023379968643188476, 0.023407615661621094, 0.023389184951782226, 0.023409664154052736, 0.02342092704772949, 0.023347200393676756, 0.023456768035888673, 0.023369728088378908, 0.023396352767944335, 0.023380992889404296, 0.02342911911010742, 0.023341056823730468, 0.023928831100463867, 0.02366361618041992, 0.023439359664916993, 0.023423999786376954]",tokens/s,41.574281977593806,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,huggyllama/llama-13b,huggyllama/llama-13b,cuda,0,42,,,True,,,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",llama,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-2.7b,facebook/opt-2.7b,cuda,0,42,,,True,,,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",opt,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,microsoft/phi-1_5,microsoft/phi-1_5,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1257.242624,3855.089664,0.0,3208.64256,2982.448128,s,10,0.24312742614746097,0.024312742614746096,0.00029150430413419476,0.024220175743103027,0.024407213592529297,0.02478699893951416,0.02509082721710205,"[0.025166784286499023, 0.02420150375366211, 0.02424515151977539, 0.024238847732543947, 0.02414384078979492, 0.024316864013671877, 0.024194528579711914, 0.02417513656616211, 0.024121952056884766, 0.024322816848754883]",tokens/s,10529.457908411025,kWh,2.8704562241527e-07,1.572854325808541e-07,1.3237267871359147e-06,1.7680578421320387e-06,tokens/kWh,144791643.06711742,MB,1257.242624,3855.089664,0.0,3208.64256,3022.612992,s,10,11.509342041015625,1.1509342041015624,0.018404604917383584,1.1566403198242186,1.167904541015625,1.1749253540039062,1.1805420043945314,"[1.1819461669921876, 1.141770263671875, 1.157016845703125, 1.12313916015625, 1.118952880859375, 1.1571549072265626, 1.14356103515625, 1.1562637939453124, 1.163192626953125, 1.1663443603515624]",tokens/s,54.738142089693824,kWh,1.3269401936517563e-05,7.271264468797721e-06,3.703438723266408e-05,5.757505363797935e-05,tokens/kWh,1094223.904612085,,s,629,11.707308048248295,0.01861257241374927,0.0029418714982966552,0.018381824493408205,0.018630451583862305,0.019184640502929687,0.04226715621948242,"[0.019560447692871095, 0.019190784454345702, 0.019102720260620116, 0.019169279098510742, 0.01920102310180664, 0.019109888076782225, 0.0192675838470459, 0.019527679443359376, 0.019131391525268555, 0.019164159774780275, 0.01925734329223633, 0.019386367797851564, 0.019516416549682617, 0.019211263656616212, 0.0192174072265625, 0.019200000762939453, 0.019345407485961915, 0.019113983154296875, 0.019107839584350587, 0.019382272720336914, 0.01926144027709961, 0.01923788833618164, 0.019006464004516603, 0.019175424575805664, 0.01843507194519043, 0.018301952362060548, 0.022184959411621095, 0.020431936264038084, 0.019533760070800783, 0.01879347229003906, 0.01834499168395996, 0.018344928741455078, 0.018372608184814454, 0.018312255859375, 0.018288576126098632, 0.018310144424438478, 0.018337791442871093, 0.018358272552490236, 0.0180316162109375, 0.01829478454589844, 0.0184453125, 0.01841663932800293, 0.01844121551513672, 0.0182794246673584, 0.01841868782043457, 0.018287616729736327, 0.017941503524780272, 0.018301952362060548, 0.018313215255737304, 0.018335744857788085, 0.018301952362060548, 0.018355199813842774, 0.018355199813842774, 0.018509824752807616, 0.01843302345275879, 0.017968128204345703, 0.01803468894958496, 0.01741209602355957, 0.017641471862792968, 0.017657855987548828, 0.01763430404663086, 0.01765376091003418, 0.04227174377441406, 0.01760153579711914, 0.01763942337036133, 0.01764761543273926, 0.01761177635192871, 0.017822719573974608, 0.018514944076538087, 0.01839411163330078, 0.01825382423400879, 0.01827020835876465, 0.018264064788818358, 0.018342912673950194, 0.0184453125, 0.018280448913574218, 0.01840025520324707, 0.017921024322509766, 0.01764556884765625, 0.017675264358520508, 0.017760255813598632, 0.01780633544921875, 0.017675264358520508, 0.01779097557067871, 0.017752063751220702, 0.017682432174682617, 0.017678335189819337, 0.017821695327758787, 0.017733631134033204, 0.017737728118896484, 0.01765273666381836, 0.01762611198425293, 0.017691648483276368, 0.01764454460144043, 0.017670143127441407, 0.017740800857543947, 0.017648639678955077, 0.017709056854248048, 0.01763737678527832, 0.017704959869384765, 0.01883647918701172, 0.019252223968505858, 0.019607551574707033, 0.01862553596496582, 0.018337791442871093, 0.018532352447509767, 0.01840025520324707, 0.018391040802001952, 0.018532352447509767, 0.018359296798706053, 0.018375680923461913, 0.018382848739624022, 0.018365440368652345, 0.018396160125732423, 0.018380800247192384, 0.018415615081787108, 0.018342912673950194, 0.018391040802001952, 0.018344959259033202, 0.01850783920288086, 0.01841043281555176, 0.01841459274291992, 0.018364416122436524, 0.018405376434326173, 0.01843302345275879, 0.04312473678588867, 0.01843507194519043, 0.018289663314819335, 0.01839820861816406, 0.018310144424438478, 0.018307071685791015, 0.01836851119995117, 0.018365440368652345, 0.018317312240600587, 0.01865216064453125, 0.018569215774536133, 0.018351104736328124, 0.01836031913757324, 0.018391040802001952, 0.01839820861816406, 0.018461696624755858, 0.01836851119995117, 0.018320383071899413, 0.018378751754760742, 0.018340864181518556, 0.018340864181518556, 0.018365440368652345, 0.018408447265625, 0.018431999206542968, 0.01835212707519531, 0.018357248306274415, 0.018359296798706053, 0.018361343383789062, 0.018380800247192384, 0.018266111373901366, 0.01837772750854492, 0.018323455810546875, 0.018348031997680665, 0.018463743209838866, 0.018340864181518556, 0.018351104736328124, 0.018239488601684572, 0.017987648010253907, 0.017995712280273437, 0.018068479537963866, 0.01801625633239746, 0.01839411163330078, 0.018344959259033202, 0.01828972816467285, 0.018481088638305665, 0.018346048355102538, 0.018390975952148437, 0.01836031913757324, 0.018343936920166014, 0.01842585563659668, 0.018274303436279296, 0.018694143295288086, 0.01875660705566406, 0.018358272552490236, 0.018363391876220703, 0.01835212707519531, 0.018415615081787108, 0.01842790412902832, 0.018310144424438478, 0.018323455810546875, 0.018395135879516602, 0.018343936920166014, 0.01827846336364746, 0.042399681091308594, 0.017666048049926757, 0.01766092872619629, 0.017682432174682617, 0.017617919921875, 0.017633279800415038, 0.01766912078857422, 0.017708032608032227, 0.01761177635192871, 0.017771520614624024, 0.017666048049926757, 0.017584127426147463, 0.017485824584960938, 0.017999872207641602, 0.018379776000976563, 0.018358272552490236, 0.018340864181518556, 0.018307071685791015, 0.018448383331298827, 0.018323455810546875, 0.01826201629638672, 0.018148351669311523, 0.018287616729736327, 0.018295808792114256, 0.018314239501953124, 0.018326528549194337, 0.018283519744873047, 0.018338815689086914, 0.01816268730163574, 0.017897472381591797, 0.01761689567565918, 0.017960960388183594, 0.01778278350830078, 0.01763737678527832, 0.017503231048583985, 0.01756979179382324, 0.017563648223876953, 0.017655807495117186, 0.017754112243652344, 0.017725439071655275, 0.01764556884765625, 0.017719295501708983, 0.01762713623046875, 0.017737728118896484, 0.01764761543273926, 0.017721376419067382, 0.01767318344116211, 0.017638431549072266, 0.01771516799926758, 0.01770086479187012, 0.01767731285095215, 0.017715200424194336, 0.017733631134033204, 0.01765068817138672, 0.01759231948852539, 0.01764659118652344, 0.01768448066711426, 0.01762918472290039, 0.01766912078857422, 0.017633279800415038, 0.01741926383972168, 0.017527807235717775, 0.01760358428955078, 0.0422369270324707, 0.01764659118652344, 0.017656831741333007, 0.01765888023376465, 0.01766912078857422, 0.01768550491333008, 0.017687551498413084, 0.017548288345336914, 0.017497087478637697, 0.0176629753112793, 0.01770086479187012, 0.01784012794494629, 0.017667072296142578, 0.01765171241760254, 0.017695743560791014, 0.017712127685546874, 0.017649728775024413, 0.017644479751586915, 0.01762611198425293, 0.01767628860473633, 0.017745920181274414, 0.017727487564086913, 0.01764249610900879, 0.01775923156738281, 0.01761996841430664, 0.017672191619873046, 0.017469440460205078, 0.01743257522583008, 0.017546239852905272, 0.017743871688842772, 0.017641471862792968, 0.01781350326538086, 0.01763942337036133, 0.01777561569213867, 0.01762918472290039, 0.01765888023376465, 0.01765171241760254, 0.017675264358520508, 0.017760255813598632, 0.017673215866088866, 0.017746944427490235, 0.017695743560791014, 0.017691648483276368, 0.017739839553833008, 0.01753081512451172, 0.017541120529174805, 0.017504255294799806, 0.01763737678527832, 0.017655807495117186, 0.017678335189819337, 0.0176629753112793, 0.01763737678527832, 0.017729536056518554, 0.01761484718322754, 0.017614912033081055, 0.017667007446289063, 0.017629215240478516, 0.01787900733947754, 0.018758655548095703, 0.019554399490356447, 0.018701215744018555, 0.018542591094970702, 0.01861222457885742, 0.043154430389404294, 0.018291744232177734, 0.018590688705444336, 0.01845248031616211, 0.01824051284790039, 0.018333696365356447, 0.018317312240600587, 0.018516992568969725, 0.018276351928710938, 0.018308095932006836, 0.01837772750854492, 0.01835526466369629, 0.018358207702636718, 0.017719295501708983, 0.017396768569946288, 0.0174366397857666, 0.017754112243652344, 0.017703935623168944, 0.01841049575805664, 0.018053119659423827, 0.018093055725097656, 0.018379776000976563, 0.01838591957092285, 0.018405376434326173, 0.01844428825378418, 0.018498559951782227, 0.01942323112487793, 0.01865216064453125, 0.01865830421447754, 0.01840640068054199, 0.01843715286254883, 0.018400224685668945, 0.018380800247192384, 0.01843302345275879, 0.01840640068054199, 0.01840947151184082, 0.018453504562377928, 0.018404352188110353, 0.018421760559082033, 0.01856208038330078, 0.018456544876098633, 0.018516992568969725, 0.018568191528320312, 0.01843609619140625, 0.018562047958374024, 0.01846681594848633, 0.01841971206665039, 0.01844428825378418, 0.01843302345275879, 0.018453504562377928, 0.018062335968017578, 0.018240575790405274, 0.01853023910522461, 0.018359296798706053, 0.01844633674621582, 0.01839308738708496, 0.018388992309570314, 0.018487295150756835, 0.018423807144165038, 0.018480127334594726, 0.018379776000976563, 0.018437120437622072, 0.018437120437622072, 0.04303974533081055, 0.018421760559082033, 0.018510847091674804, 0.018441280364990233, 0.01858246421813965, 0.018503679275512695, 0.018467872619628907, 0.018644960403442382, 0.01847091293334961, 0.017582080841064454, 0.017570816040039062, 0.017596416473388672, 0.017720319747924804, 0.017727487564086913, 0.01765273666381836, 0.017666048049926757, 0.017675264358520508, 0.017641471862792968, 0.017670143127441407, 0.0176629753112793, 0.0177357120513916, 0.017598432540893556, 0.017531904220581054, 0.0176680965423584, 0.018455551147460936, 0.01844223976135254, 0.01844633674621582, 0.018489343643188477, 0.018387968063354493, 0.018496511459350586, 0.018586624145507814, 0.01878428840637207, 0.018762720108032226, 0.018488319396972656, 0.018532384872436525, 0.018407392501831054, 0.018481151580810547, 0.01841868782043457, 0.01849241638183594, 0.018404352188110353, 0.01843302345275879, 0.018520063400268554, 0.01840332794189453, 0.01848422431945801, 0.01843609619140625, 0.018499584197998048, 0.01843507194519043, 0.01857535934448242, 0.01845452880859375, 0.018135040283203126, 0.01839820861816406, 0.018118656158447266, 0.018345983505249023, 0.017695743560791014, 0.01763020706176758, 0.01765478324890137, 0.01761075210571289, 0.0176363525390625, 0.017743871688842772, 0.01763020706176758, 0.017703935623168944, 0.01763532829284668, 0.017757183074951173, 0.0422553596496582, 0.017648639678955077, 0.01798041534423828, 0.018257919311523436, 0.018347007751464844, 0.018382848739624022, 0.01836851119995117, 0.018551807403564453, 0.018344959259033202, 0.018339839935302735, 0.01839206314086914, 0.018530303955078126, 0.018506752014160157, 0.01848320007324219, 0.018449407577514648, 0.018501632690429686, 0.018408447265625, 0.018533376693725585, 0.018387968063354493, 0.01855183982849121, 0.018471904754638672, 0.01848320007324219, 0.018421760559082033, 0.018447359085083007, 0.01846886444091797, 0.01843302345275879, 0.018361343383789062, 0.018367488861083983, 0.018464767456054687, 0.01840640068054199, 0.017683456420898438, 0.017649728775024413, 0.017678272247314452, 0.017746944427490235, 0.018516992568969725, 0.018569215774536133, 0.018386943817138672, 0.01844428825378418, 0.01843302345275879, 0.018489343643188477, 0.01846067237854004, 0.018469888687133788, 0.0184268798828125, 0.018423807144165038, 0.01858252716064453, 0.018518016815185546, 0.01843612861633301, 0.018411487579345704, 0.018341888427734376, 0.018501632690429686, 0.018509824752807616, 0.018322431564331054, 0.0184268798828125, 0.018379776000976563, 0.01843404769897461, 0.01836851119995117, 0.01865830421447754, 0.018242559432983398, 0.018181119918823242, 0.01823849678039551, 0.018117599487304688, 0.018495487213134765, 0.018361343383789062, 0.042742782592773435, 0.01839206314086914, 0.018481151580810547, 0.01843916893005371, 0.01843302345275879, 0.01839206314086914, 0.018395135879516602, 0.018397184371948243, 0.018365440368652345, 0.018166784286499024, 0.018571264266967775, 0.018506752014160157, 0.018465791702270508, 0.018153472900390624, 0.018258943557739257, 0.01841152000427246, 0.01845145606994629, 0.018416671752929686, 0.018335712432861327, 0.018386943817138672, 0.018381824493408205, 0.018463743209838866, 0.01803468894958496, 0.01767628860473633, 0.018388992309570314, 0.018495487213134765, 0.018388992309570314, 0.018378751754760742, 0.018675712585449217, 0.01847091293334961, 0.018437120437622072, 0.01843715286254883, 0.01851388740539551, 0.01881395149230957, 0.01862860870361328, 0.018493440628051756, 0.01839923286437988, 0.018497535705566406, 0.018051071166992186, 0.018557952880859374, 0.018524160385131837, 0.018498559951782227, 0.018359296798706053, 0.01820569610595703, 0.018296831130981444, 0.01846784019470215, 0.01840230369567871, 0.018494464874267577, 0.018486272811889647, 0.01843507194519043, 0.019021823883056642, 0.020165632247924805, 0.018737152099609376, 0.018740224838256835, 0.018508800506591795, 0.01844223976135254, 0.018508800506591795, 0.018473983764648438, 0.018493440628051756, 0.01846272087097168, 0.01841459274291992, 0.01845964813232422, 0.01839923286437988, 0.04325785446166992, 0.01840332794189453, 0.018486272811889647, 0.018515968322753908, 0.018489343643188477, 0.018485248565673826, 0.018229248046875, 0.018564096450805666, 0.01846272087097168, 0.018551807403564453, 0.018499584197998048, 0.01861631965637207, 0.0184770565032959, 0.018509824752807616, 0.018525184631347655, 0.018525184631347655, 0.01844121551513672, 0.018487295150756835, 0.018637823104858398, 0.01841663932800293, 0.018472959518432617, 0.018461696624755858, 0.018469888687133788, 0.01846886444091797, 0.01841971206665039, 0.01847500801086426, 0.01862041664123535, 0.018464767456054687, 0.018512895584106445, 0.018523136138916017, 0.018463743209838866, 0.018562047958374024, 0.01842995262145996, 0.01845248031616211, 0.01845964813232422, 0.018530303955078126, 0.01845248031616211, 0.01844223976135254, 0.0184453125, 0.018514944076538087, 0.018437120437622072, 0.018469888687133788, 0.018461696624755858, 0.018536447525024414, 0.018587648391723634, 0.01856716728210449, 0.018566144943237304, 0.018571264266967775, 0.01843916893005371, 0.018539520263671876, 0.018480127334594726, 0.01840540885925293, 0.01847395133972168, 0.01841971206665039, 0.018595872879028322, 0.01867158317565918, 0.018548736572265623, 0.01889587211608887, 0.018562047958374024, 0.018531328201293946, 0.018651136398315428, 0.018610176086425782, 0.018456575393676757]",tokens/s,53.727124750434356,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,google/recurrentgemma-2b,google/recurrentgemma-2b,cuda,0,42,,,True,,,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",recurrent_gemma,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,stabilityai/stablelm-2-1_6b,stabilityai/stablelm-2-1_6b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1264.66048,4276.617216,0.0,3632.267264,3408.132096,s,10,0.213487455368042,0.0213487455368042,0.00038207157218038363,0.021232367515563965,0.021424921989440918,0.021956252574920652,0.02238131704330444,"[0.02248758316040039, 0.021306848526000975, 0.02123923110961914, 0.021238431930541993, 0.021165184020996094, 0.021216960906982423, 0.021226303100585937, 0.021224639892578126, 0.021244895935058593, 0.02113737678527832]",tokens/s,11991.336894182772,kWh,2.5255655744250117e-07,1.383878192229578e-07,1.2493534448760467e-06,1.6402978215415056e-06,tokens/kWh,156069219.03939274,MB,1264.66048,4278.714368,0.0,3632.267264,3474.14784,s,10,11.683104492187502,1.16831044921875,0.03179779513772458,1.1519906616210938,1.2127111938476562,1.2202495178222657,1.2262801770019531,"[1.227787841796875, 1.2032935791015624, 1.2110360107421876, 1.154431640625, 1.1466201171875, 1.14823046875, 1.16621044921875, 1.126555908203125, 1.1495496826171876, 1.1493887939453125]",tokens/s,53.92402339817139,kWh,1.3614571337912724e-05,7.45993332269954e-06,3.769223326392422e-05,5.876673792453647e-05,tokens/kWh,1072035.001855974,,s,629,11.854563329696665,0.018846682559136175,0.0026529975843851924,0.018518016815185546,0.019255309295654298,0.019856384277343753,0.039754261932373046,"[0.020084735870361328, 0.019014656066894533, 0.018783231735229493, 0.018775039672851563, 0.01883852767944336, 0.019876863479614256, 0.01988812828063965, 0.020040704727172853, 0.01999564743041992, 0.020165632247924805, 0.02001408004760742, 0.020239360809326173, 0.020168703079223634, 0.018760704040527345, 0.018883583068847656, 0.019992576599121094, 0.020246528625488282, 0.02003763198852539, 0.019959808349609375, 0.020493312835693358, 0.01999667167663574, 0.019877887725830077, 0.02018611145019531, 0.01987481689453125, 0.019722240447998047, 0.019971071243286134, 0.020051967620849608, 0.019685375213623048, 0.019695680618286134, 0.0195665283203125, 0.01925017547607422, 0.01929113578796387, 0.019106815338134766, 0.019129344940185547, 0.019131391525268555, 0.019169279098510742, 0.01924300765991211, 0.019117055892944337, 0.01924095916748047, 0.019198976516723632, 0.019117055892944337, 0.019273727416992188, 0.019224576950073242, 0.01922150421142578, 0.01925836753845215, 0.01920204734802246, 0.01918976020812988, 0.019185663223266602, 0.019135488510131835, 0.019102783203125, 0.019727359771728514, 0.019380159378051758, 0.019219455718994142, 0.019142656326293944, 0.019133440017700197, 0.01928294372558594, 0.01922764778137207, 0.019180543899536134, 0.019107839584350587, 0.019132415771484376, 0.019449855804443358, 0.019255359649658202, 0.04104185485839844, 0.019121152877807617, 0.017993728637695314, 0.01784934425354004, 0.019305471420288087, 0.019555328369140625, 0.020343807220458983, 0.01921023941040039, 0.019116031646728517, 0.019131391525268555, 0.01904947280883789, 0.018940927505493164, 0.019759103775024413, 0.01923583984375, 0.01903411293029785, 0.01919385528564453, 0.018884607315063476, 0.0193832950592041, 0.019002368927001953, 0.01907302474975586, 0.01902899169921875, 0.01909974479675293, 0.018896799087524414, 0.018964479446411133, 0.01901158332824707, 0.01903513526916504, 0.018955263137817382, 0.018966527938842775, 0.018876415252685547, 0.01902592086791992, 0.019036384582519533, 0.018917152404785156, 0.018874368667602538, 0.019066879272460938, 0.01904947280883789, 0.018949119567871094, 0.01890928077697754, 0.018877344131469728, 0.018852863311767578, 0.018979839324951172, 0.018905120849609373, 0.01912931251525879, 0.019134464263916014, 0.019132415771484376, 0.018922496795654296, 0.01922662353515625, 0.01927884864807129, 0.0198287353515625, 0.01924095916748047, 0.019117088317871095, 0.019179487228393556, 0.01923993682861328, 0.01919692802429199, 0.019152896881103516, 0.019204095840454103, 0.019125247955322267, 0.0192225284576416, 0.019186687469482423, 0.019117055892944337, 0.0191016960144043, 0.019135488510131835, 0.019110912322998046, 0.01915395164489746, 0.040965087890625, 0.019312671661376953, 0.019208223342895507, 0.019216320037841798, 0.019125247955322267, 0.01923891258239746, 0.019151872634887695, 0.019211263656616212, 0.01926553535461426, 0.01918976020812988, 0.01921843147277832, 0.019155967712402345, 0.01922662353515625, 0.019180576324462892, 0.019234783172607423, 0.019284000396728517, 0.0191231689453125, 0.01922559928894043, 0.01924403190612793, 0.019207168579101562, 0.019216384887695313, 0.019204095840454103, 0.01925529670715332, 0.019296255111694336, 0.01922559928894043, 0.01920512008666992, 0.01949388885498047, 0.019188735961914064, 0.019368959426879884, 0.01924812889099121, 0.019203071594238282, 0.01926348876953125, 0.019186687469482423, 0.019183616638183593, 0.01922969627380371, 0.019182592391967773, 0.01921843147277832, 0.01925734329223633, 0.019155967712402345, 0.01922662353515625, 0.01921846389770508, 0.019209184646606446, 0.019187711715698243, 0.019168256759643554, 0.019188735961914064, 0.01925529670715332, 0.019190784454345702, 0.019280895233154297, 0.01904025650024414, 0.019212287902832033, 0.019185663223266602, 0.019186687469482423, 0.01919692802429199, 0.019138559341430664, 0.019219455718994142, 0.01930342483520508, 0.019212287902832033, 0.019184640884399414, 0.019152896881103516, 0.019156991958618166, 0.019207168579101562, 0.019178495407104493, 0.01923788833618164, 0.04085760116577149, 0.01927884864807129, 0.019215360641479492, 0.019145727157592773, 0.019168256759643554, 0.019159040451049804, 0.01921331214904785, 0.01912633514404297, 0.0191026554107666, 0.019134464263916014, 0.019147775650024415, 0.01918156814575195, 0.019137535095214844, 0.019098623275756836, 0.01919900894165039, 0.01869308853149414, 0.01801113510131836, 0.0180316162109375, 0.017942527770996093, 0.017954816818237306, 0.017983488082885742, 0.017966079711914062, 0.01884160041809082, 0.01899212837219238, 0.01795686340332031, 0.01794047927856445, 0.01821696090698242, 0.01801318359375, 0.0180316162109375, 0.01799884796142578, 0.017928192138671875, 0.017938432693481447, 0.017985536575317384, 0.017963008880615236, 0.017935359954833984, 0.017999872207641602, 0.018159616470336915, 0.01804697608947754, 0.018000896453857423, 0.017970176696777345, 0.01802137565612793, 0.018000896453857423, 0.01795686340332031, 0.018025472640991212, 0.017991680145263672, 0.017946624755859376, 0.017977344512939454, 0.01798860740661621, 0.01802649688720703, 0.01799065589904785, 0.018050048828125, 0.017984512329101563, 0.017992704391479493, 0.017987583160400392, 0.017967103958129883, 0.018000959396362304, 0.018033599853515624, 0.01797324752807617, 0.018017280578613282, 0.01801215934753418, 0.01802239990234375, 0.018018367767333985, 0.018074560165405273, 0.03959500885009765, 0.01806540870666504, 0.01801215934753418, 0.01803980827331543, 0.018058240890502928, 0.018025472640991212, 0.018118656158447266, 0.017977344512939454, 0.01803059196472168, 0.01804287910461426, 0.017985536575317384, 0.01800704002380371, 0.01802137565612793, 0.01799884796142578, 0.017961984634399415, 0.017984512329101563, 0.017932287216186525, 0.01799065589904785, 0.018069503784179687, 0.01820364761352539, 0.01797324752807617, 0.017994752883911135, 0.01802444839477539, 0.017979391098022462, 0.01820467185974121, 0.01805516815185547, 0.018050048828125, 0.01800499153137207, 0.018027519226074217, 0.01800704002380371, 0.018070528030395508, 0.018022464752197265, 0.018017215728759764, 0.01795180892944336, 0.017980352401733398, 0.01802137565612793, 0.018052127838134764, 0.01805513572692871, 0.01802137565612793, 0.018542591094970702, 0.018745344161987306, 0.018569215774536133, 0.018529279708862305, 0.018553855895996094, 0.018511871337890624, 0.018541568756103514, 0.018538496017456055, 0.018495487213134765, 0.018338815689086914, 0.018391103744506837, 0.0183438720703125, 0.01828556823730469, 0.01839308738708496, 0.017986560821533205, 0.01802649688720703, 0.018028543472290038, 0.018205759048461913, 0.019146688461303712, 0.018747392654418944, 0.018565120697021483, 0.018513919830322266, 0.018601984024047852, 0.018572288513183592, 0.0403056640625, 0.018661376953125, 0.018519039154052733, 0.01843609619140625, 0.018516992568969725, 0.01840230369567871, 0.018286752700805663, 0.0183089599609375, 0.01839206314086914, 0.018584575653076172, 0.01864806365966797, 0.018548736572265623, 0.01858252716064453, 0.018531328201293946, 0.01860403251647949, 0.017958911895751953, 0.01804800033569336, 0.017993728637695314, 0.018000896453857423, 0.018076671600341796, 0.017968128204345703, 0.01802956771850586, 0.01798041534423828, 0.01803980827331543, 0.01800704002380371, 0.01802649688720703, 0.017976320266723633, 0.01804902458190918, 0.017938432693481447, 0.017992704391479493, 0.018074623107910158, 0.017937408447265626, 0.01799679946899414, 0.017945600509643556, 0.017992704391479493, 0.01802444839477539, 0.017994752883911135, 0.0179814395904541, 0.01800912094116211, 0.018014175415039062, 0.01799884796142578, 0.017999872207641602, 0.018085887908935547, 0.01798041534423828, 0.017942527770996093, 0.01795686340332031, 0.018217983245849608, 0.01799577522277832, 0.018714624404907225, 0.01862246322631836, 0.01858355140686035, 0.01858252716064453, 0.018586624145507814, 0.018572288513183592, 0.01859071922302246, 0.01861734390258789, 0.01860710334777832, 0.01855897521972656, 0.017985536575317384, 0.017977344512939454, 0.01799782371520996, 0.01795686340332031, 0.0179814395904541, 0.03977318572998047, 0.018085887908935547, 0.01803878402709961, 0.01802239990234375, 0.01801523208618164, 0.01802956771850586, 0.01799782371520996, 0.018009088516235353, 0.01802137565612793, 0.01802649688720703, 0.01799884796142578, 0.01798041534423828, 0.01799065589904785, 0.01803059196472168, 0.01797222328186035, 0.018104320526123048, 0.01846886444091797, 0.019963903427124022, 0.020136991500854493, 0.01948259162902832, 0.0188590087890625, 0.01863680076599121, 0.018510847091674804, 0.018619392395019533, 0.018541568756103514, 0.018537471771240235, 0.018573312759399413, 0.018570240020751954, 0.018671615600585938, 0.018581504821777343, 0.018563072204589845, 0.018610176086425782, 0.018597888946533202, 0.01859071922302246, 0.018700288772583007, 0.018580480575561522, 0.018521120071411133, 0.018568159103393555, 0.01858355140686035, 0.018562047958374024, 0.01859584045410156, 0.018538496017456055, 0.018646015167236327, 0.018518016815185546, 0.018536447525024414, 0.018564096450805666, 0.018542591094970702, 0.018538496017456055, 0.018534400939941405, 0.018552831649780274, 0.018530303955078126, 0.01859174346923828, 0.018586624145507814, 0.018677759170532226, 0.018572288513183592, 0.018903039932250978, 0.0186562557220459, 0.018512895584106445, 0.018532352447509767, 0.018535423278808593, 0.018515968322753908, 0.018544639587402344, 0.018597888946533202, 0.04017663955688477, 0.018563072204589845, 0.01850476837158203, 0.018381759643554686, 0.01809715270996094, 0.017863679885864257, 0.01781862449645996, 0.01784524726867676, 0.01786675262451172, 0.01784012794494629, 0.017686527252197267, 0.017757183074951173, 0.01770086479187012, 0.017725439071655275, 0.01784320068359375, 0.017703935623168944, 0.017732608795166017, 0.01777663993835449, 0.017819648742675782, 0.01792527961730957, 0.017791839599609376, 0.01773465538024902, 0.01777561569213867, 0.017767520904541017, 0.017779615402221678, 0.017728511810302734, 0.017803455352783205, 0.017824575424194335, 0.017731584548950196, 0.017750015258789064, 0.017686527252197267, 0.017780736923217775, 0.01780735969543457, 0.01801625633239746, 0.017744895935058593, 0.01780633544921875, 0.017756160736083985, 0.01775324821472168, 0.017786720275878905, 0.017720319747924804, 0.01781452751159668, 0.017756160736083985, 0.017701887130737306, 0.017757183074951173, 0.017794048309326172, 0.017775680541992186, 0.01771820831298828, 0.017878015518188475, 0.017714176177978515, 0.017934335708618163, 0.01790771293640137, 0.01800396728515625, 0.01797324752807617, 0.017838079452514647, 0.017943552017211914, 0.017921024322509766, 0.018066432952880858, 0.01804287910461426, 0.017954816818237306, 0.018028543472290038, 0.017991680145263672, 0.018086912155151368, 0.01804902458190918, 0.04018483352661133, 0.019185663223266602, 0.018629663467407225, 0.018362335205078125, 0.018516992568969725, 0.01823744010925293, 0.018296831130981444, 0.01842995262145996, 0.018309152603149415, 0.018204639434814453, 0.017959936141967774, 0.017930240631103517, 0.017975296020507812, 0.01799679946899414, 0.017985536575317384, 0.018001920700073244, 0.017960960388183594, 0.018035711288452147, 0.01801420783996582, 0.01799577522277832, 0.01799679946899414, 0.017920000076293945, 0.01798963165283203, 0.017941503524780272, 0.01805721664428711, 0.018167808532714845, 0.018661376953125, 0.018552831649780274, 0.018554880142211915, 0.01859993553161621, 0.018526208877563476, 0.018514944076538087, 0.01855897521972656, 0.0184770565032959, 0.018308095932006836, 0.01848320007324219, 0.018584575653076172, 0.018572288513183592, 0.018563072204589845, 0.018593791961669923, 0.01863167953491211, 0.018544639587402344, 0.01861631965637207, 0.018153472900390624, 0.018067455291748045, 0.01807155227661133, 0.018028543472290038, 0.018010112762451173, 0.017986623764038086, 0.017970111846923827, 0.018001920700073244, 0.0180316162109375, 0.01801625633239746, 0.01801215934753418, 0.01804083251953125, 0.01799884796142578, 0.01804902458190918, 0.01800704002380371, 0.018043903350830077, 0.01805516815185547, 0.01801215934753418, 0.018124799728393554, 0.018512895584106445, 0.03970560073852539, 0.018001920700073244, 0.01805516815185547, 0.01804697608947754, 0.018001920700073244, 0.018035711288452147, 0.01800396728515625, 0.01803264045715332, 0.01801420783996582, 0.01801318359375, 0.017960960388183594, 0.017991680145263672, 0.018027519226074217, 0.017933311462402343, 0.01801625633239746, 0.01802444839477539, 0.018421760559082033, 0.01859891128540039, 0.018514944076538087, 0.018568191528320312, 0.017986560821533205, 0.018066432952880858, 0.01799884796142578, 0.018027519226074217, 0.01802342414855957, 0.01806540870666504, 0.018018304824829103, 0.018076671600341796, 0.018268159866333008, 0.018523136138916017, 0.018300928115844727, 0.018545663833618165, 0.018568191528320312, 0.01903206443786621, 0.01922867202758789, 0.019533824920654298, 0.018744319915771485, 0.01863167953491211, 0.018572288513183592, 0.018276351928710938, 0.018582592010498045, 0.01851897621154785, 0.018724863052368163, 0.018562047958374024, 0.018515968322753908, 0.018545663833618165, 0.01825279998779297, 0.018027519226074217, 0.018018304824829103, 0.018113536834716795, 0.018009088516235353, 0.01804287910461426, 0.017991680145263672, 0.018044927597045898, 0.01799782371520996, 0.018043903350830077, 0.01802444839477539, 0.018052095413208007, 0.018061311721801757, 0.01803264045715332, 0.01800396728515625, 0.017963008880615236, 0.017966079711914062]",tokens/s,53.05973594356727,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,togethercomputer/RedPajama-INCITE-Base-3B-v1,togethercomputer/RedPajama-INCITE-Base-3B-v1,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.42.3,,0.31.0,,,,1.20.0,,,,0.11.1,,,,MB,1276.674048,6791.102464,0.0,6144.65536,6020.360704,s,10,0.35853353118896486,0.03585335311889649,0.00038503052885981143,0.03573289489746094,0.035879467010498045,0.036442996978759766,0.03689382095336914,"[0.03700652694702149, 0.03573455810546875, 0.03571161651611328, 0.03568291091918945, 0.03575423812866211, 0.03569232177734375, 0.035738399505615234, 0.03573123168945312, 0.035730079650878904, 0.03575164794921875]",tokens/s,7140.196877850049,kWh,4.2245744062321525e-07,2.31488479508369e-07,2.225947217264292e-06,2.8798931373958762e-06,tokens/kWh,88892187.23979677,MB,1276.674048,6791.102464,0.0,6144.65536,6124.220416,s,10,13.198198608398439,1.3198198608398435,0.0117623360564721,1.3177906494140625,1.3273325927734374,1.3398837036132814,1.3499245922851564,"[1.352434814453125, 1.3171402587890626, 1.3097017822265624, 1.3184410400390625, 1.3097115478515624, 1.311541748046875, 1.32454345703125, 1.318809326171875, 1.3160390625, 1.3198355712890626]",tokens/s,47.733786912337486,kWh,1.5914132048095034e-05,8.720846506636083e-06,5.7757461880535354e-05,8.239244043526647e-05,tokens/kWh,764633.2560023806,,s,629,13.497218076705927,0.021458216338165234,0.004318344103743907,0.020807680130004884,0.02149314613342285,0.02197176284790039,0.05674004547119141,"[0.021798912048339843, 0.021969919204711915, 0.022676479339599608, 0.021760000228881835, 0.021797887802124022, 0.02169856071472168, 0.02163711929321289, 0.021791744232177734, 0.02284851264953613, 0.021805055618286134, 0.022370304107666016, 0.022420480728149415, 0.022611967086791994, 0.022486015319824217, 0.02271129608154297, 0.022368255615234374, 0.022510591506958007, 0.02231091117858887, 0.02262937545776367, 0.02229964828491211, 0.022326271057128907, 0.022561792373657227, 0.02214297676086426, 0.02191257667541504, 0.021569536209106444, 0.021498880386352538, 0.021012479782104493, 0.020791296005249024, 0.020798463821411133, 0.020758527755737305, 0.02085171127319336, 0.020702207565307617, 0.020790271759033203, 0.020808704376220705, 0.02085068893432617, 0.021104639053344726, 0.020928512573242186, 0.020816896438598635, 0.02149171257019043, 0.021355520248413085, 0.02089369583129883, 0.020817920684814452, 0.021572608947753907, 0.021349376678466796, 0.021098495483398438, 0.020814847946166993, 0.02087321662902832, 0.02085273551940918, 0.020956159591674805, 0.020946943283081054, 0.02087014389038086, 0.020785152435302736, 0.02086297607421875, 0.02086809539794922, 0.02081279945373535, 0.020754432678222655, 0.02066739273071289, 0.020695039749145508, 0.02083020782470703, 0.02086502456665039, 0.02086911964416504, 0.020985855102539062, 0.057055233001708984, 0.022586368560791017, 0.021505023956298826, 0.020794368743896483, 0.020815872192382814, 0.020964351654052735, 0.02084659194946289, 0.020777984619140624, 0.02084659194946289, 0.02082099151611328, 0.020876287460327148, 0.020779008865356444, 0.020783103942871094, 0.020764671325683593, 0.0208035831451416, 0.020760576248168947, 0.02081996726989746, 0.020815872192382814, 0.020806655883789063, 0.02084454345703125, 0.020985855102539062, 0.020971519470214844, 0.020727807998657227, 0.020918272018432618, 0.02079539108276367, 0.020786176681518553, 0.020858879089355468, 0.020706304550170897, 0.02088960075378418, 0.020781055450439453, 0.020744192123413087, 0.02082815933227539, 0.020771839141845702, 0.020753408432006838, 0.02081177520751953, 0.020785152435302736, 0.020818944931030273, 0.020726783752441406, 0.02087116813659668, 0.020800512313842775, 0.020807680130004884, 0.020741119384765624, 0.021073919296264648, 0.020763647079467772, 0.02086502456665039, 0.020773887634277344, 0.02090188789367676, 0.020817920684814452, 0.020762624740600585, 0.020776960372924806, 0.021785600662231445, 0.0214517765045166, 0.02088140869140625, 0.02082611274719238, 0.020839424133300782, 0.02090291213989258, 0.021540864944458008, 0.02091110420227051, 0.020771839141845702, 0.020805631637573242, 0.020786176681518553, 0.020779008865356444, 0.020935680389404295, 0.056717311859130856, 0.020732927322387695, 0.020757503509521484, 0.0206878719329834, 0.020742143630981445, 0.020766719818115235, 0.020914176940917968, 0.020775936126708985, 0.020736000061035157, 0.020791296005249024, 0.020737024307250978, 0.020770816802978515, 0.020780031204223632, 0.02091110420227051, 0.020754432678222655, 0.020730880737304686, 0.020732927322387695, 0.020740095138549804, 0.02124492835998535, 0.020855808258056642, 0.0208353271484375, 0.02082508850097656, 0.020702207565307617, 0.020767744064331056, 0.020781055450439453, 0.020768768310546876, 0.02085171127319336, 0.02071244812011719, 0.020607999801635742, 0.020579328536987306, 0.020555776596069338, 0.020576255798339844, 0.020840448379516603, 0.020780031204223632, 0.02129817581176758, 0.021410816192626952, 0.021185535430908203, 0.02068889617919922, 0.020743167877197266, 0.02083328056335449, 0.02065715217590332, 0.020740095138549804, 0.020720640182495118, 0.020770816802978515, 0.020785152435302736, 0.020732927322387695, 0.020736000061035157, 0.020770816802978515, 0.020757503509521484, 0.020716543197631835, 0.020762624740600585, 0.020746240615844725, 0.020695039749145508, 0.020723712921142577, 0.020774911880493165, 0.020773887634277344, 0.020790271759033203, 0.020709375381469726, 0.020724735260009765, 0.020772863388061523, 0.020765695571899414, 0.020762624740600585, 0.02085068893432617, 0.05674291229248047, 0.020822015762329102, 0.020727807998657227, 0.020734975814819336, 0.02067558479309082, 0.02102579116821289, 0.020921344757080077, 0.020700159072875975, 0.020802560806274413, 0.020738048553466795, 0.020775936126708985, 0.022574079513549804, 0.022766592025756836, 0.021801984786987305, 0.021498880386352538, 0.021212160110473634, 0.021575679779052736, 0.021283840179443358, 0.020747264862060546, 0.021123071670532227, 0.020797439575195312, 0.02087116813659668, 0.020715520858764647, 0.020785152435302736, 0.020738048553466795, 0.020776960372924806, 0.020742143630981445, 0.020771839141845702, 0.020737024307250978, 0.02084147262573242, 0.020776960372924806, 0.020806655883789063, 0.020769792556762694, 0.020791296005249024, 0.020814847946166993, 0.020808704376220705, 0.020988927841186524, 0.021370880126953123, 0.02149171257019043, 0.0208035831451416, 0.020760576248168947, 0.020771839141845702, 0.020723712921142577, 0.02071244812011719, 0.020734975814819336, 0.020752384185791017, 0.020695039749145508, 0.020782079696655274, 0.020703231811523438, 0.020742143630981445, 0.0208855037689209, 0.020747264862060546, 0.020719615936279297, 0.020719615936279297, 0.020561920166015626, 0.020669439315795898, 0.020858879089355468, 0.020753408432006838, 0.020747264862060546, 0.020921344757080077, 0.02106879997253418, 0.02084864044189453, 0.020824064254760744, 0.057398273468017576, 0.020781055450439453, 0.020762624740600585, 0.020719615936279297, 0.020772863388061523, 0.020798463821411133, 0.02082713508605957, 0.020800512313842775, 0.020789247512817383, 0.020807680130004884, 0.020815872192382814, 0.020799488067626954, 0.02067558479309082, 0.020677631378173827, 0.02079641532897949, 0.020715520858764647, 0.020685823440551757, 0.02087731170654297, 0.02068992042541504, 0.020783103942871094, 0.020808704376220705, 0.021004287719726563, 0.020684799194335936, 0.020728832244873048, 0.020731903076171874, 0.02068377685546875, 0.020730880737304686, 0.020722688674926756, 0.020742143630981445, 0.02072166442871094, 0.020701183319091796, 0.020756479263305663, 0.020791296005249024, 0.02072166442871094, 0.020748287200927733, 0.020716543197631835, 0.020730880737304686, 0.020761600494384767, 0.020771839141845702, 0.020736000061035157, 0.020822015762329102, 0.020770816802978515, 0.02067353630065918, 0.020754432678222655, 0.020839424133300782, 0.020755456924438476, 0.02070528030395508, 0.020935680389404295, 0.021129215240478515, 0.02079641532897949, 0.020744192123413087, 0.02085683250427246, 0.020786176681518553, 0.020759552001953126, 0.020787200927734374, 0.02082611274719238, 0.020815872192382814, 0.020763647079467772, 0.020810752868652343, 0.020792320251464845, 0.02084249687194824, 0.020784128189086915, 0.020773887634277344, 0.05680640029907227, 0.020822015762329102, 0.020724735260009765, 0.020742143630981445, 0.020674560546875, 0.020686847686767578, 0.020818944931030273, 0.020738048553466795, 0.021005311965942384, 0.020755456924438476, 0.02088960075378418, 0.020769792556762694, 0.020714496612548827, 0.020916223526000977, 0.020787200927734374, 0.02082611274719238, 0.020765695571899414, 0.02087731170654297, 0.020771839141845702, 0.020734975814819336, 0.020791296005249024, 0.020809728622436522, 0.020746240615844725, 0.020787200927734374, 0.020725759506225586, 0.020763647079467772, 0.020864000320434572, 0.02086195182800293, 0.020789247512817383, 0.020779008865356444, 0.020780031204223632, 0.020847616195678712, 0.021029888153076173, 0.020750335693359375, 0.020932607650756836, 0.02088243293762207, 0.020757503509521484, 0.020748287200927733, 0.02084454345703125, 0.020790271759033203, 0.020813823699951172, 0.02084864044189453, 0.020767744064331056, 0.021394432067871092, 0.021511167526245118, 0.020769792556762694, 0.020718591690063477, 0.020792320251464845, 0.02068172836303711, 0.020772863388061523, 0.02088960075378418, 0.020809728622436522, 0.02071347236633301, 0.02084249687194824, 0.02067967987060547, 0.020817920684814452, 0.02081177520751953, 0.020730880737304686, 0.020755456924438476, 0.020762624740600585, 0.020714496612548827, 0.020788223266601562, 0.020772863388061523, 0.05673267364501953, 0.020766719818115235, 0.020818944931030273, 0.020790271759033203, 0.02079641532897949, 0.020747264862060546, 0.02087424087524414, 0.02091007995605469, 0.020756479263305663, 0.020798463821411133, 0.020779008865356444, 0.02086502456665039, 0.020709375381469726, 0.02143539237976074, 0.021001216888427734, 0.02127257537841797, 0.020765695571899414, 0.02102374458312988, 0.02085273551940918, 0.020800512313842775, 0.02081279945373535, 0.020785152435302736, 0.020764671325683593, 0.020766719818115235, 0.020762624740600585, 0.020798463821411133, 0.02081996726989746, 0.0208855037689209, 0.020732927322387695, 0.020773887634277344, 0.02081279945373535, 0.021538816452026367, 0.0229171199798584, 0.02172313690185547, 0.02128998374938965, 0.021643264770507813, 0.02082508850097656, 0.020768768310546876, 0.02084147262573242, 0.02086809539794922, 0.020771839141845702, 0.020790271759033203, 0.020732927322387695, 0.02081279945373535, 0.02090188789367676, 0.020788223266601562, 0.020802560806274413, 0.02084659194946289, 0.020736000061035157, 0.020782079696655274, 0.020791296005249024, 0.020787200927734374, 0.02079539108276367, 0.02084351921081543, 0.020752384185791017, 0.020727807998657227, 0.02088960075378418, 0.022364160537719727, 0.023554048538208007, 0.021968896865844727, 0.021498880386352538, 0.02104729652404785, 0.02168217658996582, 0.058259456634521485, 0.021972991943359374, 0.021168127059936523, 0.020988927841186524, 0.02092748832702637, 0.02081996726989746, 0.020929536819458007, 0.020779008865356444, 0.020888576507568358, 0.020785152435302736, 0.020883455276489257, 0.02084249687194824, 0.020909055709838868, 0.020891647338867187, 0.021147647857666017, 0.021514240264892577, 0.021006336212158205, 0.02087424087524414, 0.020954111099243163, 0.02082918357849121, 0.02082815933227539, 0.020790271759033203, 0.020802560806274413, 0.0210831356048584, 0.02147532844543457, 0.020748287200927733, 0.020810752868652343, 0.020758527755737305, 0.020749311447143554, 0.020793344497680662, 0.020766719818115235, 0.020802560806274413, 0.02088652801513672, 0.02109951972961426, 0.02088243293762207, 0.020807680130004884, 0.020831232070922853, 0.020944896697998046, 0.020883455276489257, 0.020787200927734374, 0.020921344757080077, 0.021752832412719726, 0.02109644889831543, 0.020792320251464845, 0.02088140869140625, 0.020939775466918945, 0.020765695571899414, 0.020740095138549804, 0.020814847946166993, 0.020754432678222655, 0.020744192123413087, 0.020745216369628908, 0.02082815933227539, 0.020746240615844725, 0.020780031204223632, 0.02087833595275879, 0.020888576507568358, 0.020747264862060546, 0.020808704376220705, 0.020766719818115235, 0.020722688674926756, 0.020798463821411133, 0.020785152435302736, 0.056817665100097656, 0.020720640182495118, 0.020801536560058592, 0.020746240615844725, 0.020692991256713866, 0.020763647079467772, 0.02102272033691406, 0.020743167877197266, 0.02071347236633301, 0.020755456924438476, 0.020807680130004884, 0.02105958366394043, 0.020792320251464845, 0.02084249687194824, 0.020839424133300782, 0.020748287200927733, 0.020794368743896483, 0.020753408432006838, 0.020797439575195312, 0.02084556770324707, 0.02082918357849121, 0.02082508850097656, 0.020896768569946288, 0.02079539108276367, 0.020783103942871094, 0.020798463821411133, 0.02086195182800293, 0.022781951904296875, 0.02108518409729004, 0.020921344757080077, 0.020896768569946288, 0.021158912658691405, 0.021013504028320314, 0.02085785675048828, 0.020797439575195312, 0.02084351921081543, 0.020776960372924806, 0.02087833595275879, 0.020831232070922853, 0.020840448379516603, 0.02085478401184082, 0.020855808258056642, 0.020815872192382814, 0.020792320251464845, 0.02079641532897949, 0.02088243293762207, 0.020806655883789063, 0.0212992000579834, 0.021102592468261717, 0.020983808517456053, 0.02105036735534668, 0.021021696090698243, 0.020831232070922853, 0.02083737564086914, 0.020765695571899414, 0.020793344497680662, 0.020964351654052735, 0.02082099151611328, 0.020760576248168947, 0.020772863388061523, 0.02085478401184082, 0.020816896438598635, 0.020790271759033203, 0.057801727294921876, 0.02152038383483887, 0.021441535949707033, 0.02085785675048828, 0.02083020782470703, 0.021572608947753907, 0.02143539237976074, 0.020747264862060546, 0.02082713508605957, 0.020785152435302736, 0.02166169548034668, 0.021557247161865235, 0.021007360458374022, 0.020858879089355468, 0.02129817581176758, 0.021175296783447265, 0.020792320251464845, 0.020822015762329102, 0.02085273551940918, 0.02127462387084961, 0.020807680130004884, 0.020788223266601562, 0.020806655883789063, 0.020786176681518553, 0.02081996726989746, 0.020760576248168947, 0.020922367095947265, 0.020770816802978515, 0.020753408432006838, 0.020896768569946288, 0.020755456924438476, 0.020939775466918945, 0.02082099151611328, 0.020745216369628908, 0.02146303939819336, 0.020947967529296875, 0.020753408432006838, 0.020790271759033203, 0.02089472007751465, 0.02082918357849121, 0.020736000061035157, 0.020790271759033203, 0.020763647079467772, 0.02089779281616211, 0.020876287460327148, 0.02082815933227539, 0.020775936126708985, 0.020782079696655274, 0.020875263214111327, 0.020790271759033203, 0.02084454345703125, 0.02084249687194824, 0.02083737564086914, 0.02083635139465332, 0.02081279945373535, 0.020760576248168947, 0.02081996726989746, 0.02086502456665039, 0.02087321662902832, 0.020933631896972657, 0.02085683250427246, 0.020800512313842775, 0.02124083137512207]",tokens/s,46.602195832158536,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,meta-llama/Meta-Llama-3-8B,meta-llama/Meta-Llama-3-8B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.1,,,,MB,1280.335872,17201.364992,0.0,16554.917888,16436.726784,s,10,0.844628547668457,0.0844628547668457,0.00027184801111277817,0.08435808181762695,0.0848109161376953,0.08496329650878906,0.08508520080566406,"[0.0851156768798828, 0.08429670715332031, 0.08441200256347656, 0.08430416107177735, 0.0847770538330078, 0.08445945739746094, 0.08418342590332031, 0.08430364990234375, 0.08423961639404297, 0.08453679656982421]",tokens/s,3030.918155758192,kWh,9.976038147532751e-07,5.466216842923582e-07,6.793572194890711e-06,8.337797693936345e-06,tokens/kWh,30703551.392974637,MB,1280.335872,17201.364992,0.0,16554.917888,16436.729344,s,10,24.64940209960937,2.464940209960937,0.003371519870193411,2.4638629150390625,2.4660568603515625,2.4704556762695313,2.4739747290039062,"[2.4748544921875, 2.462900634765625, 2.46427392578125, 2.4631435546875, 2.4639423828125, 2.46314697265625, 2.465079345703125, 2.463783447265625, 2.464727294921875, 2.463550048828125]",tokens/s,25.55842926551082,kWh,2.904267110994884e-05,1.5916483305214613e-05,0.00012930045334690903,0.00017425960776207248,tokens/kWh,361529.56390225457,,s,629,25.368834106445345,0.04033200970818015,0.010102595982973209,0.03909427261352539,0.03917230072021484,0.03930132446289063,0.12401680328369141,"[0.03990528106689453, 0.03996979141235352, 0.04014284896850586, 0.0399288330078125, 0.04009881591796875, 0.039877632141113284, 0.03989811325073242, 0.039782398223876955, 0.03975680160522461, 0.03952230453491211, 0.03942911911010742, 0.03965849685668945, 0.039367679595947266, 0.03927763366699219, 0.039143360137939454, 0.03904512023925781, 0.0390748176574707, 0.03906150436401367, 0.03906867218017578, 0.03902566528320312, 0.03908812713623047, 0.039057407379150394, 0.03909836959838867, 0.03905535888671875, 0.03905331039428711, 0.03905539321899414, 0.039225311279296876, 0.039367679595947266, 0.03934310531616211, 0.039137279510498044, 0.03910758590698242, 0.03907583999633789, 0.03911679840087891, 0.039067649841308595, 0.039103488922119144, 0.03908095932006836, 0.03912908935546875, 0.03916595077514649, 0.03909836959838867, 0.039076862335205076, 0.03912704086303711, 0.03911990356445313, 0.03911676788330078, 0.039070720672607424, 0.039180286407470705, 0.03911475372314453, 0.039155712127685545, 0.03907993698120117, 0.039160831451416016, 0.03913523101806641, 0.03912908935546875, 0.039137279510498044, 0.03913113784790039, 0.03912192153930664, 0.03921408081054688, 0.03915059280395508, 0.039109630584716795, 0.03909632110595703, 0.039175167083740234, 0.03913216018676758, 0.03917004776000976, 0.03909427261352539, 0.12401766204833985, 0.03908512115478516, 0.03904095840454101, 0.039106559753417966, 0.03906662368774414, 0.03907891082763672, 0.03914035034179687, 0.03908915328979492, 0.039037952423095705, 0.03905228805541992, 0.039034881591796876, 0.039021568298339845, 0.039034881591796876, 0.03903692626953125, 0.039008255004882815, 0.03905228805541992, 0.03905228805541992, 0.03905331039428711, 0.03903590393066406, 0.03901542282104492, 0.03912908935546875, 0.0391464958190918, 0.03907174301147461, 0.03902975845336914, 0.039060478210449216, 0.03902566528320312, 0.039037952423095705, 0.03902668762207031, 0.039070720672607424, 0.03903385543823242, 0.039032833099365234, 0.03905228805541992, 0.039109630584716795, 0.03914035034179687, 0.03908201599121094, 0.03910857772827148, 0.03904307174682617, 0.03910246276855469, 0.039065601348876954, 0.03907276916503906, 0.039057407379150394, 0.03909939193725586, 0.039142398834228515, 0.039106559753417966, 0.03911884689331055, 0.03914448165893555, 0.03907990264892578, 0.03911372756958008, 0.03910246276855469, 0.03912908935546875, 0.03911679840087891, 0.03913523101806641, 0.03910553741455078, 0.039119873046875, 0.03909529495239258, 0.03916185760498047, 0.03908607864379883, 0.03914854431152344, 0.039180286407470705, 0.03915673446655273, 0.03913113784790039, 0.03908915328979492, 0.039103488922119144, 0.12410163116455078, 0.03906150436401367, 0.039109630584716795, 0.03908812713623047, 0.039037952423095705, 0.03909120178222656, 0.039037952423095705, 0.03905023956298828, 0.03900928115844727, 0.039060478210449216, 0.03913011169433594, 0.03923052978515625, 0.03902355194091797, 0.039041023254394534, 0.03904000091552735, 0.03905535888671875, 0.039000064849853515, 0.03905843353271484, 0.03909836959838867, 0.039090175628662106, 0.03904000091552735, 0.03904716873168945, 0.039065601348876954, 0.03911372756958008, 0.03906662368774414, 0.039056385040283206, 0.03907993698120117, 0.039057407379150394, 0.03905945587158203, 0.03905535888671875, 0.039056385040283206, 0.03909222412109375, 0.0390748176574707, 0.03908095932006836, 0.03908508682250977, 0.03909014511108398, 0.03906662368774414, 0.03913318252563477, 0.03910553741455078, 0.03908915328979492, 0.03909641647338867, 0.03911670303344727, 0.03911782455444336, 0.039174144744873046, 0.039433216094970705, 0.039223297119140625, 0.03909427261352539, 0.039106559753417966, 0.039128063201904296, 0.0391383056640625, 0.03911679840087891, 0.03913216018676758, 0.039174144744873046, 0.03930828857421875, 0.03908505630493164, 0.039172096252441405, 0.03909939193725586, 0.0391014404296875, 0.03910246276855469, 0.03915776062011719, 0.03912908935546875, 0.03929087829589844, 0.03928575897216797, 0.12413337707519531, 0.03911167907714844, 0.039065601348876954, 0.03910451126098633, 0.03904819107055664, 0.0391833610534668, 0.03905945587158203, 0.039037952423095705, 0.039047199249267577, 0.039048160552978516, 0.03905023956298828, 0.03912704086303711, 0.039065601348876954, 0.039051265716552735, 0.039076862335205076, 0.03901337432861328, 0.039024639129638675, 0.03910451126098633, 0.03903180694580078, 0.03902771377563476, 0.03903897476196289, 0.039046142578125, 0.03902771377563476, 0.03908403015136719, 0.039024639129638675, 0.039067649841308595, 0.039054336547851565, 0.03908607864379883, 0.03903590393066406, 0.03914956665039063, 0.03910553741455078, 0.03913011169433594, 0.03905331039428711, 0.03909427261352539, 0.039019519805908204, 0.03907891082763672, 0.03904716873168945, 0.03914035034179687, 0.03907788848876953, 0.03908505630493164, 0.039067649841308595, 0.039137279510498044, 0.03912089538574219, 0.03911884689331055, 0.03911679840087891, 0.03909939193725586, 0.03906867218017578, 0.03915161514282227, 0.03909024047851563, 0.03910956954956055, 0.03909836959838867, 0.039144447326660156, 0.03908403015136719, 0.03912499237060547, 0.03910758590698242, 0.03918643188476562, 0.039122943878173826, 0.03911167907714844, 0.03908505630493164, 0.03917311859130859, 0.03908505630493164, 0.039136257171630856, 0.039093246459960936, 0.12486348724365234, 0.03909939193725586, 0.039051265716552735, 0.03912089538574219, 0.03907276916503906, 0.03915776062011719, 0.0390645751953125, 0.03907276916503906, 0.03904819107055664, 0.039032833099365234, 0.039049217224121094, 0.03904409790039062, 0.039057407379150394, 0.03902668762207031, 0.03903692626953125, 0.03911782455444336, 0.03908403015136719, 0.03904409790039062, 0.03902771377563476, 0.039021568298339845, 0.03906355285644531, 0.03907891082763672, 0.03915980911254883, 0.039258113861083986, 0.039229438781738284, 0.03909632110595703, 0.039054336547851565, 0.03905228805541992, 0.039090175628662106, 0.03906355285644531, 0.03909836959838867, 0.03908812713623047, 0.03905331039428711, 0.03908607864379883, 0.03907583999633789, 0.03907891082763672, 0.039106559753417966, 0.03910553741455078, 0.03912192153930664, 0.039090175628662106, 0.03906662368774414, 0.03911888122558594, 0.03916489410400391, 0.0391383056640625, 0.03906252670288086, 0.039112735748291015, 0.03906867218017578, 0.03912393569946289, 0.039090175628662106, 0.03912089538574219, 0.03911167907714844, 0.03913318252563477, 0.03909734344482422, 0.03912396621704101, 0.03916185760498047, 0.03916799926757813, 0.0391014404296875, 0.03915980911254883, 0.03909222412109375, 0.03912908935546875, 0.03908198547363281, 0.03911475372314453, 0.03906768035888672, 0.12417839813232422, 0.03909427261352539, 0.039057407379150394, 0.03910553741455078, 0.03906355285644531, 0.039087104797363284, 0.03906662368774414, 0.03903692626953125, 0.039049217224121094, 0.039037952423095705, 0.03904716873168945, 0.03903897476196289, 0.03906355285644531, 0.039019519805908204, 0.03905228805541992, 0.0390748176574707, 0.03904307174682617, 0.039034881591796876, 0.03900928115844727, 0.03902771377563476, 0.03904307174682617, 0.039041023254394534, 0.03903590393066406, 0.03905535888671875, 0.03902361679077149, 0.039057407379150394, 0.039057407379150394, 0.039158782958984374, 0.039122943878173826, 0.03908198547363281, 0.039034881591796876, 0.039109630584716795, 0.03903385543823242, 0.039099456787109375, 0.039056320190429684, 0.039091232299804685, 0.039026657104492185, 0.03908505630493164, 0.03913216018676758, 0.03907891082763672, 0.039060478210449216, 0.0391383056640625, 0.03913523101806641, 0.03909222412109375, 0.03915059280395508, 0.03912396621704101, 0.03909836959838867, 0.039147518157958985, 0.03910451126098633, 0.03917004776000976, 0.039139328002929685, 0.039223358154296874, 0.03908806228637695, 0.039206912994384766, 0.039147518157958985, 0.03919257736206055, 0.039122943878173826, 0.039093246459960936, 0.0391014404296875, 0.03915059280395508, 0.03909120178222656, 0.03913011169433594, 0.03910758590698242, 0.12389785766601563, 0.03908812713623047, 0.03907276916503906, 0.03909734344482422, 0.03902873611450195, 0.039106559753417966, 0.039158782958984374, 0.03909734344482422, 0.038980609893798826, 0.03901235198974609, 0.039019519805908204, 0.039018497467041016, 0.03899907302856445, 0.03916387176513672, 0.03909120178222656, 0.03908607864379883, 0.03905228805541992, 0.03902873611450195, 0.039008255004882815, 0.03902361679077149, 0.03909427261352539, 0.039057407379150394, 0.039065601348876954, 0.03904822540283203, 0.039045089721679686, 0.03904512023925781, 0.03910246276855469, 0.03910041427612305, 0.03908198547363281, 0.03909939193725586, 0.03902259063720703, 0.03907174301147461, 0.039103488922119144, 0.03909939193725586, 0.03908403015136719, 0.039142398834228515, 0.03903798294067383, 0.039097312927246095, 0.03911782455444336, 0.03915161514282227, 0.03907276916503906, 0.039103488922119144, 0.03914342498779297, 0.039119873046875, 0.03911372756958008, 0.03910556793212891, 0.03914342498779297, 0.03918844985961914, 0.039314430236816404, 0.03913420867919922, 0.03914041519165039, 0.039142337799072266, 0.03908095932006836, 0.03913011169433594, 0.03906867218017578, 0.03916595077514649, 0.03947212982177734, 0.04023295974731445, 0.0392806396484375, 0.03911577606201172, 0.0390830078125, 0.039136257171630856, 0.03911782455444336, 0.12407193756103516, 0.03911270523071289, 0.0390645751953125, 0.03906252670288086, 0.03905843353271484, 0.03907788848876953, 0.039019519805908204, 0.03900931167602539, 0.039083999633789064, 0.03905945587158203, 0.03903692626953125, 0.03903897476196289, 0.03905228805541992, 0.03906870269775391, 0.03905737686157226, 0.03906355285644531, 0.03913420867919922, 0.03906969451904297, 0.03901235198974609, 0.03903286361694336, 0.03903689575195313, 0.039070720672607424, 0.03903078460693359, 0.039069759368896485, 0.03907577514648437, 0.039070720672607424, 0.03905535888671875, 0.03904716873168945, 0.03901235198974609, 0.0390830078125, 0.03925094223022461, 0.03922742462158203, 0.03906044769287109, 0.03914547348022461, 0.03908505630493164, 0.03911065673828125, 0.039057407379150394, 0.03911679840087891, 0.03910451126098633, 0.03916287994384766, 0.03905228805541992, 0.039158782958984374, 0.039090175628662106, 0.03912499237060547, 0.03908403015136719, 0.03911782455444336, 0.03905023956298828, 0.03914342498779297, 0.03914137649536133, 0.039126014709472655, 0.03909632110595703, 0.03913011169433594, 0.0391536636352539, 0.039160831451416016, 0.03912396621704101, 0.039174144744873046, 0.03907078552246094, 0.03912492752075195, 0.039147518157958985, 0.03932160186767578, 0.039128063201904296, 0.03919462585449219, 0.0391383056640625, 0.12401459503173828, 0.03906355285644531, 0.03907379150390625, 0.03907993698120117, 0.03906355285644531, 0.03908607864379883, 0.03907583999633789, 0.03904924774169922, 0.03911062240600586, 0.03904716873168945, 0.03909120178222656, 0.0391536636352539, 0.039121982574462894, 0.03906758499145508, 0.03904307174682617, 0.03909734344482422, 0.03906867218017578, 0.03915776062011719, 0.03954995346069336, 0.03940966415405273, 0.03911475372314453, 0.03915980911254883, 0.03906252670288086, 0.0391014404296875, 0.03906355285644531, 0.03909427261352539, 0.039054336547851565, 0.03906867218017578, 0.03910246276855469, 0.0390748176574707, 0.03904819107055664, 0.03909737777709961, 0.039092193603515624, 0.03908505630493164, 0.03910553741455078, 0.03909836959838867, 0.03911475372314453, 0.03909427261352539, 0.039106559753417966, 0.03906867218017578, 0.03907379150390625, 0.03913011169433594, 0.03911679840087891, 0.03917926406860352, 0.03909939193725586, 0.03911065673828125, 0.03907788848876953, 0.03909529495239258, 0.03911065673828125, 0.0391383056640625, 0.039172096252441405, 0.0391383056640625, 0.03911065673828125, 0.03912089538574219, 0.03913216018676758, 0.03917107009887695, 0.03908198547363281, 0.03906969451904297, 0.039093246459960936, 0.03916799926757813, 0.03907993698120117, 0.039155712127685545, 0.03913318252563477, 0.12430233764648438, 0.03908812713623047, 0.03913011169433594, 0.0390748176574707, 0.03912192153930664, 0.039119873046875, 0.03910758590698242, 0.039016448974609375, 0.03904716873168945, 0.03902566528320312, 0.039122943878173826, 0.03906355285644531, 0.03905228805541992, 0.03904819107055664, 0.039057407379150394, 0.03903180694580078, 0.03902771377563476, 0.03909734344482422, 0.039103488922119144, 0.03904716873168945, 0.03902774429321289, 0.03906864166259766, 0.03902361679077149, 0.039119873046875, 0.03907788848876953, 0.03908812713623047, 0.03904307174682617, 0.03908505630493164, 0.03901030349731445, 0.039070720672607424, 0.039005184173583986, 0.039103488922119144, 0.039054336547851565, 0.0390830078125, 0.03904307174682617, 0.0390984001159668, 0.039075809478759764, 0.03912089538574219, 0.03908915328979492, 0.0390830078125, 0.03906252670288086, 0.039144447326660156, 0.03911577606201172, 0.039109630584716795, 0.03910758590698242, 0.0391383056640625, 0.0390748176574707, 0.03915776062011719, 0.03918438339233398, 0.03923455810546875, 0.03913523101806641, 0.0391536636352539, 0.03912089538574219, 0.0391464958190918, 0.039172096252441405, 0.039155712127685545, 0.03907993698120117, 0.039087104797363284, 0.03909222412109375, 0.03917824172973633, 0.039109630584716795, 0.039175167083740234, 0.039142398834228515]",tokens/s,24.794202104865104,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen2-beta-14B,Qwen/Qwen2-beta-14B,cuda,0,42,,,True,,,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",qwen2,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-32B,Qwen/Qwen1.5-32B,cuda,0,42,,,True,,,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",qwen2,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,0,0,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/0/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669325a2-70c920ed6ed05a727ca34b44;d829c062-d7ec-49fc-a457-565c97992f2a) Repository Not Found for url: https://huggingface.co/0/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: 0 is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,2,2,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/2/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669324d5-765ffb613906523a5a0b117f;5be0db35-8853-4fb2-8d54-7dae8098fa60) Repository Not Found for url: https://huggingface.co/2/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: 2 is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/xglm-564M,facebook/xglm-564M,cuda,0,42,,,True,,,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",xglm,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-14B,Qwen/Qwen-14B,cuda,0,42,,,True,,,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",qwen,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,huggyllama/llama-7b,huggyllama/llama-7b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.1,,,,MB,1219.93216,14529.593344,0.0,13883.14624,13705.186304,s,10,0.9362810974121093,0.09362810974121094,0.00027920329109976734,0.09354251098632813,0.09373921356201172,0.09408875732421876,0.09436839233398438,"[0.09443830108642579, 0.0934741439819336, 0.09354873657226563, 0.09362611389160157, 0.093427490234375, 0.09344755554199219, 0.09353238677978516, 0.09353628540039062, 0.09366153717041016, 0.09358854675292969]",tokens/s,2734.221599769414,kWh,1.1074903613557943e-06,6.068454596681698e-07,6.0343921587663614e-06,7.748727979790327e-06,tokens/kWh,33037680.593212295,MB,1219.93216,14529.593344,0.0,13883.14624,13856.22528,s,10,22.526626953125,2.2526626953125,0.004122940525000741,2.2514234619140625,2.2538315185546876,2.259308093261719,2.263689353027344,"[2.26478466796875, 2.250800537109375, 2.25008447265625, 2.25150439453125, 2.251889404296875, 2.25090234375, 2.25018798828125, 2.25251611328125, 2.252614501953125, 2.251342529296875]",tokens/s,27.966903403290186,kWh,2.6531057572408135e-05,1.4539835276311148e-05,0.00011534956605943347,0.0001564204589081527,tokens/kWh,402760.61353964236,,s,629,23.331680217742928,0.03709329128416999,0.011191790459388166,0.035729408264160156,0.0358287353515625,0.03590696868896485,0.12985827392578125,"[0.03639910507202149, 0.03638579177856445, 0.03628134536743164, 0.03638476943969727, 0.036380672454833986, 0.03639807891845703, 0.03640217590332031, 0.03637247848510742, 0.03649638366699219, 0.03633049774169922, 0.036241409301757815, 0.03701760101318359, 0.035983360290527344, 0.03586764907836914, 0.03579596710205078, 0.035767295837402346, 0.035748863220214845, 0.03577241516113281, 0.03568947219848633, 0.03576422500610352, 0.035844097137451174, 0.035724288940429685, 0.03568947219848633, 0.0357212142944336, 0.035784702301025394, 0.03567923355102539, 0.035740673065185545, 0.03583795166015625, 0.03578777694702148, 0.03602841567993164, 0.036229118347167966, 0.035740673065185545, 0.035727359771728515, 0.03571916961669922, 0.0357386245727539, 0.03573452758789063, 0.035740673065185545, 0.03571814346313477, 0.035783679962158206, 0.035833854675292966, 0.03577036666870117, 0.03577241516113281, 0.03581235122680664, 0.035760128021240234, 0.035773441314697264, 0.03577958297729492, 0.035745792388916016, 0.035778560638427735, 0.03579084777832031, 0.035922943115234376, 0.036125694274902344, 0.036101119995117184, 0.035953662872314454, 0.03583180618286133, 0.035846145629882815, 0.03582668685913086, 0.03582156753540039, 0.035800064086914066, 0.03576115036010742, 0.0358021125793457, 0.035784702301025394, 0.035850238800048825, 0.129902587890625, 0.03562393569946289, 0.03563622283935547, 0.03566592025756836, 0.03564031982421875, 0.0356383056640625, 0.0356259536743164, 0.03567103958129883, 0.035714046478271484, 0.035672065734863284, 0.03563827133178711, 0.03567001724243164, 0.03570175933837891, 0.03565363311767578, 0.035683326721191407, 0.03567103958129883, 0.03567103958129883, 0.035661823272705076, 0.035659774780273434, 0.03567001724243164, 0.03573555374145508, 0.03570278549194336, 0.035700767517089844, 0.03569353485107422, 0.03562700653076172, 0.035659774780273434, 0.03570892715454101, 0.035694625854492186, 0.03568022537231445, 0.03567411041259766, 0.03571916961669922, 0.035716094970703126, 0.03577036666870117, 0.0357386245727539, 0.035705856323242184, 0.035729408264160156, 0.03571712112426758, 0.03573452758789063, 0.03570175933837891, 0.03568742370605469, 0.0357212142944336, 0.03570687866210937, 0.035745792388916016, 0.03571200180053711, 0.035683326721191407, 0.035727359771728515, 0.03578777694702148, 0.035768318176269534, 0.03580416107177734, 0.03587583923339844, 0.03578572845458984, 0.035773441314697264, 0.03578572845458984, 0.035745792388916016, 0.035762176513671876, 0.03582259368896484, 0.03579289627075195, 0.035776512145996094, 0.035762176513671876, 0.03575296020507813, 0.03580416107177734, 0.03581644821166992, 0.03585331344604492, 0.12987289428710938, 0.035609600067138675, 0.03558707046508789, 0.03564134216308594, 0.035620864868164064, 0.0356577262878418, 0.035622913360595705, 0.03563724899291992, 0.03561369705200195, 0.035800064086914066, 0.03600588989257812, 0.035783679962158206, 0.03567308807373047, 0.03566387176513672, 0.035619873046875, 0.035608543395996096, 0.03562496185302735, 0.03563417434692383, 0.03565465545654297, 0.03563417434692383, 0.03565465545654297, 0.035676158905029294, 0.03569356918334961, 0.03570175933837891, 0.03565158462524414, 0.035676158905029294, 0.03570175933837891, 0.0357283821105957, 0.03566694259643555, 0.03568851089477539, 0.035676097869873045, 0.03569561767578125, 0.035678207397460936, 0.03566694259643555, 0.03567411041259766, 0.035707904815673826, 0.035683326721191407, 0.03567411041259766, 0.03568947219848633, 0.035714046478271484, 0.035729408264160156, 0.03577446365356445, 0.035716094970703126, 0.03572326278686523, 0.0357130241394043, 0.03578777694702148, 0.035737598419189456, 0.03577036666870117, 0.035737598419189456, 0.03573964691162109, 0.03578675079345703, 0.03579091262817383, 0.035752895355224606, 0.03571916961669922, 0.03572844696044922, 0.03578156661987305, 0.03572633743286133, 0.035760128021240234, 0.03573657608032227, 0.03574272155761719, 0.03576934432983398, 0.03578777694702148, 0.03585331344604492, 0.1300305938720703, 0.03561983871459961, 0.03560857772827149, 0.03559632110595703, 0.035631103515625, 0.03564028930664063, 0.03567308807373047, 0.0357498893737793, 0.03584000015258789, 0.03570380783081055, 0.0357130241394043, 0.03570483016967774, 0.03568742370605469, 0.03564652633666992, 0.03566483306884766, 0.03566796875, 0.03568230438232422, 0.035676158905029294, 0.03566796875, 0.035691520690917966, 0.03570073699951172, 0.03570892715454101, 0.03571712112426758, 0.03568537521362305, 0.035659774780273434, 0.0356864013671875, 0.0357283821105957, 0.035697662353515625, 0.035678207397460936, 0.035707904815673826, 0.035743743896484374, 0.03578060913085938, 0.035716094970703126, 0.035722240447998044, 0.03573452758789063, 0.035740673065185545, 0.03575091171264649, 0.035730430603027344, 0.035714046478271484, 0.03570995330810547, 0.03575302505493164, 0.03575392150878906, 0.03573964691162109, 0.03570892715454101, 0.035781631469726564, 0.03574070358276367, 0.03580105590820312, 0.03578060913085938, 0.03577753448486328, 0.03578879928588867, 0.035789825439453124, 0.035806209564208984, 0.03579391860961914, 0.03576319885253906, 0.035773441314697264, 0.035784736633300784, 0.03578979110717773, 0.03577036666870117, 0.03578572845458984, 0.03580416107177734, 0.0358205451965332, 0.03585539245605469, 0.03582255935668945, 0.1298206787109375, 0.03562700653076172, 0.035604480743408204, 0.03560246276855469, 0.03565052795410156, 0.035643390655517575, 0.035664894104003905, 0.03566284942626953, 0.03566284942626953, 0.035664894104003905, 0.03566592025756836, 0.03564646530151367, 0.03564851379394531, 0.03564646530151367, 0.03565363311767578, 0.03565875244140625, 0.03564851379394531, 0.035743743896484374, 0.03563827133178711, 0.035691520690917966, 0.03563827133178711, 0.03568947219848633, 0.03566694259643555, 0.035746814727783204, 0.03581235122680664, 0.03571916961669922, 0.03567923355102539, 0.03571814346313477, 0.03576838302612305, 0.035703742980957034, 0.03570691299438476, 0.03571503829956055, 0.03571200180053711, 0.03575500869750976, 0.03575603103637695, 0.03575603103637695, 0.03574169540405273, 0.035737598419189456, 0.03575091171264649, 0.03579391860961914, 0.035746814727783204, 0.03571507263183594, 0.035705856323242184, 0.03575603103637695, 0.03575193786621094, 0.03578371047973633, 0.035756000518798826, 0.03585126495361328, 0.03609292984008789, 0.03587583923339844, 0.03577241516113281, 0.03575296020507813, 0.03578675079345703, 0.03575398254394531, 0.03580825424194336, 0.03581235122680664, 0.03579084777832031, 0.03580416107177734, 0.035794944763183595, 0.03582463836669922, 0.03589324951171875, 0.03588915252685547, 0.03580518341064453, 0.129817626953125, 0.03563520050048828, 0.03562902450561523, 0.03563212966918945, 0.035625984191894534, 0.03565363311767578, 0.03564031982421875, 0.03561369705200195, 0.035622913360595705, 0.03567411041259766, 0.035643390655517575, 0.035697662353515625, 0.03565260696411133, 0.035676158905029294, 0.035659774780273434, 0.03565465545654297, 0.03566284942626953, 0.03568947219848633, 0.03565875244140625, 0.03564646530151367, 0.03564441680908203, 0.03569356918334961, 0.035743743896484374, 0.035675136566162106, 0.03567718505859375, 0.03572326278686523, 0.03571814346313477, 0.035762176513671876, 0.035724288940429685, 0.03576422500610352, 0.035730430603027344, 0.03575091171264649, 0.03569868850708008, 0.03570175933837891, 0.03568435287475586, 0.035699710845947266, 0.035722240447998044, 0.03571507263183594, 0.035773441314697264, 0.0357314567565918, 0.03573555374145508, 0.03577753448486328, 0.03573964691162109, 0.03574480056762695, 0.03575190353393555, 0.03578060913085938, 0.0358021125793457, 0.03575091171264649, 0.035748863220214845, 0.035789825439453124, 0.03577241516113281, 0.03582361602783203, 0.035783679962158206, 0.035737598419189456, 0.03575296020507813, 0.035775489807128906, 0.03577036666870117, 0.035800064086914066, 0.03578675079345703, 0.03580723190307617, 0.035811328887939455, 0.03583283233642578, 0.03580723190307617, 0.1299752960205078, 0.03566796875, 0.03555942535400391, 0.03559628677368164, 0.03561983871459961, 0.03561881637573242, 0.035609600067138675, 0.03561983871459961, 0.035642368316650394, 0.03577958297729492, 0.03565465545654297, 0.03566592025756836, 0.03563315200805664, 0.035675136566162106, 0.03568537521362305, 0.03563520050048828, 0.0356577262878418, 0.03563315200805664, 0.03569561767578125, 0.03570483016967774, 0.03565465545654297, 0.035688449859619144, 0.035642368316650394, 0.035664894104003905, 0.03561983871459961, 0.03562393569946289, 0.03568025588989258, 0.035714046478271484, 0.03578675079345703, 0.035724288940429685, 0.035727359771728515, 0.0357130241394043, 0.035676193237304685, 0.0357611198425293, 0.03567718505859375, 0.035707904815673826, 0.03570483016967774, 0.03570995330810547, 0.035699710845947266, 0.0356864013671875, 0.0357283821105957, 0.03577964782714844, 0.035733440399169925, 0.035697662353515625, 0.03579391860961914, 0.03578060913085938, 0.035740673065185545, 0.03571814346313477, 0.03571712112426758, 0.035789825439453124, 0.03588608169555664, 0.03582156753540039, 0.03573350524902344, 0.0357386245727539, 0.035722240447998044, 0.035745792388916016, 0.035746814727783204, 0.035740673065185545, 0.035740673065185545, 0.035783679962158206, 0.03581644821166992, 0.03581951904296875, 0.035765247344970705, 0.1299752960205078, 0.03565875244140625, 0.03561369705200195, 0.03566796875, 0.03568947219848633, 0.03567308807373047, 0.03566899108886719, 0.035661823272705076, 0.03564441680908203, 0.03567411041259766, 0.035729408264160156, 0.035699710845947266, 0.0357130241394043, 0.03568537521362305, 0.03566796875, 0.03565363311767578, 0.03571814346313477, 0.03571200180053711, 0.0356802864074707, 0.03573040008544922, 0.035691520690917966, 0.03567923355102539, 0.035681278228759765, 0.0356864013671875, 0.03569359970092773, 0.035773406982421874, 0.035904510498046875, 0.035768318176269534, 0.03570073699951172, 0.035768318176269534, 0.03587071990966797, 0.03590041732788086, 0.03577446365356445, 0.035768318176269534, 0.03579084777832031, 0.03578265762329102, 0.03573555374145508, 0.03573964691162109, 0.035694591522216795, 0.035737598419189456, 0.035740734100341794, 0.03573446273803711, 0.03573964691162109, 0.03577241516113281, 0.03577139282226562, 0.03581644821166992, 0.03576115036010742, 0.035775489807128906, 0.03575603103637695, 0.035760128021240234, 0.03581951904296875, 0.035796993255615236, 0.03580825424194336, 0.03579391860961914, 0.035810302734375, 0.03578675079345703, 0.035783679962158206, 0.0357918701171875, 0.0357918701171875, 0.03580416107177734, 0.0358287353515625, 0.03581849670410156, 0.03578579330444336, 0.13018515014648438, 0.03566592025756836, 0.03563417434692383, 0.03563724899291992, 0.03563417434692383, 0.035655681610107424, 0.035729408264160156, 0.03564851379394531, 0.035643390655517575, 0.035681278228759765, 0.035625984191894534, 0.03570483016967774, 0.03566899108886719, 0.03570175933837891, 0.035645439147949216, 0.0356864013671875, 0.03568947219848633, 0.0358287353515625, 0.03568230438232422, 0.035664894104003905, 0.03567923355102539, 0.03585228729248047, 0.035817470550537106, 0.035705856323242184, 0.035692543029785154, 0.0357314567565918, 0.035722240447998044, 0.035765247344970705, 0.03571712112426758, 0.03578675079345703, 0.03579913711547852, 0.03577231979370117, 0.03571200180053711, 0.035699710845947266, 0.035699710845947266, 0.03570995330810547, 0.035737598419189456, 0.03571814346313477, 0.03575603103637695, 0.03574476623535156, 0.035778560638427735, 0.0358287353515625, 0.03575500869750976, 0.035776512145996094, 0.03581644821166992, 0.035776512145996094, 0.03572739028930664, 0.035780574798583986, 0.03574784088134766, 0.03585843276977539, 0.03590860748291016, 0.03581235122680664, 0.03577446365356445, 0.035767295837402346, 0.03578060913085938, 0.03580825424194336, 0.03580313491821289, 0.03582566452026367, 0.035794944763183595, 0.0357918701171875, 0.03583900833129883, 0.03592086410522461, 0.03580518341064453, 0.13022720336914062, 0.03563724899291992, 0.03561369705200195, 0.0355676155090332, 0.03562188720703125, 0.03561881637573242, 0.03562188720703125, 0.03565055847167969, 0.03564134216308594, 0.035620864868164064, 0.03564851379394531, 0.03566796875, 0.03566080093383789, 0.03567103958129883, 0.035620864868164064, 0.03565260696411133, 0.0356577262878418, 0.035655681610107424, 0.035676223754882816, 0.03566175842285156, 0.03569356918334961, 0.03571507263183594, 0.03568025588989258, 0.03564646530151367, 0.03566284942626953, 0.035672065734863284, 0.03570073699951172, 0.03567411041259766, 0.03570278549194336, 0.03571712112426758, 0.03573657608032227, 0.035768318176269534, 0.0357386245727539, 0.035697662353515625, 0.03568742370605469, 0.03570892715454101, 0.03570892715454101, 0.03570073699951172, 0.03566899108886719, 0.03571916961669922, 0.03576422500610352, 0.03585126495361328, 0.035762176513671876, 0.0357386245727539, 0.03573555374145508, 0.035806209564208984, 0.03589734268188476, 0.035784702301025394, 0.035827713012695314, 0.035814399719238284, 0.03584307098388672, 0.03577958297729492, 0.0358287353515625, 0.0357498893737793, 0.035740673065185545, 0.035772480010986325, 0.0357907829284668, 0.035796993255615236, 0.03585638427734375, 0.035798015594482424, 0.03581849670410156, 0.035901439666748046, 0.03587583923339844]",tokens/s,26.95905284702418,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-180B,tiiuae/falcon-180B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 403 Client Error: Forbidden for url: https://huggingface.co/tiiuae/falcon-180B/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6693270c-582f474d68de2cfc0034b8ca;70c01925-8a97-466e-b004-3ec61f44989b) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/tiiuae/falcon-180B/resolve/main/config.json. If you are trying to create or update content,make sure you have a token with the `write` role. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1826, in _raise_on_head_call_error raise LocalEntryNotFoundError( huggingface_hub.utils._errors.LocalEntryNotFoundError: An error happened while trying to locate the file on the Hub and we cannot find the requested files in the local cache. Please check your connection and try again or make sure your Internet connection is on. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 445, in cached_file raise EnvironmentError( OSError: We couldn't connect to 'https://huggingface.co' to load this file, couldn't find it in the cached files and it looks like tiiuae/falcon-180B is not the path to a directory containing a file named config.json. Checkout your internet connection or see how to run the library in offline mode at 'https://huggingface.co/docs/transformers/installation#offline-mode'. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,mistralai/Mistral-7B-v0.1,mistralai/Mistral-7B-v0.1,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.1,,,,MB,1272.81152,15987.113984,0.0,15340.66688,15114.865664,s,10,0.8057655334472655,0.08057655334472655,0.0002688233481084678,0.0805254898071289,0.08084041671752928,0.08101911964416504,0.08116208198547363,"[0.08119782257080078, 0.08051532745361328, 0.08030156707763672, 0.08016815948486328, 0.08046479797363282, 0.08053340911865234, 0.08080070495605468, 0.08072271728515625, 0.08054345703125, 0.08051757049560547]",tokens/s,3177.102883822398,kWh,9.523693217171564e-07,5.217386787029681e-07,6.48024073974407e-06,7.954348740164194e-06,tokens/kWh,32183653.038415268,MB,1273.1392,15987.113984,0.0,15340.66688,15134.09792,s,10,23.419662841796875,2.3419662841796876,0.0012973634900605752,2.3416966552734375,2.3433538330078125,2.344137487792969,2.3447644116210937,"[2.344921142578125, 2.341200927734375, 2.340545166015625, 2.342385498046875, 2.3431796875, 2.340890625, 2.340749267578125, 2.341056884765625, 2.342541259765625, 2.3421923828125]",tokens/s,26.900472660761125,kWh,2.7592292627361097e-05,1.5121688913337128e-05,0.00012333444644525545,0.0001660484279859536,tokens/kWh,379407.38593038236,,s,629,24.106319873809827,0.03832483286774215,0.009640235182452673,0.037152767181396484,0.037244108581542965,0.03730838928222656,0.11819950653076172,"[0.03721830368041992, 0.03718656158447266, 0.037456897735595705, 0.03798015975952149, 0.03725107192993164, 0.03714048004150391, 0.03715379333496094, 0.03710976028442383, 0.03713740921020508, 0.03710976028442383, 0.03711897659301758, 0.037103614807128905, 0.037310462951660156, 0.03738726425170898, 0.03752959823608398, 0.03722444915771484, 0.03710771179199219, 0.037147647857666014, 0.03712819290161133, 0.03719475173950195, 0.03738521575927734, 0.03719987106323242, 0.037215232849121094, 0.03714355087280274, 0.03711897659301758, 0.037131263732910154, 0.03713433456420898, 0.03716505432128906, 0.037179393768310545, 0.037146625518798826, 0.037250049591064455, 0.03713228988647461, 0.037185535430908204, 0.037146625518798826, 0.03715584182739258, 0.03720806503295898, 0.03713228988647461, 0.037149696350097655, 0.03719475173950195, 0.0371599349975586, 0.03720806503295898, 0.037174270629882815, 0.037168128967285156, 0.03713740921020508, 0.03715379333496094, 0.037195777893066405, 0.03729510498046875, 0.03726540756225586, 0.0371486701965332, 0.037176319122314457, 0.03718041610717773, 0.03715891265869141, 0.03727769470214844, 0.03718041610717773, 0.03719270324707031, 0.037184513092041016, 0.03719372940063476, 0.037169151306152344, 0.037177345275878904, 0.03717836761474609, 0.037182464599609374, 0.037171199798583986, 0.11826483154296875, 0.037119998931884765, 0.03713433456420898, 0.03718656158447266, 0.037179393768310545, 0.037157886505126955, 0.03720499038696289, 0.03708927917480469, 0.03711180877685547, 0.03707289505004883, 0.03713433456420898, 0.037146625518798826, 0.037108734130859376, 0.03722963333129883, 0.03708921432495117, 0.03712102508544922, 0.03709542465209961, 0.03711795043945312, 0.03707494354248047, 0.037114879608154294, 0.03710771179199219, 0.03717836761474609, 0.03710259246826172, 0.03711385726928711, 0.037108734130859376, 0.0370780143737793, 0.037114879608154294, 0.0371517448425293, 0.03719987106323242, 0.03712006378173828, 0.037123008728027346, 0.03720499038696289, 0.03709337615966797, 0.03711385726928711, 0.037082111358642575, 0.03717529678344727, 0.037116928100585936, 0.0370964469909668, 0.037171199798583986, 0.03715584182739258, 0.037136383056640625, 0.03714048004150391, 0.037195777893066405, 0.03719680023193359, 0.03715891265869141, 0.03720806503295898, 0.037215232849121094, 0.0371486701965332, 0.037171199798583986, 0.03718041610717773, 0.03719372940063476, 0.03718041610717773, 0.0372408332824707, 0.037154815673828126, 0.03716505432128906, 0.03721011352539062, 0.0371599349975586, 0.03724800109863281, 0.0373196792602539, 0.03720908737182617, 0.037169151306152344, 0.03720294570922852, 0.037176319122314457, 0.1180631332397461, 0.037182430267333984, 0.037119998931884765, 0.03714355087280274, 0.037138431549072266, 0.037181503295898435, 0.037142463684082035, 0.03712614440917969, 0.037098495483398435, 0.03711078262329102, 0.03712204742431641, 0.03705344009399414, 0.03710976028442383, 0.03707596969604492, 0.03707494354248047, 0.03710771179199219, 0.03715379333496094, 0.03710259246826172, 0.03710259246826172, 0.03717529678344727, 0.037101566314697264, 0.037104671478271484, 0.037113822937011716, 0.037190654754638675, 0.0370964469909668, 0.03711180877685547, 0.03713433456420898, 0.0370882568359375, 0.03712204742431641, 0.037085182189941404, 0.037087230682373046, 0.03712819290161133, 0.03710566329956055, 0.037133312225341795, 0.037104671478271484, 0.037227489471435546, 0.03725625610351563, 0.0371578254699707, 0.03708927917480469, 0.03713433456420898, 0.037157886505126955, 0.03713740921020508, 0.03712102508544922, 0.037138431549072266, 0.03719782257080078, 0.0371671028137207, 0.03714355087280274, 0.03726438522338867, 0.03713433456420898, 0.0371671028137207, 0.03711897659301758, 0.03715891265869141, 0.03711283111572266, 0.03717836761474609, 0.037184513092041016, 0.03715795135498047, 0.037155776977539065, 0.03716403198242187, 0.03722751998901367, 0.037253120422363284, 0.03718041610717773, 0.037220352172851565, 0.037168128967285156, 0.11792281341552735, 0.03712819290161133, 0.03709952163696289, 0.037103614807128905, 0.03717222213745117, 0.037169151306152344, 0.03712409591674805, 0.03708620834350586, 0.03711590576171875, 0.03706880187988281, 0.03707904052734375, 0.03710259246826172, 0.03715071868896484, 0.03716198348999023, 0.03727155303955078, 0.03731455993652344, 0.03717942428588867, 0.03722953414916992, 0.037157886505126955, 0.037114879608154294, 0.037130241394042966, 0.037138431549072266, 0.037101566314697264, 0.03716403198242187, 0.03708313751220703, 0.03715686416625977, 0.03714559936523437, 0.037154815673828126, 0.03709235382080078, 0.03719782257080078, 0.03716508865356445, 0.03715375900268555, 0.037177345275878904, 0.03710976028442383, 0.037223423004150394, 0.03732585525512695, 0.0371732177734375, 0.037182464599609374, 0.03712716674804688, 0.03720601654052735, 0.03713740921020508, 0.037176319122314457, 0.03718860626220703, 0.037179393768310545, 0.037190654754638675, 0.03717222213745117, 0.037179393768310545, 0.03724390411376953, 0.03723980712890625, 0.03719372940063476, 0.03726233673095703, 0.037163009643554686, 0.037160961151123044, 0.03719782257080078, 0.037149696350097655, 0.037176319122314457, 0.03717836761474609, 0.037223423004150394, 0.0372305908203125, 0.0372592658996582, 0.03713536071777344, 0.03728076934814453, 0.03731455993652344, 0.1182525405883789, 0.037101566314697264, 0.037200897216796876, 0.03729510498046875, 0.037114879608154294, 0.037171199798583986, 0.03708927917480469, 0.037116928100585936, 0.03708006286621094, 0.03710771179199219, 0.03709235382080078, 0.03706880187988281, 0.03723263931274414, 0.03739852905273437, 0.037190654754638675, 0.03706982421875, 0.03709439849853516, 0.03719168090820312, 0.037108734130859376, 0.037125118255615236, 0.03708313751220703, 0.03711283111572266, 0.03715686416625977, 0.037119998931884765, 0.03711590576171875, 0.03709337615966797, 0.03716403198242187, 0.037101566314697264, 0.03718041610717773, 0.0373493766784668, 0.037394432067871096, 0.03716403198242187, 0.037130241394042966, 0.037144577026367184, 0.037147647857666014, 0.03717529678344727, 0.03718963241577149, 0.03714355087280274, 0.03715584182739258, 0.037171199798583986, 0.037177345275878904, 0.037220352172851565, 0.03713945770263672, 0.0372490234375, 0.037149696350097655, 0.03717737579345703, 0.037146591186523435, 0.037454849243164064, 0.03723775863647461, 0.03715686416625977, 0.037181503295898435, 0.03730527877807617, 0.03714559936523437, 0.03717529678344727, 0.037130241394042966, 0.037250049591064455, 0.03769651031494141, 0.03725516891479492, 0.03715379333496094, 0.03726028823852539, 0.03718656158447266, 0.03719270324707031, 0.037212158203125, 0.11837235260009765, 0.037163009643554686, 0.03712204742431641, 0.03710566329956055, 0.037130241394042966, 0.03723571014404297, 0.037195777893066405, 0.03713740921020508, 0.03705344009399414, 0.037108734130859376, 0.03711078262329102, 0.03710976028442383, 0.03712307357788086, 0.037133312225341795, 0.03707699203491211, 0.0370780143737793, 0.037103649139404296, 0.037110752105712894, 0.037119998931884765, 0.03712614440917969, 0.03710566329956055, 0.03712716674804688, 0.037117984771728514, 0.03714249420166016, 0.03711385726928711, 0.037116928100585936, 0.03711078262329102, 0.037131263732910154, 0.03712204742431641, 0.037128257751464847, 0.03709535980224609, 0.03712102508544922, 0.03711795043945312, 0.03713228988647461, 0.037133312225341795, 0.037160961151123044, 0.03719680023193359, 0.037130241394042966, 0.037160961151123044, 0.037163009643554686, 0.03718348693847656, 0.037136383056640625, 0.03709542465209961, 0.03722547149658203, 0.03722854232788086, 0.037149696350097655, 0.037142559051513674, 0.03716912078857422, 0.03715891265869141, 0.037207038879394534, 0.03716403198242187, 0.03726233673095703, 0.037147647857666014, 0.037174270629882815, 0.03714355087280274, 0.03720294570922852, 0.037185535430908204, 0.037149696350097655, 0.03718656158447266, 0.03719680023193359, 0.037184513092041016, 0.03722547149658203, 0.03719270324707031, 0.11865702056884765, 0.03710060882568359, 0.03712505722045899, 0.03712102508544922, 0.037133312225341795, 0.03712921524047851, 0.03720191955566406, 0.03712409591674805, 0.03707596969604492, 0.037082111358642575, 0.037098495483398435, 0.03709235382080078, 0.03705241775512695, 0.03717836761474609, 0.037059585571289064, 0.037149696350097655, 0.037108734130859376, 0.03710464096069336, 0.037116928100585936, 0.037084159851074217, 0.037100543975830076, 0.037100543975830076, 0.03714048004150391, 0.037166080474853515, 0.037119998931884765, 0.03713433456420898, 0.03712102508544922, 0.03713740921020508, 0.03715071868896484, 0.037152767181396484, 0.03709542465209961, 0.03717529678344727, 0.03711494445800781, 0.037159870147705075, 0.03709439849853516, 0.037163009643554686, 0.03713228988647461, 0.037106689453125, 0.037103614807128905, 0.03713536071777344, 0.03712614440917969, 0.03715071868896484, 0.03717228698730469, 0.03719059371948242, 0.0371486701965332, 0.03721625518798828, 0.037141502380371096, 0.03720294570922852, 0.037179393768310545, 0.037187583923339845, 0.03722547149658203, 0.037177345275878904, 0.037174270629882815, 0.03718860626220703, 0.037187583923339845, 0.03727667236328125, 0.037217281341552735, 0.037190654754638675, 0.03715584182739258, 0.037169151306152344, 0.03717324829101563, 0.03719168090820312, 0.03717228698730469, 0.11852489471435547, 0.037117919921875, 0.037125118255615236, 0.037133312225341795, 0.03712307357788086, 0.03713433456420898, 0.03714048004150391, 0.03710976028442383, 0.0370882568359375, 0.03707494354248047, 0.03709747314453125, 0.037087230682373046, 0.03709132766723633, 0.03724595260620117, 0.03709542465209961, 0.03707904052734375, 0.0371701774597168, 0.037166080474853515, 0.0370964469909668, 0.037106689453125, 0.037103614807128905, 0.037138431549072266, 0.037269504547119144, 0.037395454406738284, 0.03717942428588867, 0.03721007919311523, 0.03713945770263672, 0.037116928100585936, 0.03711283111572266, 0.03711385726928711, 0.03715891265869141, 0.03710771179199219, 0.037131263732910154, 0.03709648132324219, 0.037190654754638675, 0.03717116928100586, 0.03714048004150391, 0.03713536071777344, 0.037130241394042966, 0.037125118255615236, 0.037141502380371096, 0.03723263931274414, 0.03710771179199219, 0.037168128967285156, 0.03714559936523437, 0.03719475173950195, 0.037114879608154294, 0.037149696350097655, 0.037160961151123044, 0.037119998931884765, 0.03714355087280274, 0.03718963241577149, 0.0371517448425293, 0.03716505432128906, 0.03714048004150391, 0.03717529678344727, 0.03715891265869141, 0.03717324829101563, 0.03720294570922852, 0.03719270324707031, 0.03722751998901367, 0.037184574127197265, 0.037204929351806644, 0.11828121948242187, 0.03717529678344727, 0.03712204742431641, 0.03713945770263672, 0.03711078262329102, 0.0371701774597168, 0.03714048004150391, 0.037087230682373046, 0.03711590576171875, 0.03706982421875, 0.037082111358642575, 0.03707392120361328, 0.03709132766723633, 0.03705036926269531, 0.03708108901977539, 0.037100543975830076, 0.03709439849853516, 0.037152767181396484, 0.03711795043945312, 0.0371517448425293, 0.037084159851074217, 0.03711795043945312, 0.03724492645263672, 0.03739648056030274, 0.03710976028442383, 0.0370882568359375, 0.03712102508544922, 0.037103614807128905, 0.03712105560302734, 0.037146591186523435, 0.03719782257080078, 0.037152767181396484, 0.0370964469909668, 0.037163009643554686, 0.037106689453125, 0.037114879608154294, 0.037106689453125, 0.03748659133911133, 0.037482494354248046, 0.03723468780517578, 0.037375999450683595, 0.03753062438964844, 0.037256191253662106, 0.03719782257080078, 0.03722547149658203, 0.037154815673828126, 0.037187583923339845, 0.037198848724365234, 0.037149696350097655, 0.03729715347290039, 0.03713740921020508, 0.03714559936523437, 0.03715584182739258, 0.03717222213745117, 0.037131263732910154, 0.03718860626220703, 0.03725414276123047, 0.03717529678344727, 0.037133312225341795, 0.03726233673095703, 0.03720191955566406, 0.03722956848144531, 0.03722956848144531, 0.11867852783203126, 0.037163009643554686, 0.037144577026367184, 0.03711590576171875, 0.03724492645263672, 0.03735551834106445, 0.037152767181396484, 0.037154815673828126, 0.0371701774597168, 0.03707699203491211, 0.037141502380371096, 0.0370780143737793, 0.03707904052734375, 0.03713536071777344, 0.03706163024902344, 0.03714559936523437, 0.037103614807128905, 0.0370964469909668, 0.03710259246826172, 0.037136383056640625, 0.03709132766723633, 0.03746105575561524, 0.03715372848510742, 0.037163009643554686, 0.037131263732910154, 0.037084159851074217, 0.037185535430908204, 0.03710259246826172, 0.037101566314697264, 0.0371486701965332, 0.03712204742431641, 0.03716505432128906, 0.0370882568359375, 0.037179393768310545, 0.03715686416625977, 0.03719372940063476, 0.037144577026367184, 0.03712307357788086, 0.037087230682373046, 0.03711795043945312, 0.03715071868896484, 0.037294078826904296, 0.037146625518798826, 0.03721932983398438, 0.03718656158447266, 0.037266433715820314, 0.03711795043945312, 0.037207038879394534, 0.037207038879394534, 0.037157886505126955, 0.03719168090820312, 0.03716505432128906, 0.037148735046386716, 0.03719776153564453, 0.0371671028137207, 0.03719680023193359, 0.03718656158447266, 0.03718963241577149, 0.03717324829101563, 0.03718656158447266, 0.03723161697387695, 0.03723468780517578, 0.03722444915771484]",tokens/s,26.092742620717225,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,google/recurrentgemma-7b,google/recurrentgemma-7b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/google/recurrentgemma-7b/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66947042-458157e14426494755b52e7a;ab95f370-378f-46fb-ac48-4911724784c0) Repository Not Found for url: https://huggingface.co/google/recurrentgemma-7b/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: google/recurrentgemma-7b is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,mistralai/Mixtral-8x7B-v0.1,mistralai/Mixtral-8x7B-v0.1,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ self.load_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 250, in load_model_with_no_weights self.pretrained_model = self.automodel_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3710, in from_pretrained model = cls(config, *model_args, **model_kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/mixtral/modeling_mixtral.py"", line 1323, in __init__ self.model = MixtralModel(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/mixtral/modeling_mixtral.py"", line 1087, in __init__ [MixtralDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/mixtral/modeling_mixtral.py"", line 1087, in [MixtralDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/mixtral/modeling_mixtral.py"", line 877, in __init__ self.block_sparse_moe = MixtralSparseMoeBlock(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/mixtral/modeling_mixtral.py"", line 824, in __init__ self.experts = nn.ModuleList([MixtralBlockSparseTop2MLP(config) for _ in range(self.num_experts)]) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/mixtral/modeling_mixtral.py"", line 824, in self.experts = nn.ModuleList([MixtralBlockSparseTop2MLP(config) for _ in range(self.num_experts)]) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/mixtral/modeling_mixtral.py"", line 791, in __init__ self.w2 = nn.Linear(self.ffn_dim, self.hidden_dim, bias=False) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/linear.py"", line 98, in __init__ self.weight = Parameter(torch.empty((out_features, in_features), **factory_kwargs)) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_device.py"", line 78, in __torch_function__ return func(*args, **kwargs) torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 112.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,t,t,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/t/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669473a6-2ca4942377d64462366409e8;cb17bce6-021f-4019-babe-483053c85f61) Repository Not Found for url: https://huggingface.co/t/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: t is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-72B,Qwen/Qwen-72B,cuda,0,42,,,True,,,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",qwen,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-30b,facebook/opt-30b,cuda,0,42,,,True,,,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",opt,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-MoE-A2.7B,Qwen/Qwen1.5-MoE-A2.7B,cuda,0,42,,,True,,,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",qwen2_moe,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,x,x,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/x/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6693246d-62643a15140d18b21a51e400;695326cd-a7c8-4446-bd5b-269d7a834505) Repository Not Found for url: https://huggingface.co/x/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: x is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,01-ai/Yi-34B,01-ai/Yi-34B,cuda,0,42,,,True,,,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",llama,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/xglm-4.5B,facebook/xglm-4.5B,cuda,0,42,,,True,,,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",xglm,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/xglm-7.5B,facebook/xglm-7.5B,cuda,0,42,,,True,,,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",xglm,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-14B,Qwen/Qwen1.5-14B,cuda,0,42,,,True,,,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",qwen2,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,meta-llama/Llama-2-70b-hf,meta-llama/Llama-2-70b-hf,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ self.load_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 250, in load_model_with_no_weights self.pretrained_model = self.automodel_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3710, in from_pretrained model = cls(config, *model_args, **model_kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 1101, in __init__ self.model = LlamaModel(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 884, in __init__ [LlamaDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 884, in [LlamaDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 679, in __init__ self.mlp = LlamaMLP(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 192, in __init__ self.gate_proj = nn.Linear(self.hidden_size, self.intermediate_size, bias=config.mlp_bias) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/linear.py"", line 98, in __init__ self.weight = Parameter(torch.empty((out_features, in_features), **factory_kwargs)) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_device.py"", line 78, in __torch_function__ return func(*args, **kwargs) torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 448.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Deci/DeciCoder-1b,Deci/DeciCoder-1b,cuda,0,42,,,True,,,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",llama,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,meta-llama/Meta-Llama-3-70B,meta-llama/Meta-Llama-3-70B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ self.load_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 250, in load_model_with_no_weights self.pretrained_model = self.automodel_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3710, in from_pretrained model = cls(config, *model_args, **model_kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 1101, in __init__ self.model = LlamaModel(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 884, in __init__ [LlamaDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 884, in [LlamaDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 679, in __init__ self.mlp = LlamaMLP(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 193, in __init__ self.up_proj = nn.Linear(self.hidden_size, self.intermediate_size, bias=config.mlp_bias) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/linear.py"", line 98, in __init__ self.weight = Parameter(torch.empty((out_features, in_features), **factory_kwargs)) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_device.py"", line 78, in __torch_function__ return func(*args, **kwargs) torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 448.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,r,r,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/r/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669473e3-33adb42f7affe0995bf78662;77caad29-b50e-4056-911c-3ab5fbe58e49) Repository Not Found for url: https://huggingface.co/r/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: r is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,togethercomputer/RedPajama-INCITE-Base-7B-v0.1,togethercomputer/RedPajama-INCITE-Base-7B-v0.1,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.42.3,,0.31.0,,,,1.20.0,,,,0.11.1,,,,MB,1275.551744,15150.350336,0.0,14503.903232,14287.962624,s,10,0.8546160125732423,0.08546160125732423,0.00029379681964468055,0.08537347412109375,0.0855491554260254,0.08594084892272949,0.08625420372009278,"[0.0863325424194336, 0.08546211242675782, 0.08531110382080079, 0.08531433868408203, 0.0853482894897461, 0.08531056213378906, 0.0853846435546875, 0.0853623046875, 0.08539552307128906, 0.08539459228515625]",tokens/s,2995.4973489109566,kWh,1.008620978860981e-06,5.526742358524293e-07,5.46066350223712e-06,7.0219587169505315e-06,tokens/kWh,36457064.23508777,MB,1275.551744,15150.350336,0.0,14503.903232,14465.569792,s,10,21.003234619140624,2.100323461914063,0.0018054849471219484,2.09967041015625,2.1017933349609375,2.1034835815429687,2.1048357788085936,"[2.105173828125, 2.100306396484375, 2.099818603515625, 2.101417724609375, 2.0991279296875, 2.100860107421875, 2.099522216796875, 2.09895361328125, 2.099132568359375, 2.098921630859375]",tokens/s,29.99537982715622,kWh,2.4792883441048597e-05,1.3587280156876905e-05,0.0001181602131721619,0.0001565403767700874,tokens/kWh,402452.07849811687,,s,629,21.73788667678832,0.03455943827788288,0.010147726548790146,0.033334270477294925,0.03345100784301758,0.033492581939697265,0.11874189483642579,"[0.033708030700683594, 0.03364556884765625, 0.03374182510375977, 0.03372851181030274, 0.03376537704467773, 0.03371929550170898, 0.033675262451171875, 0.033742847442626955, 0.033739776611328126, 0.03365785598754883, 0.03336601638793945, 0.03354111862182617, 0.03340595245361328, 0.03326976013183594, 0.033285118103027346, 0.03325747299194336, 0.033261566162109374, 0.03321855926513672, 0.0333199348449707, 0.03328102493286133, 0.033312767028808594, 0.03326566314697266, 0.033304576873779294, 0.0332492790222168, 0.033317886352539065, 0.03331071853637695, 0.03332198333740234, 0.03329945755004883, 0.03328204727172852, 0.03332403182983398, 0.03334348678588867, 0.03327078247070313, 0.03326259231567383, 0.033329151153564454, 0.033347583770751955, 0.03328204727172852, 0.033288192749023435, 0.033363967895507815, 0.03332505416870117, 0.03332403182983398, 0.03335270309448242, 0.03337113571166992, 0.03335782241821289, 0.033470462799072266, 0.033363967895507815, 0.03340800094604492, 0.033377281188964845, 0.03341107177734375, 0.03338137435913086, 0.033372161865234375, 0.03340185546875, 0.033432575225830076, 0.033326080322265625, 0.03335372924804687, 0.03339263916015625, 0.033391616821289063, 0.03337625503540039, 0.03339571380615235, 0.03341721725463867, 0.03345612716674805, 0.033452030181884765, 0.033501182556152344, 0.11888127899169922, 0.033170433044433595, 0.03321855926513672, 0.03319500732421875, 0.03321855926513672, 0.03322880172729492, 0.03323494338989258, 0.03317964935302734, 0.03325961685180664, 0.0332850227355957, 0.0332410888671875, 0.0332677116394043, 0.03323699188232422, 0.03342131042480469, 0.03335372924804687, 0.03327590560913086, 0.0332492790222168, 0.033393665313720705, 0.03348889541625977, 0.03327078247070313, 0.03325235366821289, 0.03328921508789062, 0.03324006271362305, 0.033259521484375, 0.03330047988891602, 0.033326080322265625, 0.03333324813842774, 0.03328204727172852, 0.033329151153564454, 0.0334284782409668, 0.0334919662475586, 0.03330355072021484, 0.03332710266113281, 0.03342131042480469, 0.03334860610961914, 0.03333631896972656, 0.03331379318237305, 0.033320960998535154, 0.033337345123291014, 0.03338854217529297, 0.033331199645996096, 0.03334041595458984, 0.03337932968139649, 0.03337625503540039, 0.03339468765258789, 0.03342540740966797, 0.033372161865234375, 0.03335782241821289, 0.03341721725463867, 0.033339393615722655, 0.03340288162231445, 0.033380352020263675, 0.03337113571166992, 0.03336806488037109, 0.033448959350585936, 0.03339263916015625, 0.03344076919555664, 0.03341721725463867, 0.03339468765258789, 0.03341721725463867, 0.03341516876220703, 0.0334571533203125, 0.03347353744506836, 0.11873894500732422, 0.033261566162109374, 0.03314790344238281, 0.03323392105102539, 0.03320627212524414, 0.03321036911010742, 0.03322982406616211, 0.03325132751464844, 0.03324006271362305, 0.03324518585205078, 0.03323904037475586, 0.03332710266113281, 0.033459201812744144, 0.033255424499511715, 0.033296382904052735, 0.03326259231567383, 0.03321343994140625, 0.033274879455566404, 0.03322982406616211, 0.033258495330810545, 0.03326566314697266, 0.033255424499511715, 0.03328921508789062, 0.033298431396484376, 0.033277950286865234, 0.033293312072753906, 0.03331174468994141, 0.03331891250610351, 0.033314815521240236, 0.03331891250610351, 0.03328102493286133, 0.03331891250610351, 0.033309696197509765, 0.03349401473999023, 0.03337932968139649, 0.033363967895507815, 0.033306625366210936, 0.033309696197509765, 0.0333199348449707, 0.03331174468994141, 0.03346124649047852, 0.03338854217529297, 0.0333383674621582, 0.03332710266113281, 0.03335168075561523, 0.033372161865234375, 0.033393665313720705, 0.03339263916015625, 0.03338854217529297, 0.033358848571777344, 0.03344793701171875, 0.03344998550415039, 0.033438720703125, 0.033377281188964845, 0.03338956832885742, 0.03340083312988281, 0.03339980697631836, 0.03336191940307617, 0.033396736145019534, 0.033492992401123044, 0.033430526733398434, 0.033438720703125, 0.03344998550415039, 0.11879730987548828, 0.033173503875732424, 0.033175552368164066, 0.03319705581665039, 0.03317862319946289, 0.033291263580322264, 0.03321753692626953, 0.0332492790222168, 0.03323187255859375, 0.03322060775756836, 0.033296382904052735, 0.03323392105102539, 0.03322675323486328, 0.03330252838134766, 0.03330047988891602, 0.03322982406616211, 0.033242111206054685, 0.033285118103027346, 0.033298431396484376, 0.033279998779296875, 0.0332861442565918, 0.03328409576416016, 0.033279998779296875, 0.0333199348449707, 0.03338240051269531, 0.033291263580322264, 0.03335270309448242, 0.03334860610961914, 0.033344512939453126, 0.03334041595458984, 0.033329151153564454, 0.03335372924804687, 0.03338240051269531, 0.033344512939453126, 0.033396736145019534, 0.03338649749755859, 0.03338240051269531, 0.03336601638793945, 0.03338956832885742, 0.03335270309448242, 0.03338137435913086, 0.033380352020263675, 0.03343667221069336, 0.03341209411621094, 0.033438720703125, 0.03338854217529297, 0.03342950439453125, 0.033821697235107424, 0.03347251129150391, 0.033393665313720705, 0.03346432113647461, 0.033416191101074216, 0.03340083312988281, 0.03343564987182617, 0.03345510482788086, 0.033462272644042966, 0.033459201812744144, 0.03346432113647461, 0.033445888519287106, 0.033481727600097655, 0.033465343475341795, 0.033442817687988284, 0.03347967910766601, 0.1187430419921875, 0.03321036911010742, 0.033181697845458984, 0.03320729446411133, 0.03317657470703125, 0.03323187255859375, 0.03323392105102539, 0.033259521484375, 0.03324620819091797, 0.03325439834594727, 0.03323494338989258, 0.033260543823242186, 0.03321139144897461, 0.03327590560913086, 0.03323494338989258, 0.03328716659545899, 0.033261566162109374, 0.03331071853637695, 0.03323494338989258, 0.033260543823242186, 0.03324518585205078, 0.03328307342529297, 0.033304576873779294, 0.03335168075561523, 0.03328716659545899, 0.03332403182983398, 0.033274879455566404, 0.03336703872680664, 0.033328128814697267, 0.03329228973388672, 0.03328307342529297, 0.033314815521240236, 0.033312767028808594, 0.0333568000793457, 0.03329433441162109, 0.03333631896972656, 0.03332198333740234, 0.033323009490966796, 0.03337830352783203, 0.03331584167480469, 0.03339468765258789, 0.03335168075561523, 0.03337625503540039, 0.03336703872680664, 0.033344512939453126, 0.033360897064208986, 0.033331199645996096, 0.03339980697631836, 0.033351711273193356, 0.033365985870361325, 0.03335987091064453, 0.03338547134399414, 0.03336601638793945, 0.03339263916015625, 0.033420352935791015, 0.03337311935424805, 0.03344793701171875, 0.033377281188964845, 0.03339263916015625, 0.03345100784301758, 0.03338854217529297, 0.03342438507080078, 0.03347251129150391, 0.118687744140625, 0.033259521484375, 0.03322163009643555, 0.033219585418701174, 0.03322880172729492, 0.03323801422119141, 0.03319705581665039, 0.03324825668334961, 0.03322060775756836, 0.033235969543457033, 0.03322367858886719, 0.03323699188232422, 0.033288192749023435, 0.03324518585205078, 0.03323392105102539, 0.033263614654541016, 0.03327385711669922, 0.03339468765258789, 0.0332492790222168, 0.033301502227783206, 0.033462272644042966, 0.03325030517578125, 0.033328128814697267, 0.03328102493286133, 0.03323699188232422, 0.03330559921264648, 0.033277950286865234, 0.03326259231567383, 0.03329433441162109, 0.033350654602050785, 0.03326668930053711, 0.033344512939453126, 0.03336601638793945, 0.03332710266113281, 0.03338751983642578, 0.03355648040771484, 0.033358848571777344, 0.03332403182983398, 0.033337345123291014, 0.0333496322631836, 0.03333631896972656, 0.03339775848388672, 0.033314815521240236, 0.03340288162231445, 0.03334143829345703, 0.03339775848388672, 0.03338649749755859, 0.033410049438476565, 0.03337011337280273, 0.03339468765258789, 0.03338444900512695, 0.033966079711914066, 0.03351551818847656, 0.03338137435913086, 0.03337318420410156, 0.03352166366577149, 0.03348787307739258, 0.03344179153442383, 0.033454078674316406, 0.03340390396118164, 0.03345100784301758, 0.03348992156982422, 0.03343155288696289, 0.11880242919921875, 0.03344793701171875, 0.03323187255859375, 0.033165313720703124, 0.0331960334777832, 0.03321139144897461, 0.033189888000488284, 0.03321036911010742, 0.03325747299194336, 0.03323904037475586, 0.033219585418701174, 0.03325030517578125, 0.03326259231567383, 0.03322163009643555, 0.03321139144897461, 0.03325235366821289, 0.03322777557373047, 0.033235969543457033, 0.03328921508789062, 0.03326259231567383, 0.03323699188232422, 0.03328102493286133, 0.033247230529785156, 0.03322777557373047, 0.033288192749023435, 0.033317886352539065, 0.03329945755004883, 0.03329536056518555, 0.033342464447021485, 0.03328307342529297, 0.03330867385864258, 0.033377281188964845, 0.03328921508789062, 0.033326080322265625, 0.03347353744506836, 0.033301502227783206, 0.03332505416870117, 0.03335372924804687, 0.03365273666381836, 0.0333834228515625, 0.03336601638793945, 0.0333568000793457, 0.03336703872680664, 0.03333324813842774, 0.03335782241821289, 0.03336703872680664, 0.033393665313720705, 0.033358848571777344, 0.03335987091064453, 0.0333834228515625, 0.03342335891723633, 0.03340595245361328, 0.03338444900512695, 0.03339571380615235, 0.03340390396118164, 0.033364990234375, 0.033391616821289063, 0.033414142608642575, 0.03339468765258789, 0.03340800094604492, 0.03348787307739258, 0.033396736145019534, 0.033445888519287106, 0.11874918365478515, 0.033159168243408206, 0.03314176177978516, 0.03319705581665039, 0.03319705581665039, 0.03342745590209961, 0.03323801422119141, 0.033188865661621096, 0.033247230529785156, 0.03325030517578125, 0.03323494338989258, 0.03321036911010742, 0.03322470474243164, 0.03321036911010742, 0.03326566314697266, 0.03321651077270508, 0.03318067169189453, 0.03325337600708008, 0.03328102493286133, 0.03325439834594727, 0.03319500732421875, 0.03329433441162109, 0.03332198333740234, 0.03323187255859375, 0.03330559921264648, 0.03326566314697266, 0.033312767028808594, 0.033345535278320314, 0.03330355072021484, 0.033279998779296875, 0.03336294555664063, 0.03329945755004883, 0.033309696197509765, 0.03334860610961914, 0.033309696197509765, 0.033347583770751955, 0.0333568000793457, 0.03335270309448242, 0.033314815521240236, 0.03335270309448242, 0.03331379318237305, 0.03330867385864258, 0.03338444900512695, 0.033334270477294925, 0.03336294555664063, 0.03336294555664063, 0.033372161865234375, 0.033364990234375, 0.03338854217529297, 0.03338137435913086, 0.03339059066772461, 0.03340902328491211, 0.03343462371826172, 0.03339878463745117, 0.03342233657836914, 0.03340697479248047, 0.03347455978393555, 0.03341209411621094, 0.03339263916015625, 0.03343155288696289, 0.03343769454956055, 0.03342745590209961, 0.03339980697631836, 0.1188485107421875, 0.03319705581665039, 0.033152000427246094, 0.033154048919677735, 0.03319705581665039, 0.03320729446411133, 0.03322163009643555, 0.03324415969848633, 0.033247230529785156, 0.03319807815551758, 0.03321548843383789, 0.033242111206054685, 0.03320217514038086, 0.03323392105102539, 0.03325961685180664, 0.03322768020629883, 0.03321548843383789, 0.033263614654541016, 0.03324006271362305, 0.03325439834594727, 0.03356159973144531, 0.033334270477294925, 0.033277950286865234, 0.033260543823242186, 0.033259521484375, 0.033291263580322264, 0.03326668930053711, 0.03331686401367188, 0.033277950286865234, 0.03328204727172852, 0.03332505416870117, 0.033328128814697267, 0.033288192749023435, 0.033344512939453126, 0.03337420654296875, 0.033317886352539065, 0.03337011337280273, 0.033317886352539065, 0.0333383674621582, 0.03335270309448242, 0.03332403182983398, 0.033360897064208986, 0.03337011337280273, 0.03335782241821289, 0.03335782241821289, 0.03337113571166992, 0.03339980697631836, 0.033426433563232424, 0.03337830352783203, 0.033426433563232424, 0.03339059066772461, 0.03337830352783203, 0.03339059066772461, 0.03336908721923828, 0.03340185546875, 0.03340390396118164, 0.03341926574707031, 0.03338649749755859, 0.03344179153442383, 0.03338751983642578, 0.033468414306640625, 0.03342335891723633, 0.03341823959350586, 0.11877887725830077, 0.03318272018432617, 0.03319091033935547, 0.033186817169189455, 0.03319910430908203, 0.033185791015625, 0.03317452621459961, 0.033235969543457033, 0.03319398498535156, 0.033186817169189455, 0.033242111206054685, 0.033314815521240236, 0.033235969543457033, 0.033209342956542966, 0.03325439834594727, 0.03321651077270508, 0.033247230529785156, 0.033260543823242186, 0.03322265625, 0.033255424499511715, 0.03321139144897461, 0.03321548843383789, 0.0332564468383789, 0.03326976013183594, 0.0332677116394043, 0.03329945755004883, 0.03329740905761719, 0.03326668930053711, 0.03331686401367188, 0.03332505416870117, 0.03334656143188477, 0.033288192749023435, 0.03333017730712891, 0.03331379318237305, 0.033342464447021485, 0.033317886352539065, 0.03338137435913086, 0.03329536056518555, 0.033345535278320314, 0.0333383674621582, 0.03336703872680664, 0.03351347351074219, 0.0334571533203125, 0.03331686401367188, 0.03336908721923828, 0.03335168075561523, 0.033375232696533204, 0.03336294555664063, 0.03337625503540039, 0.03334656143188477, 0.03340595245361328, 0.033350654602050785, 0.03339878463745117, 0.03343667221069336, 0.03353395080566406, 0.03336908721923828, 0.033459201812744144, 0.0333834228515625, 0.03341926574707031, 0.03342335891723633, 0.03344998550415039, 0.033414142608642575, 0.03346432113647461]",tokens/s,28.93565549183053,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-72B,Qwen/Qwen1.5-72B,cuda,0,42,,,True,,,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",qwen2,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,internlm/internlm-20b,internlm/internlm-20b,cuda,0,42,,,True,,,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",internlm,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-110B,Qwen/Qwen1.5-110B,cuda,0,42,,,True,,,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",qwen2,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Salesforce/codegen-16B-nl,Salesforce/codegen-16B-nl,cuda,0,42,,,True,,,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",codegen,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,m,m,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/m/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669472e3-1863effa6690288a682bf23f;4faff99d-4442-4995-ba9b-1f62773e0070) Repository Not Found for url: https://huggingface.co/m/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: m is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,openai-community/gpt2-large,openai-community/gpt2-large,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.42.3,,0.31.0,,,,1.20.0,,,,0.11.1,,,,MB,1200.164864,2663.907328,0.0,2017.460224,1692.386816,s,10,0.1861420783996582,0.01861420783996582,0.0003858354528170821,0.018703680038452148,0.018857903671264647,0.019122215843200684,0.01933366558074951,"[0.01938652801513672, 0.018666143417358397, 0.01878748893737793, 0.01864860725402832, 0.01879430389404297, 0.01879916763305664, 0.01808742332458496, 0.018135679244995116, 0.01809552001953125, 0.0187412166595459]",tokens/s,13752.935510387537,kWh,2.1717296913266185e-07,1.1900093450595873e-07,8.63903366775683e-07,1.2000772704143036e-06,tokens/kWh,213319597.25528416,MB,1200.164864,2663.907328,0.0,2017.460224,1714.454528,s,10,11.390044067382814,1.1390044067382814,0.020196892256602473,1.14722705078125,1.1590804931640624,1.16129775390625,1.1630715624999999,"[1.1526417236328126, 1.1544222412109375, 1.1585877685546875, 1.1635150146484374, 1.1583651123046874, 1.1418123779296876, 1.114403564453125, 1.1138953857421876, 1.11367529296875, 1.1187255859375]",tokens/s,55.31146291207988,kWh,1.3607678394764665e-05,7.456672973387375e-06,3.121365007302644e-05,5.227800144117846e-05,tokens/kWh,1205095.8005899594,,s,629,11.537998861312872,0.018343400415441757,0.0022692666039140995,0.018156543731689453,0.01854074935913086,0.01869414405822754,0.036071833801269534,"[0.0188538875579834, 0.019216384887695313, 0.01801625633239746, 0.017704959869384765, 0.017724416732788087, 0.01757798385620117, 0.017627168655395507, 0.017593311309814452, 0.017697792053222656, 0.01762099266052246, 0.01762611198425293, 0.017588224411010742, 0.017655807495117186, 0.01763430404663086, 0.017729536056518554, 0.017870847702026366, 0.01759846305847168, 0.01763532829284668, 0.017572864532470703, 0.01764454460144043, 0.01757695960998535, 0.017567743301391603, 0.017625087738037108, 0.018534400939941405, 0.018585695266723632, 0.0188691520690918, 0.018703359603881836, 0.018618368148803712, 0.02002230453491211, 0.021888992309570313, 0.01965670394897461, 0.018687999725341797, 0.0184268798828125, 0.01839411163330078, 0.01772652816772461, 0.017787839889526365, 0.018359296798706053, 0.018298879623413086, 0.018518016815185546, 0.018404352188110353, 0.018315263748168945, 0.01842790412902832, 0.018319360733032225, 0.01842995262145996, 0.018267135620117187, 0.018502656936645507, 0.018408447265625, 0.018309120178222657, 0.018361343383789062, 0.01845145606994629, 0.018386943817138672, 0.018568191528320312, 0.018312192916870116, 0.018371583938598633, 0.018335744857788085, 0.018351104736328124, 0.018473983764648438, 0.01846784019470215, 0.018514944076538087, 0.018495487213134765, 0.018571264266967775, 0.01843097686767578, 0.037220352172851565, 0.018495487213134765, 0.018571264266967775, 0.01837772750854492, 0.018448383331298827, 0.01833171272277832, 0.018357183456420897, 0.01841459274291992, 0.018518016815185546, 0.018568191528320312, 0.017897472381591797, 0.01781350326538086, 0.018496511459350586, 0.018347007751464844, 0.018504703521728515, 0.01844121551513672, 0.018554943084716797, 0.018451391220092775, 0.0184586238861084, 0.018523136138916017, 0.01842483139038086, 0.01838489532470703, 0.018539520263671876, 0.018367488861083983, 0.018339839935302735, 0.018367488861083983, 0.018423807144165038, 0.018297855377197265, 0.017757183074951173, 0.01779097557067871, 0.018340864181518556, 0.018336767196655272, 0.0184268798828125, 0.018395135879516602, 0.018316287994384766, 0.018273279190063475, 0.01843302345275879, 0.01840025520324707, 0.01822822380065918, 0.01777663993835449, 0.017853439331054686, 0.018317312240600587, 0.018325504302978517, 0.018354175567626953, 0.01844633674621582, 0.018284543991088868, 0.01843302345275879, 0.01840332794189453, 0.01842995262145996, 0.01843609619140625, 0.018409536361694335, 0.01834899139404297, 0.018333696365356447, 0.018380800247192384, 0.01842995262145996, 0.018348031997680665, 0.01839308738708496, 0.018345983505249023, 0.018297855377197265, 0.01840336036682129, 0.017725408554077147, 0.01776639938354492, 0.018256895065307616, 0.03736576080322265, 0.018355199813842774, 0.01845964813232422, 0.018353151321411132, 0.01879756736755371, 0.01860403251647949, 0.018423871994018556, 0.01840224075317383, 0.018610176086425782, 0.018290687561035156, 0.018391040802001952, 0.018564096450805666, 0.018338815689086914, 0.018349056243896485, 0.018290687561035156, 0.01840230369567871, 0.018463743209838866, 0.01768550491333008, 0.018156543731689453, 0.018343936920166014, 0.01843097686767578, 0.018390016555786134, 0.018388992309570314, 0.018512895584106445, 0.018320383071899413, 0.018355199813842774, 0.01843507194519043, 0.0183767032623291, 0.018293760299682618, 0.018390016555786134, 0.018423807144165038, 0.018386943817138672, 0.018380800247192384, 0.01824460792541504, 0.018478080749511717, 0.01839923286437988, 0.018322431564331054, 0.01840640068054199, 0.019556352615356445, 0.018787328720092773, 0.018453504562377928, 0.01824460792541504, 0.018388992309570314, 0.018259967803955078, 0.018350080490112306, 0.01841971206665039, 0.01835212707519531, 0.018565120697021483, 0.018230335235595703, 0.01796908760070801, 0.018452543258666992, 0.018313152313232422, 0.018275327682495117, 0.018387968063354493, 0.018541568756103514, 0.01862553596496582, 0.018324480056762696, 0.01843097686767578, 0.01824051284790039, 0.01782067108154297, 0.017944576263427735, 0.018402368545532226, 0.018464736938476563, 0.036668384552001956, 0.017708032608032227, 0.018750463485717773, 0.019629056930541993, 0.019812416076660157, 0.0185916805267334, 0.01843609619140625, 0.018540544509887694, 0.01839206314086914, 0.018509824752807616, 0.018663423538208008, 0.01869824028015137, 0.0194201602935791, 0.0185743350982666, 0.018491392135620118, 0.01804902458190918, 0.01765478324890137, 0.018145280838012694, 0.018362367630004883, 0.018480159759521483, 0.018425823211669922, 0.018395135879516602, 0.018565120697021483, 0.01827020835876465, 0.018337791442871093, 0.01842995262145996, 0.018347007751464844, 0.01838489532470703, 0.018312192916870116, 0.017744960784912108, 0.017971136093139647, 0.018413568496704103, 0.018320383071899413, 0.018547712326049806, 0.018554880142211915, 0.018355199813842774, 0.0185743350982666, 0.018362367630004883, 0.018367488861083983, 0.01841868782043457, 0.018477088928222658, 0.018400224685668945, 0.01845257568359375, 0.018327455520629882, 0.019276800155639647, 0.0188272647857666, 0.018405376434326173, 0.018327552795410155, 0.018354175567626953, 0.018735103607177735, 0.018611200332641603, 0.01859689521789551, 0.018347999572753907, 0.018498559951782227, 0.018305023193359374, 0.01843404769897461, 0.018299903869628906, 0.018566144943237304, 0.018456575393676757, 0.018527231216430663, 0.01842483139038086, 0.018387968063354493, 0.01848320007324219, 0.037495807647705076, 0.018420736312866212, 0.018413568496704103, 0.018404352188110353, 0.018437120437622072, 0.018374656677246092, 0.018511871337890624, 0.018328575134277342, 0.01845145606994629, 0.01863270378112793, 0.018440191268920898, 0.018343936920166014, 0.018497535705566406, 0.01846886444091797, 0.01861324882507324, 0.01860505676269531, 0.018373632431030275, 0.017717248916625978, 0.017743871688842772, 0.018053119659423827, 0.017613824844360353, 0.01825484848022461, 0.01839411163330078, 0.01843097686767578, 0.018391040802001952, 0.019404800415039062, 0.018715648651123046, 0.018421760559082033, 0.018264064788818358, 0.01861222457885742, 0.018321407318115233, 0.01835212707519531, 0.01843404769897461, 0.01838489532470703, 0.01836031913757324, 0.018482175827026368, 0.018481151580810547, 0.01842483139038086, 0.018461696624755858, 0.01841663932800293, 0.01841152000427246, 0.018342912673950194, 0.01844121551513672, 0.018350080490112306, 0.018337888717651366, 0.018459552764892577, 0.01817190361022949, 0.01838591957092285, 0.018347007751464844, 0.01825279998779297, 0.01843404769897461, 0.018315263748168945, 0.01836953544616699, 0.01835212707519531, 0.018473983764648438, 0.018785280227661134, 0.01880473518371582, 0.018529279708862305, 0.018333759307861328, 0.0177509765625, 0.017853439331054686, 0.01841459274291992, 0.018328575134277342, 0.03745280075073242, 0.018158592224121094, 0.018306047439575195, 0.017712127685546874, 0.017868799209594728, 0.018305023193359374, 0.018338815689086914, 0.018469888687133788, 0.01862451171875, 0.018423807144165038, 0.01845964813232422, 0.018328575134277342, 0.0186060791015625, 0.018348031997680665, 0.01840742492675781, 0.018330623626708984, 0.01767628860473633, 0.018075647354125975, 0.018358272552490236, 0.018349056243896485, 0.01839923286437988, 0.018478080749511717, 0.018331647872924805, 0.018332672119140626, 0.017779712677001954, 0.018081823348999024, 0.01833776092529297, 0.018529279708862305, 0.018341888427734376, 0.018354175567626953, 0.01882316780090332, 0.018474048614501953, 0.01814316749572754, 0.018331647872924805, 0.018239488601684572, 0.018315263748168945, 0.01841049575805664, 0.018364416122436524, 0.018361343383789062, 0.018347007751464844, 0.01830297660827637, 0.01841868782043457, 0.018295808792114256, 0.018610176086425782, 0.01777564811706543, 0.01790665626525879, 0.017731584548950196, 0.01764659118652344, 0.01765068817138672, 0.017717248916625978, 0.01761996841430664, 0.017583103179931642, 0.01764761543273926, 0.017758207321166994, 0.017632255554199217, 0.017614912033081055, 0.01757792091369629, 0.017667072296142578, 0.01765273666381836, 0.017661951065063478, 0.017580032348632812, 0.017695743560791014, 0.017530879974365234, 0.03600598526000977, 0.017718175888061523, 0.01760665512084961, 0.01776646423339844, 0.017686464309692382, 0.017617919921875, 0.017612800598144532, 0.017733631134033204, 0.017711103439331053, 0.017723392486572266, 0.017707008361816406, 0.017661951065063478, 0.017656831741333007, 0.017796096801757814, 0.01778892707824707, 0.017754112243652344, 0.017695743560791014, 0.017737728118896484, 0.017678335189819337, 0.017707008361816406, 0.01842585563659668, 0.017943552017211914, 0.01766092872619629, 0.01767731285095215, 0.017688575744628905, 0.017509376525878906, 0.017566719055175782, 0.01760051155090332, 0.017597536087036132, 0.017682336807250978, 0.017716224670410157, 0.017674240112304687, 0.017622016906738282, 0.01778892707824707, 0.017558528900146485, 0.017657855987548828, 0.017709056854248048, 0.01763737678527832, 0.017666048049926757, 0.01764556884765625, 0.017566719055175782, 0.017565696716308594, 0.01762303924560547, 0.017633279800415038, 0.017520639419555666, 0.017719295501708983, 0.017745920181274414, 0.01764659118652344, 0.01767628860473633, 0.017572864532470703, 0.01762099266052246, 0.01758720016479492, 0.017665023803710937, 0.017582080841064454, 0.01757798385620117, 0.017562623977661132, 0.01767731285095215, 0.01780735969543457, 0.01784934425354004, 0.01785753631591797, 0.017752063751220702, 0.017670143127441407, 0.017667072296142578, 0.03609190368652344, 0.01759235191345215, 0.01767830467224121, 0.01763532829284668, 0.017588224411010742, 0.017705984115600586, 0.01764761543273926, 0.01763839912414551, 0.017687551498413084, 0.017864704132080078, 0.017716224670410157, 0.017690624237060547, 0.017709056854248048, 0.01765478324890137, 0.01802137565612793, 0.017870847702026366, 0.01776742362976074, 0.01758720016479492, 0.01761075210571289, 0.01780019187927246, 0.017625087738037108, 0.017641471862792968, 0.017581056594848633, 0.01764249610900879, 0.01765888023376465, 0.0175883846282959, 0.01771811294555664, 0.01763942337036133, 0.01780121612548828, 0.01765888023376465, 0.01761894416809082, 0.017680448532104494, 0.01775507164001465, 0.01761587142944336, 0.01756159973144531, 0.017590272903442384, 0.01758515167236328, 0.017756256103515625, 0.01760041618347168, 0.01761484718322754, 0.01757695960998535, 0.017717248916625978, 0.0176680965423584, 0.01763839912414551, 0.017666048049926757, 0.017675264358520508, 0.01780019187927246, 0.017612800598144532, 0.017725439071655275, 0.017697792053222656, 0.017674240112304687, 0.01773465538024902, 0.017765375137329103, 0.017741823196411134, 0.017726495742797853, 0.017797088623046874, 0.01765888023376465, 0.017688575744628905, 0.01758720016479492, 0.017582080841064454, 0.01760972785949707, 0.017560575485229494, 0.017673215866088866, 0.03602022552490235, 0.017697792053222656, 0.017649663925170898, 0.01762099266052246, 0.017605632781982423, 0.017727487564086913, 0.017672191619873046, 0.017661951065063478, 0.0177838077545166, 0.017710079193115236, 0.017709056854248048, 0.017693695068359376, 0.017798240661621095, 0.017563552856445314, 0.017704959869384765, 0.017745920181274414, 0.017758207321166994, 0.017649663925170898, 0.017571840286254883, 0.017674240112304687, 0.017590272903442384, 0.017656831741333007, 0.01762099266052246, 0.01760870361328125, 0.0176629753112793, 0.017673215866088866, 0.017748992919921876, 0.017605632781982423, 0.01784832000732422, 0.017720319747924804, 0.01764352035522461, 0.01763020706176758, 0.017625087738037108, 0.0175994873046875, 0.01763020706176758, 0.0175994873046875, 0.017622016906738282, 0.01761894416809082, 0.017666048049926757, 0.0176680965423584, 0.017762304306030274, 0.01761894416809082, 0.01759231948852539, 0.017727487564086913, 0.017819711685180664, 0.01775507164001465, 0.017670143127441407, 0.01758720016479492, 0.0176312313079834, 0.01764352035522461, 0.01761075210571289, 0.01763020706176758, 0.01787392044067383, 0.01762303924560547, 0.017747968673706056, 0.017694719314575197, 0.017698816299438477, 0.01765990447998047, 0.017681407928466796, 0.01762099266052246, 0.01784217643737793, 0.017648639678955077, 0.01760767936706543, 0.036905025482177736, 0.017848255157470704, 0.01768448066711426, 0.01759436798095703, 0.017740800857543947, 0.01774284744262695, 0.017714176177978515, 0.017804288864135744, 0.01757798385620117, 0.017640447616577147, 0.017812543869018555, 0.017627071380615235, 0.01768448066711426, 0.017779712677001954, 0.017781759262084963, 0.01764249610900879, 0.017755136489868165, 0.017580032348632812, 0.017670143127441407, 0.017572864532470703, 0.017741823196411134, 0.017833984375, 0.017663999557495116, 0.017683456420898438, 0.0176363525390625, 0.01761075210571289, 0.017699840545654297, 0.01761587142944336, 0.017729536056518554, 0.01765171241760254, 0.01763430404663086, 0.017655839920043947, 0.017599456787109374, 0.017699840545654297, 0.017681407928466796, 0.01760358428955078, 0.01780019187927246, 0.01845964813232422, 0.018276351928710938, 0.018343936920166014, 0.01838591957092285, 0.018339839935302735, 0.01847500801086426, 0.017976320266723633, 0.01766092872619629, 0.017718271255493166, 0.017812480926513673, 0.01762918472290039, 0.017655807495117186, 0.01764556884765625, 0.017613824844360353, 0.01763839912414551, 0.01760256004333496, 0.017648639678955077, 0.01765888023376465, 0.0176629753112793, 0.017763328552246094, 0.01764361572265625, 0.017587104797363282, 0.01762303924560547, 0.01757900810241699, 0.017698816299438477, 0.017690624237060547]",tokens/s,54.51551933403714,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,huggyllama/llama-65b,huggyllama/llama-65b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ self.load_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 250, in load_model_with_no_weights self.pretrained_model = self.automodel_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3710, in from_pretrained model = cls(config, *model_args, **model_kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 1101, in __init__ self.model = LlamaModel(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 884, in __init__ [LlamaDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 884, in [LlamaDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 677, in __init__ self.self_attn = LLAMA_ATTENTION_CLASSES[config._attn_implementation](config=config, layer_idx=layer_idx) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 264, in __init__ self.v_proj = nn.Linear(self.hidden_size, self.num_key_value_heads * self.head_dim, bias=config.attention_bias) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/linear.py"", line 98, in __init__ self.weight = Parameter(torch.empty((out_features, in_features), **factory_kwargs)) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_device.py"", line 78, in __torch_function__ return func(*args, **kwargs) torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 128.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,l,l,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/l/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66947455-584228bf50d38f264d1608e5;64741d5e-945a-417c-b062-5d1bd8c01159) Repository Not Found for url: https://huggingface.co/l/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: l is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,B,B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/B/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66932508-78894615374a64d13bb62ed9;4774bf55-1dd2-486b-af8a-17c792db2def) Repository Not Found for url: https://huggingface.co/B/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: B is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,huggyllama/llama-30b,huggyllama/llama-30b,cuda,0,42,,,True,,,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",llama,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,meta-llama/Llama-2-13b-hf,meta-llama/Llama-2-13b-hf,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ self.load_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 250, in load_model_with_no_weights self.pretrained_model = self.automodel_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3710, in from_pretrained model = cls(config, *model_args, **model_kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 1101, in __init__ self.model = LlamaModel(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 884, in __init__ [LlamaDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 884, in [LlamaDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 679, in __init__ self.mlp = LlamaMLP(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 192, in __init__ self.gate_proj = nn.Linear(self.hidden_size, self.intermediate_size, bias=config.mlp_bias) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/linear.py"", line 98, in __init__ self.weight = Parameter(torch.empty((out_features, in_features), **factory_kwargs)) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_device.py"", line 78, in __torch_function__ return func(*args, **kwargs) torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 136.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-7B,,cuda,0,42,,,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,217063f5c507ed7cc255df7e1f64c4333a0b4dfe,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,,MB,1297.969152,17102.798848,0.0,16456.351744,16359.853056,s,10,1.0316815719604493,0.10316815719604491,0.0002858635983448075,0.10308147048950195,0.10334511642456055,0.10366564598083496,0.10392206962585449,"[0.10398617553710937, 0.10306684875488281, 0.1030960922241211, 0.10312809753417969, 0.10292483520507813, 0.1030156478881836, 0.10302108764648438, 0.10305548858642578, 0.10311341094970704, 0.10327388763427735]",tokens/s,2481.385797301167,kWh,1.2201533830998827e-06,6.685899876493308e-07,5.9327739329280695e-06,7.821517303677283e-06,tokens/kWh,32730222.290711,MB,1297.969152,17102.798848,0.0,16456.351744,16416.242176,s,10,23.347927978515628,2.334792797851563,0.0034280681689723264,2.3335817871093747,2.337003662109375,2.340813720703125,2.343861767578125,"[2.344623779296875, 2.333737060546875, 2.33296240234375, 2.332305419921875, 2.336156982421875, 2.333426513671875, 2.332965576171875, 2.3332919921875, 2.3338671875, 2.334591064453125]",tokens/s,26.983122467214883,kWh,2.7499233605116095e-05,1.5070586503284129e-05,0.0001175103525992714,0.0001600801727076716,tokens/kWh,393552.79879068263,,s,629,24.23761405563358,0.03853356765601517,0.012321844083848062,0.037031936645507815,0.03714355087280274,0.03729612808227539,0.14069289001464844,"[0.03764432144165039, 0.037763038635253904, 0.03763711929321289, 0.03765657424926758, 0.03772313690185547, 0.037664768218994144, 0.03767091369628906, 0.037716991424560545, 0.037797889709472655, 0.03747945785522461, 0.037622753143310546, 0.037645313262939455, 0.03738521575927734, 0.03702579116821289, 0.03698995208740234, 0.036985855102539066, 0.036997119903564454, 0.03695001602172852, 0.03700428771972656, 0.03702169418334961, 0.03701145553588867, 0.03698483276367188, 0.03700838470458984, 0.037005313873291014, 0.036994049072265625, 0.03697971343994141, 0.03704524612426758, 0.03710464096069336, 0.03711180877685547, 0.03726131057739258, 0.03727052688598633, 0.037032958984375, 0.03700735855102539, 0.037064704895019535, 0.03705753707885742, 0.0370513916015625, 0.03701964950561523, 0.03704217529296875, 0.03706985473632812, 0.037058528900146485, 0.0370145263671875, 0.037154815673828126, 0.03728179168701172, 0.03710464096069336, 0.037187583923339845, 0.037100543975830076, 0.03710566329956055, 0.03720191955566406, 0.0371517448425293, 0.0370780143737793, 0.03706572723388672, 0.037108734130859376, 0.03710976028442383, 0.037082111358642575, 0.037103614807128905, 0.03712819290161133, 0.037190654754638675, 0.037100543975830076, 0.037100543975830076, 0.03712819290161133, 0.0370882568359375, 0.037119998931884765, 0.14084197998046874, 0.03694182586669922, 0.036893695831298826, 0.0369090576171875, 0.03691417694091797, 0.036942848205566405, 0.036949024200439454, 0.03697865676879883, 0.036947967529296875, 0.03698483276367188, 0.036969470977783206, 0.036982784271240236, 0.036975616455078124, 0.037026817321777344, 0.03704934310913086, 0.03696332931518555, 0.03697049713134765, 0.03696537780761719, 0.036992000579833983, 0.037012481689453126, 0.036961280822753906, 0.03697663879394531, 0.037015552520751956, 0.03709542465209961, 0.037125118255615236, 0.037087230682373046, 0.03700940704345703, 0.03706265640258789, 0.03699609756469727, 0.036994049072265625, 0.03698390579223633, 0.03700624084472656, 0.036961280822753906, 0.03701145553588867, 0.03718860626220703, 0.037341182708740234, 0.03707904052734375, 0.03705651092529297, 0.03700940704345703, 0.03700940704345703, 0.03709648132324219, 0.037037025451660155, 0.03700844955444336, 0.03702880096435547, 0.03706265640258789, 0.03703807830810547, 0.037064704895019535, 0.03703091049194336, 0.0370513916015625, 0.03708006286621094, 0.03705753707885742, 0.03709542465209961, 0.03706265640258789, 0.037070846557617186, 0.03704729461669922, 0.03712102508544922, 0.03710259246826172, 0.03708108901977539, 0.037133312225341795, 0.0370964469909668, 0.03709337615966797, 0.03706982421875, 0.037147647857666014, 0.14093618774414063, 0.03697663879394531, 0.036964351654052735, 0.03702169418334961, 0.036977664947509765, 0.03695718383789062, 0.036937759399414065, 0.03700323104858398, 0.037026817321777344, 0.036947967529296875, 0.03691622543334961, 0.03695820617675781, 0.036977664947509765, 0.036980735778808595, 0.03702169418334961, 0.036992000579833983, 0.036972545623779295, 0.036945919036865234, 0.0370145263671875, 0.03695315170288086, 0.036950977325439456, 0.03697356796264648, 0.036967422485351564, 0.03697049713134765, 0.03697868728637695, 0.03699609756469727, 0.03701251220703125, 0.037018592834472654, 0.03703091049194336, 0.03703398513793945, 0.036999168395996096, 0.03703500747680664, 0.03705855941772461, 0.036992000579833983, 0.03702783966064453, 0.037006401062011716, 0.037001152038574216, 0.03702067184448242, 0.03702889633178711, 0.037015518188476565, 0.037026817321777344, 0.03704627227783203, 0.037031936645507815, 0.0370513916015625, 0.0370513916015625, 0.03706367874145508, 0.037043201446533204, 0.03709337615966797, 0.037015552520751956, 0.037054462432861326, 0.03705548858642578, 0.03706880187988281, 0.03702998352050781, 0.03703081512451172, 0.03711795043945312, 0.03706367874145508, 0.03707289505004883, 0.03706675338745117, 0.03709030532836914, 0.03709952163696289, 0.03707289505004883, 0.03709132766723633, 0.037160961151123044, 0.14081639099121093, 0.036934654235839845, 0.036910079956054685, 0.03691424179077148, 0.036932544708251955, 0.03693875122070313, 0.03694079971313476, 0.03696640014648438, 0.036951038360595705, 0.036994049072265625, 0.03695206451416016, 0.03693670272827149, 0.03699507141113281, 0.03695718383789062, 0.036928512573242187, 0.03695206451416016, 0.03703807830810547, 0.036967422485351564, 0.036945919036865234, 0.036944896697998046, 0.03699814224243164, 0.036951038360595705, 0.03695513534545898, 0.036975616455078124, 0.037031936645507815, 0.037012481689453126, 0.036961280822753906, 0.036956161499023435, 0.037101566314697264, 0.037103614807128905, 0.036992000579833983, 0.037012481689453126, 0.0370513916015625, 0.037005313873291014, 0.03699507141113281, 0.0370063362121582, 0.03708313751220703, 0.037026817321777344, 0.037015552520751956, 0.03701760101318359, 0.03703807830810547, 0.037028865814208986, 0.03701964950561523, 0.036999168395996096, 0.03704524612426758, 0.03703603363037109, 0.03701760101318359, 0.0370513916015625, 0.03707596969604492, 0.03703398513793945, 0.0370247688293457, 0.03702374267578125, 0.03705548858642578, 0.03708313751220703, 0.037043201446533204, 0.03712102508544922, 0.03710566329956055, 0.03708313751220703, 0.037059585571289064, 0.03709030532836914, 0.03706060791015625, 0.03708006286621094, 0.03710464096069336, 0.1406781463623047, 0.0369356803894043, 0.03690496063232422, 0.03687628936767578, 0.03694079971313476, 0.036945919036865234, 0.036918270111083985, 0.0369172477722168, 0.03697868728637695, 0.03707392120361328, 0.03714355087280274, 0.036956161499023435, 0.037002239227294925, 0.036961280822753906, 0.03694387054443359, 0.036951038360595705, 0.0369879035949707, 0.03704012680053711, 0.036939777374267575, 0.037004318237304684, 0.037002208709716794, 0.036969470977783206, 0.03695513534545898, 0.03702272033691406, 0.03696844863891602, 0.03700735855102539, 0.036967422485351564, 0.03703910446166992, 0.03711795043945312, 0.037043201446533204, 0.0370247688293457, 0.037010433197021485, 0.03704627227783203, 0.03705855941772461, 0.037002239227294925, 0.03704934310913086, 0.03711590576171875, 0.03730636978149414, 0.03718656158447266, 0.037070846557617186, 0.03702067184448242, 0.0370513916015625, 0.03704934310913086, 0.037131263732910154, 0.03704524612426758, 0.03706675338745117, 0.03709750366210938, 0.0372408332824707, 0.03707183837890625, 0.037154815673828126, 0.03719987106323242, 0.03711078262329102, 0.03705036926269531, 0.037222400665283206, 0.03783679962158203, 0.03758182525634766, 0.037302272796630856, 0.037326847076416016, 0.03707904052734375, 0.03711078262329102, 0.037098495483398435, 0.03709132766723633, 0.03709132766723633, 0.14067097473144533, 0.036929534912109374, 0.03692134475708008, 0.03688857650756836, 0.03690598297119141, 0.03700735855102539, 0.036959232330322264, 0.03698175811767578, 0.03690393447875977, 0.03695315170288086, 0.03698271942138672, 0.036926464080810545, 0.036913150787353514, 0.036928512573242187, 0.036979774475097656, 0.03693766403198242, 0.03695001602172852, 0.036913150787353514, 0.03695718383789062, 0.03700428771972656, 0.037103614807128905, 0.0370882568359375, 0.036994049072265625, 0.03695513534545898, 0.03705855941772461, 0.037085182189941404, 0.037029888153076174, 0.03695718383789062, 0.03699507141113281, 0.03702783966064453, 0.03706060791015625, 0.03711283111572266, 0.037101566314697264, 0.03706982421875, 0.03702169418334961, 0.037064704895019535, 0.03701145553588867, 0.037082111358642575, 0.037136383056640625, 0.03704729461669922, 0.037031936645507815, 0.03706367874145508, 0.037028865814208986, 0.03703807830810547, 0.03701862335205078, 0.0370780143737793, 0.03705036926269531, 0.03707289505004883, 0.03707494354248047, 0.03707596969604492, 0.03717836761474609, 0.037256191253662106, 0.03707289505004883, 0.037087230682373046, 0.037131263732910154, 0.03706367874145508, 0.03707904052734375, 0.037048320770263675, 0.03702272033691406, 0.03710259246826172, 0.037116928100585936, 0.037087230682373046, 0.037106689453125, 0.14069862365722657, 0.03696844863891602, 0.03688652801513672, 0.036893695831298826, 0.03691212844848633, 0.036928512573242187, 0.03691526412963867, 0.03693049621582031, 0.0369244155883789, 0.03696230316162109, 0.036942848205566405, 0.036910079956054685, 0.036934654235839845, 0.036951038360595705, 0.03695718383789062, 0.036931583404541016, 0.03693670272827149, 0.0369879035949707, 0.03695001602172852, 0.036985855102539066, 0.03702783966064453, 0.03713945770263672, 0.03702783966064453, 0.03705344009399414, 0.036972545623779295, 0.03703910446166992, 0.03699097442626953, 0.03700940704345703, 0.03700121688842774, 0.03721932983398438, 0.03701760101318359, 0.0369879035949707, 0.037010433197021485, 0.03705753707885742, 0.03703807830810547, 0.03701862335205078, 0.03708108901977539, 0.03707596969604492, 0.03706060791015625, 0.03703398513793945, 0.03700121688842774, 0.037048320770263675, 0.03703807830810547, 0.03704934310913086, 0.037070846557617186, 0.03707596969604492, 0.037048320770263675, 0.03709036636352539, 0.03710867309570313, 0.03710976028442383, 0.03707187271118164, 0.037087230682373046, 0.03706880187988281, 0.03706982421875, 0.03706060791015625, 0.03709439849853516, 0.03705548858642578, 0.03709542465209961, 0.03706265640258789, 0.03716505432128906, 0.037064704895019535, 0.03708927917480469, 0.03703919982910156, 0.1407169647216797, 0.03688857650756836, 0.03689471817016601, 0.03688550567626953, 0.036910079956054685, 0.03694899368286133, 0.036918270111083985, 0.03693059158325195, 0.03693052673339844, 0.036985855102539066, 0.03700326538085937, 0.036980735778808595, 0.03696230316162109, 0.037007423400878904, 0.036967361450195316, 0.036947967529296875, 0.03702169418334961, 0.03698688125610351, 0.036956161499023435, 0.036977664947509765, 0.036961280822753906, 0.037000190734863284, 0.03704012680053711, 0.037002239227294925, 0.03697356796264648, 0.0370513916015625, 0.03700428771972656, 0.037000190734863284, 0.03700428771972656, 0.03714355087280274, 0.03699609756469727, 0.03700940704345703, 0.03699507141113281, 0.03702272033691406, 0.037002239227294925, 0.037012481689453126, 0.03709337615966797, 0.03709542465209961, 0.0370882568359375, 0.03703091049194336, 0.037032958984375, 0.03714355087280274, 0.03703705596923828, 0.03713433456420898, 0.03708313751220703, 0.03708313751220703, 0.037059585571289064, 0.03708006286621094, 0.03706265640258789, 0.03706880187988281, 0.037144577026367184, 0.037059585571289064, 0.03706777572631836, 0.03707904052734375, 0.037068832397460935, 0.037074913024902345, 0.03707494354248047, 0.03709542465209961, 0.03706880187988281, 0.037119998931884765, 0.03708620834350586, 0.037085182189941404, 0.03719168090820312, 0.14086860656738281, 0.03693670272827149, 0.03689471817016601, 0.036898815155029296, 0.03698995208740234, 0.036945919036865234, 0.03689779281616211, 0.036951038360595705, 0.03698483276367188, 0.03694899368286133, 0.03696025466918945, 0.036953086853027346, 0.0370063362121582, 0.036982784271240236, 0.036951072692871095, 0.03691312026977539, 0.036961280822753906, 0.03705548858642578, 0.03698175811767578, 0.036988929748535154, 0.03696230316162109, 0.036985855102539066, 0.03697868728637695, 0.03695820617675781, 0.036992000579833983, 0.03703603363037109, 0.03704940795898438, 0.0369950065612793, 0.03700326538085937, 0.03703705596923828, 0.036992000579833983, 0.03703094482421875, 0.03701961517333984, 0.03702169418334961, 0.036992000579833983, 0.03700121688842774, 0.03713228988647461, 0.037054462432861326, 0.037013504028320314, 0.03700735855102539, 0.03702272033691406, 0.03704524612426758, 0.03699097442626953, 0.037130241394042966, 0.03733606338500976, 0.03712102508544922, 0.03712409591674805, 0.03708006286621094, 0.03706163024902344, 0.03706163024902344, 0.03701760101318359, 0.037132350921630856, 0.037069759368896484, 0.03706265640258789, 0.037048320770263675, 0.03711590576171875, 0.037179393768310545, 0.03737702560424805, 0.03711897659301758, 0.03707494354248047, 0.037103614807128905, 0.03722649765014648, 0.03718348693847656, 0.1409617919921875, 0.03692031860351563, 0.03689779281616211, 0.036915199279785156, 0.03690496063232422, 0.036934654235839845, 0.03690393447875977, 0.03700838470458984, 0.03692236709594727, 0.03696025466918945, 0.03695513534545898, 0.036994049072265625, 0.03700940704345703, 0.036982784271240236, 0.03694079971313476, 0.03701766586303711, 0.03702982330322266, 0.036939777374267575, 0.03697049713134765, 0.0370247688293457, 0.036988929748535154, 0.03696844863891602, 0.03695513534545898, 0.03698995208740234, 0.03698995208740234, 0.036964351654052735, 0.03716201782226562, 0.03732579040527344, 0.037166080474853515, 0.03703910446166992, 0.03698995208740234, 0.03705344009399414, 0.03705241775512695, 0.037000190734863284, 0.03699302291870117, 0.037064704895019535, 0.03705344009399414, 0.03704217529296875, 0.0373831672668457, 0.03728691101074219, 0.03709439849853516, 0.03712204742431641, 0.037125118255615236, 0.037101566314697264, 0.03703398513793945, 0.037087230682373046, 0.03706163024902344, 0.03706367874145508, 0.03706572723388672, 0.037070846557617186, 0.03707187271118164, 0.037163009643554686, 0.03707289505004883, 0.03718659210205078, 0.037117919921875, 0.03708313751220703, 0.037114879608154294, 0.03711180877685547, 0.037108734130859376, 0.037108734130859376, 0.03711078262329102, 0.037101566314697264, 0.03708313751220703]",tokens/s,25.951399281968584,,,main,False,False,,,,,,,,,,,,,,,,,,,,,,, float16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,M,M,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/M/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6693229b-314bffff461549c804b0290b;e60fc87f-8544-4e72-809a-3f0e4f4c8f1e) Repository Not Found for url: https://huggingface.co/M/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: M is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,stabilityai/stablelm-base-alpha-3b,stabilityai/stablelm-base-alpha-3b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.42.3,,0.31.0,,,,1.20.0,,,,0.11.1,,,,MB,1290.100736,8529.641472,0.0,7883.194368,7764.230656,s,10,0.42583157730102533,0.04258315773010254,0.00026664190842415537,0.042499326705932615,0.042636421966552734,0.043006819152832026,0.043303136901855464,"[0.043377216339111326, 0.04255411148071289, 0.04250620651245117, 0.04252684783935547, 0.0424411506652832, 0.042481983184814456, 0.04249244689941406, 0.04244927978515625, 0.042485118865966794, 0.042517215728759765]",tokens/s,6011.766474026199,kWh,5.021194525750569e-07,2.751298363198087e-07,2.732881188186325e-06,3.51013047708119e-06,tokens/kWh,72931761.84518188,MB,1290.100736,8529.641472,0.0,7883.194368,7840.651264,s,10,11.050873046875,1.1050873046874998,0.0011993969748205176,1.1047346801757811,1.10561923828125,1.1070929443359374,1.1082719091796875,"[1.108566650390625, 1.10418017578125, 1.104793212890625, 1.1050654296875, 1.1046761474609375, 1.1045972900390626, 1.104452880859375, 1.1044144287109374, 1.1048350830078124, 1.105291748046875]",tokens/s,57.00907044427168,kWh,1.3056253617850398e-05,7.1545030364062364e-06,6.128375335841789e-05,8.149451001267454e-05,tokens/kWh,773058.2095677592,,s,629,11.415499748229976,0.018148648248378346,0.005067467928377624,0.017535999298095704,0.01759050178527832,0.017624064254760742,0.060142386016845704,"[0.017986560821533205, 0.01797324752807617, 0.01793340873718262, 0.018050975799560547, 0.017483776092529296, 0.017472511291503907, 0.017489919662475584, 0.017489919662475584, 0.017501184463500977, 0.017501184463500977, 0.017475584030151366, 0.017488895416259767, 0.017469440460205078, 0.017489919662475584, 0.017537023544311522, 0.017523712158203125, 0.017494016647338868, 0.017512447357177736, 0.017572864532470703, 0.017547264099121093, 0.017520639419555666, 0.017490943908691405, 0.017494016647338868, 0.017491968154907226, 0.017550336837768556, 0.017538047790527343, 0.017510400772094727, 0.017517568588256836, 0.017501184463500977, 0.017574911117553712, 0.017693695068359376, 0.017886207580566405, 0.017597440719604493, 0.017543167114257813, 0.017557504653930665, 0.01757699203491211, 0.01760867118835449, 0.017667072296142578, 0.017560575485229494, 0.017580032348632812, 0.017738752365112305, 0.017580032348632812, 0.017564672470092774, 0.01754521560668945, 0.017550336837768556, 0.01756064033508301, 0.01755027198791504, 0.017575935363769533, 0.017548288345336914, 0.017557504653930665, 0.01764249610900879, 0.017574911117553712, 0.017580032348632812, 0.01755340766906738, 0.017559551239013673, 0.017583103179931642, 0.017682432174682617, 0.017590303421020508, 0.017572832107543946, 0.01759846305847168, 0.01763942337036133, 0.01761484718322754, 0.06020403289794922, 0.017459199905395507, 0.017451007843017577, 0.017472511291503907, 0.017480703353881837, 0.017472511291503907, 0.017482751846313475, 0.017509376525878906, 0.01747148895263672, 0.017473535537719728, 0.01745510482788086, 0.017505279541015627, 0.017487871170043946, 0.017515520095825195, 0.017464319229125978, 0.017552383422851564, 0.01746227264404297, 0.017496063232421876, 0.017494016647338868, 0.017508352279663086, 0.017483776092529296, 0.017527807235717775, 0.017491968154907226, 0.017520639419555666, 0.017503231048583985, 0.017520639419555666, 0.017514495849609374, 0.017512447357177736, 0.017552383422851564, 0.017544191360473634, 0.017507328033447265, 0.017557504653930665, 0.017530879974365234, 0.01755340766906738, 0.017515615463256837, 0.01754313659667969, 0.017556480407714844, 0.01753286361694336, 0.017517568588256836, 0.017541120529174805, 0.017522687911987304, 0.01754521560668945, 0.017527807235717775, 0.017543167114257813, 0.017537023544311522, 0.017551359176635743, 0.017562623977661132, 0.017539072036743163, 0.017540096282958984, 0.017580032348632812, 0.017552383422851564, 0.017580032348632812, 0.017548288345336914, 0.017562623977661132, 0.017565696716308594, 0.017560575485229494, 0.017557504653930665, 0.017589248657226563, 0.017550336837768556, 0.01757798385620117, 0.01757695960998535, 0.017575935363769533, 0.01757695960998535, 0.06051430511474609, 0.017483776092529296, 0.017474559783935546, 0.017474559783935546, 0.017459199905395507, 0.01748684883117676, 0.017473600387573243, 0.01748371124267578, 0.017480703353881837, 0.017494016647338868, 0.017513471603393553, 0.017489919662475584, 0.017481727600097655, 0.017477632522583008, 0.017512447357177736, 0.017511423110961915, 0.017494016647338868, 0.017512447357177736, 0.017482751846313475, 0.017508352279663086, 0.01748684883117676, 0.017509376525878906, 0.017493120193481447, 0.017488767623901367, 0.017491968154907226, 0.017490943908691405, 0.017504255294799806, 0.017515520095825195, 0.017513471603393553, 0.017552383422851564, 0.017503231048583985, 0.017523712158203125, 0.017520639419555666, 0.017518592834472657, 0.017538047790527343, 0.017527807235717775, 0.017513471603393553, 0.01755340766906738, 0.017527807235717775, 0.017540096282958984, 0.017537023544311522, 0.017530879974365234, 0.01757798385620117, 0.017542144775390626, 0.017612800598144532, 0.01756876754760742, 0.01754521560668945, 0.0175861759185791, 0.017582080841064454, 0.017596416473388672, 0.017558528900146485, 0.017559551239013673, 0.017588224411010742, 0.017555456161499023, 0.017583103179931642, 0.017588224411010742, 0.017596416473388672, 0.017589248657226563, 0.017560575485229494, 0.017555456161499023, 0.017613824844360353, 0.017573888778686524, 0.01759846305847168, 0.060217342376708984, 0.01745408058166504, 0.017458175659179686, 0.01749504089355469, 0.01765171241760254, 0.017540096282958984, 0.017491968154907226, 0.017468416213989257, 0.017736703872680663, 0.017537023544311522, 0.017514495849609374, 0.017510400772094727, 0.017518592834472657, 0.017515520095825195, 0.017496095657348634, 0.017481695175170897, 0.017498111724853514, 0.01759539222717285, 0.017497087478637697, 0.017488895416259767, 0.017506303787231444, 0.017523712158203125, 0.017501184463500977, 0.017502208709716797, 0.017513471603393553, 0.017512447357177736, 0.017533952713012696, 0.017505279541015627, 0.017510400772094727, 0.017534975051879884, 0.017505279541015627, 0.017535999298095704, 0.017539072036743163, 0.017509376525878906, 0.017521663665771483, 0.017540096282958984, 0.017521663665771483, 0.017556480407714844, 0.01756159973144531, 0.017549312591552735, 0.017502208709716797, 0.017528831481933595, 0.017563648223876953, 0.017525760650634766, 0.017617919921875, 0.017563648223876953, 0.017551359176635743, 0.017572864532470703, 0.017532928466796875, 0.017556480407714844, 0.01757900810241699, 0.017543167114257813, 0.017562623977661132, 0.01759436798095703, 0.017567743301391603, 0.01756159973144531, 0.017566719055175782, 0.01759334373474121, 0.017572864532470703, 0.017558528900146485, 0.017583103179931642, 0.0175994873046875, 0.017588224411010742, 0.060112895965576174, 0.01744179153442383, 0.01745510482788086, 0.01747865676879883, 0.017481727600097655, 0.017447935104370118, 0.017515520095825195, 0.01746227264404297, 0.017498111724853514, 0.017506303787231444, 0.017476608276367187, 0.017491968154907226, 0.01747148895263672, 0.017485824584960938, 0.017539072036743163, 0.017528831481933595, 0.017496063232421876, 0.017494016647338868, 0.017490943908691405, 0.017509376525878906, 0.017464319229125978, 0.017514495849609374, 0.017513471603393553, 0.017486879348754883, 0.017535968780517577, 0.017516544342041016, 0.017499135971069335, 0.017528831481933595, 0.017531904220581054, 0.017527807235717775, 0.017539072036743163, 0.017516544342041016, 0.017533952713012696, 0.017546239852905272, 0.017519615173339845, 0.017539072036743163, 0.017539072036743163, 0.017559551239013673, 0.017547264099121093, 0.017514495849609374, 0.017560575485229494, 0.017541120529174805, 0.017535999298095704, 0.017567743301391603, 0.017555456161499023, 0.01760051155090332, 0.01758720016479492, 0.017564672470092774, 0.01756979179382324, 0.017571840286254883, 0.017574975967407227, 0.017547199249267578, 0.017558528900146485, 0.01760767936706543, 0.017578079223632814, 0.017565599441528322, 0.017571840286254883, 0.017580032348632812, 0.017604608535766602, 0.01757695960998535, 0.017597440719604493, 0.017707008361816406, 0.01756979179382324, 0.06015385437011719, 0.017485824584960938, 0.01743974494934082, 0.017476608276367187, 0.017483776092529296, 0.017469440460205078, 0.01748684883117676, 0.01747865676879883, 0.01744486427307129, 0.017523712158203125, 0.017474559783935546, 0.017488895416259767, 0.017489919662475584, 0.017512447357177736, 0.01748796844482422, 0.017477535247802736, 0.017512447357177736, 0.017523712158203125, 0.017481727600097655, 0.017491968154907226, 0.017507328033447265, 0.017515552520751952, 0.01748476791381836, 0.017509376525878906, 0.017489919662475584, 0.017535999298095704, 0.017514495849609374, 0.017518592834472657, 0.017522687911987304, 0.017527807235717775, 0.017516544342041016, 0.017530879974365234, 0.017532928466796875, 0.017546239852905272, 0.017527807235717775, 0.017559551239013673, 0.017542144775390626, 0.01754521560668945, 0.017544191360473634, 0.017529855728149413, 0.017657855987548828, 0.017540096282958984, 0.017544191360473634, 0.017543167114257813, 0.017540096282958984, 0.017555456161499023, 0.017575935363769533, 0.017567743301391603, 0.017559551239013673, 0.017567743301391603, 0.017571840286254883, 0.017580032348632812, 0.017574911117553712, 0.01758720016479492, 0.017560575485229494, 0.017565696716308594, 0.017589248657226563, 0.017632255554199217, 0.017583103179931642, 0.017562623977661132, 0.017572864532470703, 0.01760256004333496, 0.01759231948852539, 0.06016409683227539, 0.017464319229125978, 0.01744179153442383, 0.01744076728820801, 0.017474559783935546, 0.017497087478637697, 0.0174704647064209, 0.017504255294799806, 0.017479679107666016, 0.017473535537719728, 0.017479679107666016, 0.017474559783935546, 0.017488895416259767, 0.017472511291503907, 0.017479679107666016, 0.017490943908691405, 0.017498111724853514, 0.017503231048583985, 0.017530879974365234, 0.017508352279663086, 0.017510400772094727, 0.01749504089355469, 0.017505279541015627, 0.017542144775390626, 0.017501184463500977, 0.017504255294799806, 0.01748684883117676, 0.017506303787231444, 0.017517568588256836, 0.017511423110961915, 0.017520639419555666, 0.01754323196411133, 0.01751545524597168, 0.017533952713012696, 0.017527807235717775, 0.017527807235717775, 0.017546239852905272, 0.017524736404418945, 0.017524736404418945, 0.017573888778686524, 0.017528831481933595, 0.017539072036743163, 0.017517568588256836, 0.01757695960998535, 0.01756470489501953, 0.017557472229003907, 0.01762918472290039, 0.017566719055175782, 0.017552383422851564, 0.017564672470092774, 0.017574911117553712, 0.0176629753112793, 0.01756979179382324, 0.017564672470092774, 0.017581056594848633, 0.01759539222717285, 0.017563648223876953, 0.01756979179382324, 0.017581056594848633, 0.01759539222717285, 0.017588224411010742, 0.017590272903442384, 0.017588224411010742, 0.060104705810546874, 0.01745510482788086, 0.01743667221069336, 0.01745715141296387, 0.0174704647064209, 0.017490943908691405, 0.0174704647064209, 0.017479679107666016, 0.017473535537719728, 0.017476608276367187, 0.017479679107666016, 0.017464319229125978, 0.017460224151611328, 0.017515520095825195, 0.017484800338745117, 0.017479679107666016, 0.017520639419555666, 0.017503231048583985, 0.017497087478637697, 0.01754832077026367, 0.017505247116088866, 0.017533952713012696, 0.017494016647338868, 0.017510400772094727, 0.017510400772094727, 0.017531904220581054, 0.017532928466796875, 0.017514495849609374, 0.017502208709716797, 0.017540096282958984, 0.017524736404418945, 0.01757798385620117, 0.017542144775390626, 0.017533952713012696, 0.017535999298095704, 0.017521663665771483, 0.017537023544311522, 0.01755763244628906, 0.017526655197143554, 0.017537023544311522, 0.017524736404418945, 0.017542144775390626, 0.017540096282958984, 0.01755340766906738, 0.017542144775390626, 0.01757798385620117, 0.017556480407714844, 0.017572864532470703, 0.01756159973144531, 0.01757695960998535, 0.01756876754760742, 0.017557504653930665, 0.017573888778686524, 0.017560575485229494, 0.01757695960998535, 0.017580032348632812, 0.017554431915283202, 0.017571840286254883, 0.017574911117553712, 0.017574911117553712, 0.01763020706176758, 0.01757494354248047, 0.017605600357055665, 0.060165153503417966, 0.01746224021911621, 0.01744895935058594, 0.017487871170043946, 0.01744895935058594, 0.017508352279663086, 0.01747865676879883, 0.017475584030151366, 0.017485824584960938, 0.017509376525878906, 0.017509376525878906, 0.017530879974365234, 0.017469440460205078, 0.017467391967773437, 0.017483776092529296, 0.017463327407836914, 0.0175001277923584, 0.017502208709716797, 0.017487871170043946, 0.017523712158203125, 0.017489919662475584, 0.017499135971069335, 0.017544191360473634, 0.017509376525878906, 0.017505279541015627, 0.017508352279663086, 0.017512447357177736, 0.017555456161499023, 0.017521663665771483, 0.01752579116821289, 0.01751139259338379, 0.017538047790527343, 0.017564672470092774, 0.017510400772094727, 0.017518592834472657, 0.01756979179382324, 0.017528831481933595, 0.017550336837768556, 0.017542144775390626, 0.017547264099121093, 0.017559551239013673, 0.017532928466796875, 0.017572864532470703, 0.017529855728149413, 0.017544191360473634, 0.01756979179382324, 0.017542144775390626, 0.01758515167236328, 0.01779097557067871, 0.017588224411010742, 0.01762611198425293, 0.017597440719604493, 0.01756876754760742, 0.017551359176635743, 0.017562623977661132, 0.0175994873046875, 0.01757695960998535, 0.017583103179931642, 0.01758515167236328, 0.01759334373474121, 0.0175861759185791, 0.017562688827514647, 0.017584064483642577, 0.06021017456054688, 0.017466367721557616, 0.01746329689025879, 0.01749510383605957, 0.01748371124267578, 0.017497087478637697, 0.017491968154907226, 0.017473535537719728, 0.01746227264404297, 0.017488895416259767, 0.017481727600097655, 0.017481727600097655, 0.017468416213989257, 0.017565696716308594, 0.01762099266052246, 0.017521663665771483, 0.017542240142822265, 0.017532831192016603, 0.017485824584960938, 0.017514495849609374, 0.017485824584960938, 0.017544191360473634, 0.01759129524230957, 0.01757798385620117, 0.017539072036743163, 0.01754521560668945, 0.017529855728149413, 0.01759129524230957, 0.017521663665771483, 0.017529855728149413, 0.017510400772094727, 0.017521663665771483, 0.017543167114257813, 0.017552383422851564, 0.017530879974365234, 0.017663999557495116, 0.017533952713012696, 0.017550336837768556, 0.017542144775390626, 0.017597440719604493, 0.017711103439331053, 0.017558528900146485, 0.017543167114257813, 0.017566719055175782, 0.017530879974365234, 0.01756979179382324, 0.017546239852905272, 0.01756979179382324, 0.017555551528930666, 0.01755843162536621, 0.01756159973144531, 0.01754521560668945, 0.017564672470092774, 0.01756159973144531, 0.017550336837768556, 0.017557504653930665, 0.017584127426147463, 0.017574911117553712, 0.01756979179382324, 0.017570816040039062, 0.01757798385620117, 0.017583103179931642, 0.01759436798095703]",tokens/s,55.10052243639435,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,google/recurrentgemma-9b,google/recurrentgemma-9b,cuda,0,42,,,True,,,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",recurrent_gemma,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,mistralai/Mixtral-8x22B-v0.1,mistralai/Mixtral-8x22B-v0.1,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/base.py"", line 43, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 54, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1021, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 403 Client Error: Forbidden for url: https://huggingface.co/mistralai/Mixtral-8x22B-v0.1/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 398, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 321, in hf_raise_for_status raise GatedRepoError(message, response) from e huggingface_hub.utils._errors.GatedRepoError: 403 Client Error. (Request ID: Root=1-66475083-2aa3337f4d4cab261edcd666;3893eafe-312d-4d29-8fbf-3a2f15984a7a) Cannot access gated repo for url https://huggingface.co/mistralai/Mixtral-8x22B-v0.1/resolve/main/config.json. Access to model mistralai/Mixtral-8x22B-v0.1 is restricted and you are not in the authorized list. Visit https://huggingface.co/mistralai/Mixtral-8x22B-v0.1 to ask for access. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 87, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/base.py"", line 56, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 37, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 62, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 928, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 631, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 686, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 416, in cached_file raise EnvironmentError( OSError: You are trying to access a gated repo. Make sure to have access to it at https://huggingface.co/mistralai/Mixtral-8x22B-v0.1. 403 Client Error. (Request ID: Root=1-66475083-2aa3337f4d4cab261edcd666;3893eafe-312d-4d29-8fbf-3a2f15984a7a) Cannot access gated repo for url https://huggingface.co/mistralai/Mixtral-8x22B-v0.1/resolve/main/config.json. Access to model mistralai/Mixtral-8x22B-v0.1 is restricted and you are not in the authorized list. Visit https://huggingface.co/mistralai/Mixtral-8x22B-v0.1 to ask for access. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,01-ai/Yi-6B,01-ai/Yi-6B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1262.51008,13032.226816,0.0,12385.779712,12315.692032,s,10,0.8511878662109376,0.08511878662109376,0.00029604892667445655,0.08502067184448242,0.08526025924682616,0.08561133079528808,0.08589218803405761,"[0.08596240234375, 0.08518224334716797, 0.08496489715576172, 0.0849315185546875, 0.08495494079589844, 0.08505094146728516, 0.08511305236816406, 0.08514911651611329, 0.08499040222167968, 0.0848883514404297]",tokens/s,3007.5616695475687,kWh,1.0067948345410623e-06,5.516770416592006e-07,5.331550875406776e-06,6.890022751607039e-06,tokens/kWh,37155174.84180879,MB,1262.83776,13032.226816,0.0,12385.779712,12315.694592,s,10,20.729945068359378,2.0729945068359372,0.0114368267618131,2.07840673828125,2.08414736328125,2.0857937744140624,2.087110903320313,"[2.07699853515625, 2.0828701171875, 2.056541748046875, 2.087440185546875, 2.07981494140625, 2.083781494140625, 2.08099853515625, 2.058927978515625, 2.0584501953125, 2.064121337890625]",tokens/s,30.390818592258817,kWh,2.4479510869178396e-05,1.341476889907303e-05,0.00010454192591219334,0.00014243620568044478,tokens/kWh,442303.2732375666,,s,629,21.461549064636248,0.03412010980069353,0.010155662547651918,0.032658432006835936,0.03377766418457031,0.03395112915039063,0.11812827392578125,"[0.033203201293945314, 0.034149375915527344, 0.03452934265136719, 0.032828353881835935, 0.032546817779541014, 0.032557056427001956, 0.03263180923461914, 0.03250790405273438, 0.03252633666992188, 0.032807937622070314, 0.0330332145690918, 0.03315302276611328, 0.03368243026733399, 0.03264102554321289, 0.03362508773803711, 0.034255870819091795, 0.03418009567260742, 0.03371212768554688, 0.03260006332397461, 0.0331253776550293, 0.03382681655883789, 0.032626686096191404, 0.032666622161865236, 0.032688129425048826, 0.03266969680786133, 0.0326901741027832, 0.03282640075683594, 0.03265840148925781, 0.03264614486694336, 0.032655361175537106, 0.032720897674560545, 0.03260313415527344, 0.0327014389038086, 0.03278131103515625, 0.032696319580078126, 0.03259801483154297, 0.03257651138305664, 0.03260723114013672, 0.032694271087646484, 0.03262156677246094, 0.032671745300292966, 0.033481727600097655, 0.03377664184570312, 0.032835582733154296, 0.0330618896484375, 0.032830463409423825, 0.03261030578613281, 0.03262054443359375, 0.03250688171386719, 0.03251200103759765, 0.032671745300292966, 0.03280998229980469, 0.032522239685058595, 0.03259494400024414, 0.032707584381103515, 0.032589824676513675, 0.03265126419067383, 0.03266969680786133, 0.03259088134765625, 0.033699806213378904, 0.033724414825439454, 0.03396915054321289, 0.1184194564819336, 0.03253247833251953, 0.03244441604614258, 0.03252019119262695, 0.032471038818359374, 0.03262054443359375, 0.03260825729370117, 0.032814079284667966, 0.03323904037475586, 0.03265126419067383, 0.03253760147094727, 0.03262156677246094, 0.032540672302246096, 0.03258879852294922, 0.03300352096557617, 0.03371417617797851, 0.03584000015258789, 0.032955390930175785, 0.032628734588623046, 0.03355756759643555, 0.03384620666503906, 0.03255807876586914, 0.032543743133544925, 0.033807361602783206, 0.033670143127441404, 0.032584705352783204, 0.03274444961547852, 0.03273932647705078, 0.03276083374023438, 0.03268198394775391, 0.03266559982299805, 0.03266873550415039, 0.03273926544189453, 0.03289907073974609, 0.032851966857910156, 0.03264716720581055, 0.03276700973510742, 0.03276489639282226, 0.03374899291992187, 0.033876991271972655, 0.0337520637512207, 0.0336814079284668, 0.033724414825439454, 0.03361996841430664, 0.033620990753173825, 0.03379814529418945, 0.03268505477905274, 0.032691200256347655, 0.03258883285522461, 0.03270038223266602, 0.033140735626220705, 0.032525310516357424, 0.03369676971435547, 0.03400191879272461, 0.034078720092773435, 0.03381862258911133, 0.03359334564208984, 0.03346636962890625, 0.03275372695922851, 0.03251603317260742, 0.032589824676513675, 0.03265945434570312, 0.03260211181640625, 0.11776921844482421, 0.032522239685058595, 0.03256934356689453, 0.03258777618408203, 0.03253452682495117, 0.03318783950805664, 0.03413094329833984, 0.03322060775756836, 0.03261439895629883, 0.03263180923461914, 0.032514049530029294, 0.032570369720458986, 0.032522239685058595, 0.03253657531738281, 0.032645118713378905, 0.032555007934570314, 0.03256934356689453, 0.03254988861083984, 0.03250483322143555, 0.0326379508972168, 0.03249049758911133, 0.032519168853759765, 0.032589824676513675, 0.032570369720458986, 0.032517120361328124, 0.03247411346435547, 0.03256524658203125, 0.03256217575073242, 0.03253145599365234, 0.032543743133544925, 0.032656383514404294, 0.03259084701538086, 0.032519168853759765, 0.03265740966796875, 0.03255094528198242, 0.03251811218261719, 0.03259699249267578, 0.03343974304199219, 0.03258879852294922, 0.03268710327148437, 0.032595966339111326, 0.03256115341186523, 0.03278335952758789, 0.03255094528198242, 0.032594913482666014, 0.032589824676513675, 0.032538623809814454, 0.03255295944213867, 0.032626686096191404, 0.032920574188232424, 0.03264102554321289, 0.0325478401184082, 0.032586753845214846, 0.03259699249267578, 0.03263590240478516, 0.032435199737548825, 0.032484352111816404, 0.03255807876586914, 0.03265433502197266, 0.03256012725830078, 0.032672767639160154, 0.03263283157348633, 0.03263897705078125, 0.11889868927001954, 0.033642494201660156, 0.03368038558959961, 0.03371929550170898, 0.03222937774658203, 0.03249049758911133, 0.03277107238769531, 0.03256115341186523, 0.03263897705078125, 0.03338137435913086, 0.033734657287597655, 0.03258572769165039, 0.03275571060180664, 0.03257244873046875, 0.03265020751953125, 0.03294105529785156, 0.03377766418457031, 0.03377766418457031, 0.033983486175537106, 0.033701889038085936, 0.032527359008789065, 0.03256217575073242, 0.03258879852294922, 0.03273625564575195, 0.03409411239624023, 0.03379299163818359, 0.03375718307495117, 0.0327014389038086, 0.03269123077392578, 0.03282531356811524, 0.03269222259521484, 0.032553985595703126, 0.03262259292602539, 0.03258879852294922, 0.03265945434570312, 0.032758785247802735, 0.03265126419067383, 0.03268000030517578, 0.03261536026000977, 0.0325591049194336, 0.03265433502197266, 0.03259494400024414, 0.03253247833251953, 0.03263897705078125, 0.032612350463867186, 0.032538623809814454, 0.03252838516235351, 0.03260313415527344, 0.032709632873535156, 0.03395276641845703, 0.03375001525878906, 0.03422003173828125, 0.03374182510375977, 0.03375513458251953, 0.03407564926147461, 0.03382374572753906, 0.03377151870727539, 0.03380428695678711, 0.033797119140625, 0.033432575225830076, 0.034348033905029295, 0.033737728118896484, 0.03365683364868164, 0.11833241271972657, 0.033788928985595705, 0.03246387100219727, 0.032723968505859374, 0.0337336311340332, 0.03363430404663086, 0.033723392486572266, 0.03366195297241211, 0.03363123321533203, 0.03377459335327149, 0.033734657287597655, 0.03368755340576172, 0.033754112243652344, 0.034100223541259765, 0.03375001525878906, 0.033896446228027344, 0.033708030700683594, 0.03259904098510742, 0.03335782241821289, 0.03378073501586914, 0.03256115341186523, 0.032578624725341794, 0.03258975982666015, 0.03259084701538086, 0.03269222259521484, 0.03264614486694336, 0.03271478271484375, 0.032535518646240234, 0.032557056427001956, 0.032551937103271485, 0.032557056427001956, 0.03257241439819336, 0.032385025024414066, 0.03226521682739258, 0.03249356842041016, 0.032320510864257815, 0.03223251342773437, 0.032474048614501955, 0.032745471954345705, 0.032470016479492186, 0.03246694564819336, 0.03256217575073242, 0.03325030517578125, 0.03349708938598633, 0.03481292724609375, 0.033919998168945316, 0.03376332855224609, 0.03398553466796875, 0.0327086067199707, 0.03273523330688476, 0.03277414321899414, 0.032645118713378905, 0.032671745300292966, 0.03280588912963867, 0.03271481704711914, 0.03261536026000977, 0.03253049468994141, 0.03266656112670899, 0.03266559982299805, 0.0326379508972168, 0.03260723114013672, 0.03273625564575195, 0.03267379379272461, 0.11821056365966796, 0.03265126419067383, 0.03259801483154297, 0.032601089477539064, 0.03262771224975586, 0.03258879852294922, 0.03260416030883789, 0.032626686096191404, 0.03331686401367188, 0.03261439895629883, 0.032601089477539064, 0.032707584381103515, 0.03359027099609375, 0.0326379508972168, 0.03272499084472656, 0.032551937103271485, 0.032718849182128903, 0.03271372985839844, 0.03332710266113281, 0.03268204879760742, 0.03227027130126953, 0.03339059066772461, 0.033838081359863284, 0.03378585433959961, 0.03394867324829102, 0.03394668960571289, 0.03391788864135742, 0.03379404830932617, 0.03377766418457031, 0.03273011016845703, 0.032715774536132815, 0.03275468826293945, 0.03276595306396484, 0.03273830413818359, 0.032763904571533206, 0.03273011016845703, 0.032712703704833986, 0.032761856079101564, 0.03285913467407227, 0.03279257583618164, 0.03271171188354492, 0.03263792037963867, 0.032699390411376955, 0.032718849182128903, 0.03267583847045898, 0.03392409515380859, 0.033903617858886716, 0.03371929550170898, 0.03268403244018555, 0.032723968505859374, 0.03378585433959961, 0.03275263977050781, 0.03239833450317383, 0.032648193359375, 0.03273113632202149, 0.03268403244018555, 0.03258879852294922, 0.032901153564453126, 0.040007648468017576, 0.032881664276123046, 0.03379916763305664, 0.03299327850341797, 0.032611328125, 0.11826585388183594, 0.032677921295166015, 0.03260208129882813, 0.03321446228027344, 0.03369472122192383, 0.03253657531738281, 0.03257855987548828, 0.03259494400024414, 0.03285094451904297, 0.03274649429321289, 0.03256422424316406, 0.03253763198852539, 0.032717792510986325, 0.032464897155761716, 0.03360255813598633, 0.034044929504394535, 0.03382374572753906, 0.03384729766845703, 0.03271683120727539, 0.03332710266113281, 0.03371209716796875, 0.03338547134399414, 0.03368755340576172, 0.033783809661865234, 0.033751041412353515, 0.03385036849975586, 0.0326297607421875, 0.03293798446655274, 0.033721343994140625, 0.0337448959350586, 0.03371212768554688, 0.03263180923461914, 0.03266969680786133, 0.0327116813659668, 0.032742401123046876, 0.03266048049926758, 0.03256729507446289, 0.03263488006591797, 0.03263590240478516, 0.032658432006835936, 0.032672767639160154, 0.03256115341186523, 0.032643070220947264, 0.03252633666992188, 0.03265027236938477, 0.032568286895751954, 0.033718273162841796, 0.03429580688476563, 0.03369375991821289, 0.033929153442382814, 0.033732608795166014, 0.03369881439208984, 0.03353190231323242, 0.03253657531738281, 0.0327116813659668, 0.032500736236572264, 0.03248230361938476, 0.03244441604614258, 0.032589824676513675, 0.03262156677246094, 0.032658432006835936, 0.03262259292602539, 0.03252121734619141, 0.11833241271972657, 0.03250790405273438, 0.032551937103271485, 0.03275468826293945, 0.032487422943115234, 0.03252019119262695, 0.0326308479309082, 0.03257030487060547, 0.03252633666992188, 0.0326932487487793, 0.032584705352783204, 0.03366502380371094, 0.03373056030273437, 0.03370393753051758, 0.03265228652954102, 0.03262771224975586, 0.03271987152099609, 0.03266559982299805, 0.03253046417236328, 0.032643104553222654, 0.03249452972412109, 0.032541694641113283, 0.03258163070678711, 0.032715774536132815, 0.032661502838134765, 0.03267379379272461, 0.032666622161865236, 0.03265433502197266, 0.03268505477905274, 0.032656383514404294, 0.03260825729370117, 0.03262464141845703, 0.03262054443359375, 0.032612350463867186, 0.032778240203857424, 0.032679935455322266, 0.03264409637451172, 0.03260723114013672, 0.032723968505859374, 0.03260927963256836, 0.03280179214477539, 0.03262259292602539, 0.03259494400024414, 0.03279564666748047, 0.03251200103759765, 0.032573440551757815, 0.032628734588623046, 0.03260518264770508, 0.03260927963256836, 0.03260518264770508, 0.03251200103759765, 0.03250175857543945, 0.032626686096191404, 0.03259392166137695, 0.03267583847045898, 0.03256217575073242, 0.03256732940673828, 0.03249046325683594, 0.03257548904418945, 0.03263283157348633, 0.03271680068969727, 0.03259392166137695, 0.03276595306396484, 0.11832217407226563, 0.032677886962890625, 0.03251200103759765, 0.032573440551757815, 0.032658432006835936, 0.03266764831542969, 0.03248537445068359, 0.032656383514404294, 0.032645118713378905, 0.03252326583862305, 0.03254579162597656, 0.03249356842041016, 0.03256630325317383, 0.0324986572265625, 0.032574462890625, 0.032709632873535156, 0.03260825729370117, 0.03259392166137695, 0.03254886245727539, 0.03253868865966797, 0.032444351196289065, 0.03255807876586914, 0.03250380706787109, 0.03258265686035156, 0.032606208801269534, 0.03258367919921875, 0.032655361175537106, 0.032525310516357424, 0.03259494400024414, 0.032531486511230466, 0.03253654479980469, 0.03264716720581055, 0.03266355133056641, 0.03263897705078125, 0.032656417846679685, 0.03266352081298828, 0.032606208801269534, 0.03256320190429687, 0.032674816131591795, 0.03265126419067383, 0.032720897674560545, 0.032555007934570314, 0.0326379508972168, 0.03251507186889648, 0.03259084701538086, 0.032524288177490236, 0.032648193359375, 0.03284787368774414, 0.032753662109375, 0.03268198394775391, 0.03256012725830078, 0.032557056427001956, 0.03260416030883789, 0.03254886245727539, 0.03351039886474609, 0.03366400146484375, 0.032625663757324216, 0.032525344848632814, 0.03264713668823242, 0.03281203079223633, 0.032817150115966795, 0.03339059066772461, 0.03344179153442383, 0.11791667175292969, 0.032593952178955075, 0.03262255859375, 0.03264716720581055, 0.03250688171386719, 0.03260006332397461, 0.03254579162597656, 0.032704513549804685, 0.032673854827880856, 0.03270137786865234, 0.03256524658203125, 0.0325478401184082, 0.032586753845214846, 0.03259904098510742, 0.03274444961547852, 0.03266969680786133, 0.0326195182800293, 0.03266873550415039, 0.03248121643066406, 0.03337932968139649, 0.03469619369506836, 0.03421798324584961, 0.03388518524169922, 0.032702465057373044, 0.032718849182128903, 0.03277414321899414, 0.03269734573364258, 0.03265433502197266, 0.03261439895629883, 0.032658432006835936, 0.03268198394775391, 0.03277721786499024, 0.03261030578613281, 0.03240140914916992, 0.032710655212402344, 0.032625663757324216, 0.03275059127807617, 0.032648193359375, 0.03260825729370117, 0.03327283096313476, 0.03284275054931641, 0.03254272079467774, 0.03265539169311524, 0.03262255859375, 0.03250995254516602, 0.03253657531738281, 0.032470016479492186, 0.03268915176391601, 0.032435199737548825, 0.03263897705078125, 0.032691200256347655, 0.033667072296142575, 0.032553985595703126, 0.03260211181640625, 0.03303833770751953, 0.03361382293701172, 0.032726016998291016, 0.032704513549804685, 0.03250175857543945, 0.03256934356689453, 0.03243724822998047, 0.03249151992797852, 0.032440319061279296]",tokens/s,29.308229247834188,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,/,/,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 373, in cached_file raise EnvironmentError( OSError: / does not appear to have a file named config.json. Checkout 'https://huggingface.co///tree/None' for available files. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,TencentARC/Mistral_Pro_8B_v0.1,,cuda,0,42,,,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,d35829e539df8480b726c647eeabf91e41eae047,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,,MB,1288.060928,19665.518592,0.0,19019.071488,18747.268096,s,10,0.9966524124145507,0.09966524124145507,0.0003593623538139137,0.0997276496887207,0.09999786834716796,0.1002132209777832,0.10038550308227538,"[0.10042857360839844, 0.09909222412109375, 0.09995001220703124, 0.09973529815673828, 0.09974636840820313, 0.09949440002441406, 0.09933567810058594, 0.09931929779052734, 0.09972000122070312, 0.09983055877685547]",tokens/s,2568.598608814871,kWh,1.1789327095968628e-06,6.459987422384093e-07,7.668251459128825e-06,9.493182910964097e-06,tokens/kWh,26966719.423928328,MB,1288.060928,19665.518592,0.0,19019.071488,18774.921728,s,10,28.885788818359373,2.8885788818359375,0.0010954497532957945,2.8886032714843752,2.889635400390625,2.8903216064453123,2.8908705712890623,"[2.8910078125, 2.88893798828125, 2.887427978515625, 2.888744384765625, 2.88948291015625, 2.8885087890625, 2.888665283203125, 2.888541259765625, 2.886933837890625, 2.88753857421875]",tokens/s,21.810032745222504,kWh,3.4044481393593236e-05,1.8657960823366193e-05,0.00014615584382226746,0.00019885828603922686,tokens/kWh,316808.5235712662,,s,629,29.73557656478883,0.047274366557692876,0.011902277958038209,0.045825023651123044,0.04592947387695313,0.046009739685058594,0.14583502868652343,"[0.0458526725769043, 0.0459233283996582, 0.045870079040527346, 0.04580147171020508, 0.04575948715209961, 0.04586393737792969, 0.045805599212646486, 0.04583318328857422, 0.04580659103393555, 0.046117889404296876, 0.046840831756591796, 0.04631961441040039, 0.045841407775878903, 0.045778942108154294, 0.04576768112182617, 0.04577689743041992, 0.0458076171875, 0.045739009857177736, 0.0457523193359375, 0.04578303909301758, 0.04578815841674805, 0.04580249786376953, 0.0459315185546875, 0.045771839141845704, 0.04593657684326172, 0.04578713607788086, 0.0458158073425293, 0.0457891845703125, 0.04581990432739258, 0.04576870346069336, 0.04578713607788086, 0.045806655883789064, 0.045793216705322264, 0.04587724685668945, 0.04581276702880859, 0.04584854507446289, 0.04579328155517578, 0.045864959716796876, 0.04583628845214844, 0.045886463165283206, 0.04584447860717773, 0.04591513442993164, 0.04583020782470703, 0.0458485107421875, 0.045861888885498046, 0.045795326232910154, 0.045927425384521485, 0.04580966567993164, 0.045868030548095705, 0.04583628845214844, 0.045864959716796876, 0.04578201675415039, 0.045963264465332034, 0.04584960174560547, 0.04605952072143555, 0.04587212753295898, 0.045842494964599606, 0.04585772705078125, 0.045871105194091794, 0.04597760009765625, 0.04619673538208008, 0.04598476791381836, 0.14551756286621093, 0.04574924850463867, 0.04592947387695313, 0.046868480682373044, 0.04589875030517578, 0.04576358413696289, 0.04570214462280273, 0.04584243011474609, 0.04578406524658203, 0.045818878173828126, 0.04574105453491211, 0.04581478500366211, 0.04580147171020508, 0.04574003219604492, 0.045778942108154294, 0.046061569213867185, 0.04575027084350586, 0.04577280044555664, 0.04581273651123047, 0.04580044937133789, 0.045808639526367184, 0.04583833694458008, 0.045886463165283206, 0.045832191467285156, 0.045830142974853515, 0.04576563262939453, 0.04578611373901367, 0.04576768112182617, 0.045810688018798826, 0.045884414672851564, 0.04582297515869141, 0.04584550476074219, 0.04578815841674805, 0.045876224517822264, 0.04578819274902344, 0.045829086303710936, 0.04579225540161133, 0.045825023651123044, 0.04582809448242187, 0.04581273651123047, 0.04580352020263672, 0.04579228973388672, 0.04582806396484375, 0.04582297515869141, 0.045832191467285156, 0.04580147171020508, 0.04584960174560547, 0.0458076171875, 0.04581990432739258, 0.045832191467285156, 0.04581683349609375, 0.045873153686523435, 0.045810688018798826, 0.045818878173828126, 0.04579635238647461, 0.04585574340820313, 0.04592947387695313, 0.04621619033813477, 0.04594483184814453, 0.045876224517822264, 0.04585062408447266, 0.04590694427490234, 0.045841407775878903, 0.14633676147460936, 0.045764606475830076, 0.04578611373901367, 0.04578611373901367, 0.04578303909301758, 0.04574310302734375, 0.045723648071289064, 0.045823040008544924, 0.04583212661743164, 0.04576563262939453, 0.045778942108154294, 0.04583731079101563, 0.04576563262939453, 0.04582195281982422, 0.04576358413696289, 0.04579942321777344, 0.04578303909301758, 0.04577689743041992, 0.04578611373901367, 0.045862911224365234, 0.04579225540161133, 0.04581273651123047, 0.045808639526367184, 0.0458260498046875, 0.045780990600585936, 0.04584755325317383, 0.0457523193359375, 0.045813758850097655, 0.045830177307128905, 0.045817825317382814, 0.04578508758544922, 0.045943809509277345, 0.04578508758544922, 0.04582809448242187, 0.04589158248901367, 0.045841407775878903, 0.04578713607788086, 0.04579840087890625, 0.04585369491577149, 0.045879295349121094, 0.045870079040527346, 0.04581683349609375, 0.0458526725769043, 0.045871105194091794, 0.04587724685668945, 0.045881343841552735, 0.04578303909301758, 0.045884414672851564, 0.0458158073425293, 0.045832191467285156, 0.045830142974853515, 0.045875263214111325, 0.045781951904296875, 0.045848575592041016, 0.04581171035766601, 0.045982719421386715, 0.04584755325317383, 0.04584550476074219, 0.04586598587036133, 0.04585779190063476, 0.04581990432739258, 0.04588236618041992, 0.045878273010253906, 0.14644940185546876, 0.045810688018798826, 0.04577689743041992, 0.045791233062744144, 0.04574515151977539, 0.04576870346069336, 0.04577382278442383, 0.0459048957824707, 0.045798465728759764, 0.04587001419067383, 0.045795326232910154, 0.04576665496826172, 0.045827072143554685, 0.045753345489501954, 0.04589056015014648, 0.04580659103393555, 0.04584550476074219, 0.045795326232910154, 0.045818878173828126, 0.04578815841674805, 0.04581990432739258, 0.04586086273193359, 0.045840385437011716, 0.04580352020263672, 0.045770751953125, 0.0459417610168457, 0.04578713607788086, 0.04589056015014648, 0.04588032150268555, 0.045862911224365234, 0.04574515151977539, 0.04581683349609375, 0.04583935928344727, 0.045861888885498046, 0.045943809509277345, 0.04586393737792969, 0.04581990432739258, 0.045861888885498046, 0.04583935928344727, 0.04583935928344727, 0.045827072143554685, 0.04584447860717773, 0.045808639526367184, 0.045827072143554685, 0.045873153686523435, 0.04581683349609375, 0.04583321762084961, 0.04585884857177734, 0.045957088470458984, 0.04590182495117188, 0.045836318969726564, 0.045878238677978515, 0.04592947387695313, 0.045818878173828126, 0.04586598587036133, 0.04596121597290039, 0.045794303894042966, 0.04586598587036133, 0.045856769561767576, 0.04591107177734375, 0.04583830261230469, 0.045878273010253906, 0.045861888885498046, 0.1462804412841797, 0.045830142974853515, 0.04577177429199219, 0.04601958465576172, 0.045748222351074216, 0.045794303894042966, 0.045736961364746094, 0.045810688018798826, 0.04574720001220703, 0.04576358413696289, 0.04574105453491211, 0.04577280044555664, 0.04581683349609375, 0.045742080688476565, 0.04574924850463867, 0.04574720001220703, 0.04579942321777344, 0.04582297515869141, 0.04597555160522461, 0.04594278335571289, 0.045851646423339845, 0.04588236618041992, 0.04591513442993164, 0.04585062408447266, 0.04577385711669922, 0.0459007682800293, 0.046445568084716796, 0.04599808120727539, 0.04585369491577149, 0.045878273010253906, 0.04575948715209961, 0.04582195281982422, 0.04575743865966797, 0.04583116912841797, 0.045932544708251956, 0.04592639923095703, 0.04576153564453125, 0.04582297515869141, 0.04592230224609375, 0.045864959716796876, 0.045777919769287106, 0.045864959716796876, 0.04586908721923828, 0.04600726318359375, 0.045851646423339845, 0.04591001510620117, 0.045835262298583986, 0.04581785583496094, 0.04586393737792969, 0.045861888885498046, 0.045843456268310545, 0.04579635238647461, 0.045856769561767576, 0.045859840393066405, 0.045846527099609374, 0.045870079040527346, 0.045859840393066405, 0.04588236618041992, 0.045827072143554685, 0.04589567947387695, 0.04581273651123047, 0.045924415588378904, 0.04594169616699219, 0.14593536376953126, 0.045758464813232425, 0.04585062408447266, 0.04576665496826172, 0.045808639526367184, 0.04585062408447266, 0.04576358413696289, 0.04575641632080078, 0.045774848937988284, 0.04582912063598633, 0.04575129699707031, 0.04597350311279297, 0.045932544708251956, 0.046137344360351565, 0.045988895416259765, 0.046106590270996096, 0.04581273651123047, 0.045892608642578124, 0.04577689743041992, 0.0457891845703125, 0.04578406524658203, 0.04579020690917969, 0.04573497772216797, 0.04586387252807617, 0.04575849533081055, 0.04576150512695312, 0.045780990600585936, 0.04581999969482422, 0.04578192138671875, 0.0458342399597168, 0.045758464813232425, 0.045818878173828126, 0.045810688018798826, 0.04588748931884765, 0.045823040008544924, 0.0459222412109375, 0.04577177429199219, 0.04586905670166016, 0.04582092666625977, 0.045859840393066405, 0.045774848937988284, 0.04580556869506836, 0.04575337600708008, 0.0459007682800293, 0.04578406524658203, 0.0458076171875, 0.04578713607788086, 0.04579942321777344, 0.04582912063598633, 0.045927425384521485, 0.045879295349121094, 0.04584755325317383, 0.045824001312255856, 0.04580352020263672, 0.04590694427490234, 0.04589670562744141, 0.045862911224365234, 0.045873153686523435, 0.045894657135009766, 0.045835262298583986, 0.045849632263183594, 0.046041057586669924, 0.04583116912841797, 0.14579820251464845, 0.04584748840332031, 0.0457564811706543, 0.04577990341186523, 0.04573497772216797, 0.04581267166137695, 0.04574105453491211, 0.045764606475830076, 0.045748222351074216, 0.04573699188232422, 0.04574614334106445, 0.04577996826171875, 0.04576358413696289, 0.04582912063598633, 0.04576972961425781, 0.04579225540161133, 0.04581683349609375, 0.045846527099609374, 0.04579942321777344, 0.04583628845214844, 0.0458158073425293, 0.04591001510620117, 0.04629094314575195, 0.045878273010253906, 0.04581683349609375, 0.045797374725341795, 0.04576972961425781, 0.04584550476074219, 0.04579225540161133, 0.04580147171020508, 0.04592537689208984, 0.04579840087890625, 0.04574310302734375, 0.045808639526367184, 0.04579635238647461, 0.045868030548095705, 0.04595097732543945, 0.045862911224365234, 0.04591308975219727, 0.04589977645874024, 0.045825023651123044, 0.045810688018798826, 0.045827072143554685, 0.04587212753295898, 0.04591308975219727, 0.04581990432739258, 0.04592025756835937, 0.045879295349121094, 0.04582195281982422, 0.04581785583496094, 0.045832191467285156, 0.04581478500366211, 0.045797374725341795, 0.04586086273193359, 0.04595199966430664, 0.046058494567871096, 0.045911041259765625, 0.045900798797607424, 0.04587519836425781, 0.045859840393066405, 0.045848575592041016, 0.045859840393066405, 0.04597350311279297, 0.14584934997558593, 0.04575539016723633, 0.04584550476074219, 0.04573388671875, 0.04578611373901367, 0.045830142974853515, 0.04573491287231445, 0.045744129180908207, 0.0457154541015625, 0.045856769561767576, 0.0457154541015625, 0.04578201675415039, 0.04576153564453125, 0.045851646423339845, 0.045742080688476565, 0.04577177429199219, 0.045846527099609374, 0.045720577239990234, 0.04580659103393555, 0.04575027084350586, 0.04579635238647461, 0.045742080688476565, 0.04577280044555664, 0.045900798797607424, 0.04581171035766601, 0.04576972961425781, 0.04580044937133789, 0.04581171035766601, 0.04577382278442383, 0.04580249786376953, 0.04576563262939453, 0.04589567947387695, 0.04576358413696289, 0.04580044937133789, 0.04583116912841797, 0.04589363098144531, 0.045840385437011716, 0.04577996826171875, 0.045810688018798826, 0.045873153686523435, 0.04601446533203125, 0.046198814392089844, 0.04603596878051758, 0.045823966979980466, 0.04578508758544922, 0.0458076171875, 0.04578713607788086, 0.045813758850097655, 0.0459417610168457, 0.04611174392700195, 0.046132225036621094, 0.045916160583496096, 0.04587833786010742, 0.045926334381103516, 0.045856769561767576, 0.04585472106933594, 0.04584755325317383, 0.04591001510620117, 0.045859840393066405, 0.045879295349121094, 0.045848575592041016, 0.046020606994628906, 0.045846527099609374, 0.14615858459472655, 0.04580556869506836, 0.04575641632080078, 0.04570521545410156, 0.045794303894042966, 0.04577177429199219, 0.04575436782836914, 0.045742080688476565, 0.045753345489501954, 0.04578303909301758, 0.04576051330566406, 0.045797374725341795, 0.04578611373901367, 0.045835262298583986, 0.04572979354858398, 0.045825023651123044, 0.0458076171875, 0.04582406234741211, 0.045811649322509765, 0.04576972961425781, 0.04581478500366211, 0.04573593521118164, 0.04576358413696289, 0.045739009857177736, 0.04574617767333984, 0.0457749137878418, 0.04577171325683594, 0.045736961364746094, 0.04577280044555664, 0.04581171035766601, 0.0457523193359375, 0.04582912063598633, 0.045770751953125, 0.045825023651123044, 0.04580044937133789, 0.045841407775878903, 0.04589567947387695, 0.046066688537597655, 0.04601139068603516, 0.04586905670166016, 0.04578201675415039, 0.045797374725341795, 0.04579942321777344, 0.04585472106933594, 0.04590182495117188, 0.04581171035766601, 0.04582092666625977, 0.045810688018798826, 0.04580147171020508, 0.045841407775878903, 0.045825023651123044, 0.04583628845214844, 0.04580044937133789, 0.04584550476074219, 0.04583321762084961, 0.045861888885498046, 0.045871105194091794, 0.045982719421386715, 0.04586700820922852, 0.04588851165771484, 0.04584447860717773, 0.04587724685668945, 0.045841407775878903, 0.14620262145996094, 0.04572774505615235, 0.045754398345947266, 0.045727710723876956, 0.045753345489501954, 0.04582297515869141, 0.04572671890258789, 0.045795326232910154, 0.04575129699707031, 0.04578406524658203, 0.04577689743041992, 0.045797374725341795, 0.04576563262939453, 0.045774848937988284, 0.04583833694458008, 0.04572774505615235, 0.0457625617980957, 0.04573798370361328, 0.04577177429199219, 0.04580556869506836, 0.0457523193359375, 0.04581273651123047, 0.0458158073425293, 0.04590796661376953, 0.04577593612670899, 0.04578911972045899, 0.04577382278442383, 0.045864959716796876, 0.045832191467285156, 0.045774848937988284, 0.045794303894042966, 0.045848575592041016, 0.045808639526367184, 0.045873153686523435, 0.04583116912841797, 0.04592230224609375, 0.04586393737792969, 0.04580556869506836, 0.04583321762084961, 0.04583935928344727, 0.04579635238647461, 0.045808639526367184, 0.045982719421386715, 0.045861888885498046, 0.04592537689208984, 0.04600320053100586, 0.04585369491577149, 0.045825023651123044, 0.04575743865966797, 0.04589670562744141, 0.045876224517822264, 0.04588544082641602, 0.045813758850097655, 0.04588544082641602, 0.045879295349121094, 0.04588032150268555, 0.045835262298583986, 0.04595814514160156, 0.04593664169311523, 0.045843456268310545, 0.04585369491577149, 0.045902847290039066, 0.045825023651123044]",tokens/s,21.153112623510587,,,main,False,False,,,,,,,,,,,,,,,,,,,,,,, float16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-6.7b,EleutherAI/pythia-6.7b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.42.1,,0.31.0,,,,1.20.0,,,,0.11.1,,,,MB,1286.8608,15104.212992,0.0,14457.765888,14237.630976,s,10,0.8225084838867188,0.08225084838867189,0.00014683798844551083,0.0822090721130371,0.08233436660766602,0.08249967994689943,0.08263193061828614,"[0.08266499328613282, 0.08226649475097657, 0.08220537567138672, 0.0822138900756836, 0.0822976303100586, 0.08220201873779297, 0.08212246704101563, 0.0822127685546875, 0.08213043212890625, 0.08219241333007812]",tokens/s,3112.4299021243646,kWh,9.707098015887709e-07,5.318561317990321e-07,5.26513627039345e-06,6.7677022037812524e-06,tokens/kWh,37826723.501067705,MB,1286.8608,15104.212992,0.0,14457.765888,14415.238144,s,10,21.22832958984375,2.122832958984375,0.002310017020319638,2.1222681884765624,2.1238395263671874,2.1266297241210936,2.1288618823242187,"[2.129419921875, 2.123219482421875, 2.122244140625, 2.122619873046875, 2.122738525390625, 2.121811279296875, 2.120616943359375, 2.122292236328125, 2.122200439453125, 2.121166748046875]",tokens/s,29.67732328319466,kWh,2.5057277665329325e-05,1.3732193833213321e-05,0.00011736137849740645,0.00015615084999594906,tokens/kWh,403456.01706064603,,s,629,21.9342674293518,0.034871649331242933,0.009778497120117672,0.03368956756591797,0.03381063690185546,0.03388743743896484,0.1159414144897461,"[0.03401318359375, 0.03401113510131836, 0.03406131362915039, 0.034067489624023437, 0.034218975067138675, 0.03408281707763672, 0.034116607666015625, 0.034100223541259765, 0.03417497634887695, 0.03400191879272461, 0.03394047927856445, 0.033939456939697264, 0.03370598220825195, 0.033667072296142575, 0.03364352035522461, 0.03358003234863281, 0.03363225555419922, 0.03360768127441406, 0.03359641647338867, 0.033737728118896484, 0.0336363525390625, 0.03363942337036133, 0.03370086288452148, 0.03364352035522461, 0.03369472122192383, 0.03368243026733399, 0.033691646575927735, 0.03384012985229492, 0.03362815856933594, 0.033672191619873046, 0.03376230239868164, 0.03368038558959961, 0.03365683364868164, 0.03369472122192383, 0.033675262451171875, 0.03366809463500976, 0.03373158264160156, 0.03367935943603516, 0.03369881439208984, 0.033827838897705076, 0.033772544860839845, 0.03369881439208984, 0.033704959869384765, 0.03370700836181641, 0.03402649688720703, 0.03376332855224609, 0.03374694442749023, 0.03375513458251953, 0.03376025772094727, 0.03376332855224609, 0.03379507064819336, 0.03375820922851563, 0.03390566253662109, 0.0337336311340332, 0.03377151870727539, 0.03380121612548828, 0.033783809661865234, 0.033775615692138675, 0.033808383941650394, 0.03389440155029297, 0.03379609680175781, 0.03386675262451172, 0.11630899047851563, 0.03387187194824219, 0.03362918472290039, 0.033532928466796875, 0.03369676971435547, 0.03366502380371094, 0.03359436798095703, 0.033598464965820314, 0.03361075210571289, 0.03358003234863281, 0.033582080841064454, 0.03361587142944336, 0.03361075210571289, 0.03360665512084961, 0.033770496368408204, 0.03361587142944336, 0.03363020706176758, 0.03373056030273437, 0.03360870361328125, 0.033582080841064454, 0.03360153579711914, 0.03362611389160156, 0.03363020706176758, 0.03362201690673828, 0.03370905685424805, 0.03364147186279297, 0.03364761734008789, 0.0336814079284668, 0.03368038558959961, 0.03368960189819336, 0.03367935943603516, 0.03367731094360352, 0.03367935943603516, 0.0336732177734375, 0.03369881439208984, 0.03370905685424805, 0.033734657287597655, 0.03368038558959961, 0.033691646575927735, 0.03367424011230469, 0.03371110534667969, 0.033670143127441404, 0.03372851181030274, 0.033791999816894534, 0.03371724700927734, 0.033739776611328126, 0.033753089904785157, 0.033710079193115236, 0.033751041412353515, 0.033983486175537106, 0.03376128005981445, 0.03370086288452148, 0.03378278350830078, 0.03375718307495117, 0.033811454772949216, 0.03372851181030274, 0.03376025772094727, 0.033753089904785157, 0.03381452941894531, 0.03379404830932617, 0.03377971267700195, 0.03380326461791992, 0.033800193786621094, 0.11612774658203125, 0.03352371215820313, 0.03353190231323242, 0.03349708938598633, 0.03355648040771484, 0.03352883148193359, 0.033585151672363284, 0.033549312591552735, 0.033568767547607424, 0.03363020706176758, 0.033598464965820314, 0.03358924865722656, 0.03355648040771484, 0.033604606628417966, 0.03363430404663086, 0.03359436798095703, 0.03359027099609375, 0.03360563278198242, 0.03357183837890625, 0.033667072296142575, 0.033620990753173825, 0.03363020706176758, 0.0336097297668457, 0.03368447875976562, 0.03366092681884766, 0.033723392486572266, 0.03365273666381836, 0.033683456420898435, 0.03363123321533203, 0.033704959869384765, 0.0336517105102539, 0.03370086288452148, 0.033650688171386715, 0.03377459335327149, 0.03363430404663086, 0.03369676971435547, 0.033685504913330076, 0.03370086288452148, 0.033701889038085936, 0.033718273162841796, 0.03366912078857422, 0.033683456420898435, 0.033791999816894534, 0.033734657287597655, 0.03367935943603516, 0.033718273162841796, 0.033713153839111325, 0.03386368179321289, 0.03372748947143555, 0.033767425537109375, 0.033716224670410154, 0.03379814529418945, 0.03375718307495117, 0.03378483200073242, 0.03401113510131836, 0.03378176116943359, 0.033721343994140625, 0.03383295822143555, 0.033870849609375, 0.03379609680175781, 0.03385548782348633, 0.03386982345581055, 0.03376537704467773, 0.1160099868774414, 0.03354009628295898, 0.03360153579711914, 0.033546241760253906, 0.033547264099121094, 0.03359231948852539, 0.03361894226074219, 0.033530879974365234, 0.033585151672363284, 0.03355852890014648, 0.03355955123901367, 0.03359641647338867, 0.033587200164794925, 0.03359436798095703, 0.03364863967895508, 0.03359436798095703, 0.0336732177734375, 0.03357491302490234, 0.03364556884765625, 0.03364352035522461, 0.03360870361328125, 0.03366809463500976, 0.03363430404663086, 0.03361996841430664, 0.03362815856933594, 0.0336363525390625, 0.033753089904785157, 0.033650688171386715, 0.03376230239868164, 0.03364147186279297, 0.03363225555419922, 0.03381043243408203, 0.03365990447998047, 0.03370700836181641, 0.03364863967895508, 0.033683456420898435, 0.03375820922851563, 0.03368960189819336, 0.033686527252197264, 0.033708030700683594, 0.03390259170532227, 0.03371212768554688, 0.03366400146484375, 0.033726463317871096, 0.03367833709716797, 0.033745918273925785, 0.033783809661865234, 0.03379507064819336, 0.033740798950195314, 0.03375513458251953, 0.033702911376953124, 0.03375718307495117, 0.03388313674926758, 0.033898494720458985, 0.03376435089111328, 0.03378483200073242, 0.03376947021484375, 0.03382271957397461, 0.03383302307128906, 0.03377657699584961, 0.03376435089111328, 0.03384012985229492, 0.033890304565429685, 0.11609088134765624, 0.03351244735717773, 0.03356364822387695, 0.03356364822387695, 0.03352268981933594, 0.03360255813598633, 0.033582080841064454, 0.03360358428955078, 0.033584129333496096, 0.03357593536376953, 0.033600513458251956, 0.033683456420898435, 0.033588222503662106, 0.03363225555419922, 0.03359641647338867, 0.03361382293701172, 0.033624095916748045, 0.03361891174316406, 0.03364556884765625, 0.03362611389160156, 0.03359027099609375, 0.03358310317993164, 0.03362815856933594, 0.0336097297668457, 0.03364352035522461, 0.03363840103149414, 0.03362406539916992, 0.03365785598754883, 0.03367731094360352, 0.033704959869384765, 0.03378790283203125, 0.0336732177734375, 0.03367427062988281, 0.03374691009521484, 0.033805313110351565, 0.03367628860473633, 0.03367935943603516, 0.03380428695678711, 0.0336814079284668, 0.0337336311340332, 0.03374182510375977, 0.03373158264160156, 0.03380223846435547, 0.03371212768554688, 0.03384627151489258, 0.03371929550170898, 0.033844223022460936, 0.03375820922851563, 0.03379507064819336, 0.033770496368408204, 0.03382681655883789, 0.03375718307495117, 0.0337367057800293, 0.033783809661865234, 0.0338974723815918, 0.03377356719970703, 0.03375513458251953, 0.03379507064819336, 0.033740798950195314, 0.03376844787597656, 0.03381760025024414, 0.03381862258911133, 0.03386470413208008, 0.115957763671875, 0.03366502380371094, 0.03350425720214844, 0.03351244735717773, 0.03354009628295898, 0.03357900619506836, 0.03352883148193359, 0.03353497695922852, 0.03355340957641602, 0.033546241760253906, 0.033582080841064454, 0.03356671905517578, 0.033570816040039066, 0.033604606628417966, 0.03377766418457031, 0.03373158264160156, 0.03359436798095703, 0.03357388687133789, 0.033585151672363284, 0.03362815856933594, 0.033633281707763675, 0.03359231948852539, 0.03358617782592774, 0.03361996841430664, 0.033616897583007815, 0.03362918472290039, 0.03364044952392578, 0.033642494201660156, 0.03364659118652344, 0.03364352035522461, 0.03367935943603516, 0.03366400146484375, 0.033691646575927735, 0.03366604614257813, 0.033767425537109375, 0.033838081359863284, 0.03395891189575195, 0.03371212768554688, 0.03369779205322266, 0.03373158264160156, 0.03384832000732422, 0.033734657287597655, 0.033844223022460936, 0.03370086288452148, 0.033724414825439454, 0.03373875045776367, 0.033718273162841796, 0.03372236633300781, 0.03371724700927734, 0.03378688049316406, 0.03374182510375977, 0.03369881439208984, 0.03373158264160156, 0.03374899291992187, 0.033745918273925785, 0.033783809661865234, 0.03372748947143555, 0.0337336311340332, 0.0337336311340332, 0.03374694442749023, 0.033753089904785157, 0.03376230239868164, 0.033808383941650394, 0.11587891387939453, 0.03348889541625977, 0.03351244735717773, 0.033546241760253906, 0.033538047790527346, 0.033546241760253906, 0.03354111862182617, 0.03354009628295898, 0.03355955123901367, 0.033546241760253906, 0.03358617782592774, 0.03355648040771484, 0.03355340957641602, 0.03360153579711914, 0.03356364822387695, 0.03359334564208984, 0.03356467056274414, 0.033617919921875, 0.03358310317993164, 0.03359539031982422, 0.03357593536376953, 0.03362815856933594, 0.03359027099609375, 0.03362406539916992, 0.03360870361328125, 0.03360870361328125, 0.03362406539916992, 0.03366502380371094, 0.03380223846435547, 0.033686527252197264, 0.03366912078857422, 0.03378278350830078, 0.03367731094360352, 0.03363942337036133, 0.03377356719970703, 0.03368755340576172, 0.03364044952392578, 0.03369062423706055, 0.033655807495117186, 0.033701889038085936, 0.03367119979858398, 0.0336844482421875, 0.033675262451171875, 0.03370905685424805, 0.03368755340576172, 0.03370700836181641, 0.03370086288452148, 0.03369779205322266, 0.0337336311340332, 0.03379302215576172, 0.03377459335327149, 0.03376128005981445, 0.033710079193115236, 0.03374899291992187, 0.03373875045776367, 0.03376230239868164, 0.03375820922851563, 0.0337520637512207, 0.03378073501586914, 0.03380223846435547, 0.033759231567382815, 0.03376435089111328, 0.03374899291992187, 0.11599565124511718, 0.03353497695922852, 0.0335175666809082, 0.03361996841430664, 0.03352883148193359, 0.033570816040039066, 0.03361177444458008, 0.03357388687133789, 0.033585151672363284, 0.03356671905517578, 0.03352268981933594, 0.03357798385620117, 0.03367833709716797, 0.033585151672363284, 0.03358924865722656, 0.033797119140625, 0.03359231948852539, 0.033562625885009766, 0.033614849090576174, 0.03358003234863281, 0.03359743881225586, 0.033600513458251956, 0.033805313110351565, 0.033675262451171875, 0.03360358428955078, 0.033658878326416015, 0.03367628860473633, 0.03368960189819336, 0.03386470413208008, 0.03367935943603516, 0.03373567962646484, 0.03364863967895508, 0.033683456420898435, 0.033672191619873046, 0.033753089904785157, 0.033693695068359376, 0.03369676971435547, 0.03370393753051758, 0.033701889038085936, 0.03369574356079102, 0.03371724700927734, 0.03370086288452148, 0.03372851181030274, 0.03371110534667969, 0.033737728118896484, 0.033716224670410154, 0.03376947021484375, 0.033718273162841796, 0.03377459335327149, 0.03381350326538086, 0.03374694442749023, 0.033791999816894534, 0.033742847442626955, 0.03377151870727539, 0.033756160736083986, 0.03374182510375977, 0.033753089904785157, 0.03378073501586914, 0.033808383941650394, 0.033783809661865234, 0.03386368179321289, 0.033783809661865234, 0.03379609680175781, 0.1159362564086914, 0.0335022087097168, 0.03358924865722656, 0.033565696716308595, 0.033514495849609374, 0.033582080841064454, 0.033568767547607424, 0.03361075210571289, 0.0335728645324707, 0.03359436798095703, 0.033562625885009766, 0.03357593536376953, 0.03357388687133789, 0.03360563278198242, 0.03358924865722656, 0.03359334564208984, 0.033576961517333984, 0.03366809463500976, 0.0336445426940918, 0.03359231948852539, 0.03362918472290039, 0.03358003234863281, 0.03369574356079102, 0.03365273666381836, 0.03363942337036133, 0.033650688171386715, 0.033661983489990235, 0.03367420959472656, 0.0336363525390625, 0.033675262451171875, 0.033670143127441404, 0.03364761734008789, 0.034233345031738284, 0.03369779205322266, 0.03366092681884766, 0.03367116928100586, 0.03367424011230469, 0.033732608795166014, 0.03366809463500976, 0.03372236633300781, 0.03373567962646484, 0.03369574356079102, 0.033685504913330076, 0.033716224670410154, 0.033740798950195314, 0.03370086288452148, 0.033739776611328126, 0.03375001525878906, 0.03376639938354492, 0.033734657287597655, 0.033772544860839845, 0.03379404830932617, 0.03376025772094727, 0.03377151870727539, 0.03376537704467773, 0.033788928985595705, 0.03374694442749023, 0.03380121612548828, 0.03377356719970703, 0.03379302215576172, 0.033827838897705076, 0.03378995132446289, 0.03381964874267578, 0.11594342041015625, 0.03358620834350586, 0.033544158935546874, 0.03350630569458008, 0.033544193267822264, 0.03362406539916992, 0.03353395080566406, 0.03360358428955078, 0.03355340957641602, 0.03357593536376953, 0.033598464965820314, 0.03358617782592774, 0.03361894226074219, 0.03357798385620117, 0.033567745208740236, 0.03364659118652344, 0.033617919921875, 0.0335994873046875, 0.03358924865722656, 0.03361996841430664, 0.03362406539916992, 0.03361075210571289, 0.0336097297668457, 0.033642494201660156, 0.03360665512084961, 0.03364761734008789, 0.03361177444458008, 0.033685504913330076, 0.03366400146484375, 0.03364659118652344, 0.03367116928100586, 0.033701889038085936, 0.03366092681884766, 0.03364863967895508, 0.033667102813720706, 0.03368956756591797, 0.03367424011230469, 0.03369574356079102, 0.033672191619873046, 0.03370905685424805, 0.033710079193115236, 0.0337367057800293, 0.033702911376953124, 0.033710079193115236, 0.03370393753051758, 0.0337151985168457, 0.03371417617797851, 0.033739776611328126, 0.03369267272949219, 0.033759231567382815, 0.03384524917602539, 0.03381350326538086, 0.03373875045776367, 0.033737728118896484, 0.0337448959350586, 0.033775615692138675, 0.0337336311340332, 0.03375001525878906, 0.03376435089111328, 0.03379609680175781, 0.03378073501586914, 0.03378073501586914, 0.03377459335327149]",tokens/s,28.67659027254725,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-13b,facebook/opt-13b,cuda,0,42,,,True,,,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",opt,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-6.7b,facebook/opt-6.7b,cuda,0,42,,,True,,,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",opt,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,microsoft/rho-math-1b-v0.1,microsoft/rho-math-1b-v0.1,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1270.849536,3041.394688,0.0,2394.947584,2279.417856,s,10,0.24058313369750975,0.024058313369750974,0.000541909354183334,0.024157487869262695,0.024371293067932127,0.024764110088348387,0.025078363704681396,"[0.02515692710876465, 0.02312175941467285, 0.02419491195678711, 0.024222591400146484, 0.024074432373046874, 0.024284000396728515, 0.024172800064086914, 0.02320191955566406, 0.02401161575317383, 0.024142175674438476]",tokens/s,10640.812432091529,kWh,2.8420888229758824e-07,1.5573286271826592e-07,1.1061206765625098e-06,1.546062421578364e-06,tokens/kWh,165581930.21640837,MB,1270.849536,3041.394688,0.0,2394.947584,2279.420416,s,10,14.563057861328126,1.4563057861328124,0.006222302227402314,1.4575923461914062,1.4609573974609376,1.4624493408203125,1.4636428955078125,"[1.459085205078125, 1.45394970703125, 1.4639412841796875, 1.4589393310546874, 1.457260009765625, 1.4557384033203125, 1.4579246826171874, 1.439390380859375, 1.4562030029296875, 1.4606258544921875]",tokens/s,43.260145362256026,kWh,1.7034398785784336e-05,9.334815048509355e-06,3.722665998963787e-05,6.359587382393158e-05,tokens/kWh,990630.3068406405,,s,629,14.753925128936764,0.023456160777323955,0.002903868648751255,0.02310655975341797,0.023323033905029298,0.023503257369995117,0.04722995300292972,"[0.023405567169189453, 0.02228428840637207, 0.022031360626220704, 0.022098943710327147, 0.022079488754272462, 0.02208460807800293, 0.02220953559875488, 0.023191551208496093, 0.023191551208496093, 0.02304204750061035, 0.023015424728393553, 0.02310041618347168, 0.022992895126342772, 0.023149568557739256, 0.02305433654785156, 0.023166976928710937, 0.023104511260986327, 0.023161855697631836, 0.023040000915527343, 0.02300214385986328, 0.02331644821166992, 0.02328166389465332, 0.023373823165893554, 0.023451711654663084, 0.023891904830932616, 0.023974912643432617, 0.023771135330200196, 0.023152639389038086, 0.02307891273498535, 0.023611391067504883, 0.023446527481079102, 0.023044095993041993, 0.02313523292541504, 0.023104511260986327, 0.023196672439575194, 0.023022592544555662, 0.02327961540222168, 0.023166976928710937, 0.023228416442871092, 0.02304310417175293, 0.023060447692871095, 0.0230645751953125, 0.023128063201904296, 0.02294272041320801, 0.02314854431152344, 0.023028736114501954, 0.02309529685974121, 0.023006208419799806, 0.02309939193725586, 0.023005184173583985, 0.023079935073852538, 0.022964223861694336, 0.02327347183227539, 0.023152639389038086, 0.02309225654602051, 0.024512544631958007, 0.024862655639648436, 0.023362560272216795, 0.023192575454711914, 0.023189504623413085, 0.022988800048828126, 0.022981632232666017, 0.0464005126953125, 0.02307891273498535, 0.02307379150390625, 0.023059455871582032, 0.022936576843261718, 0.022932479858398438, 0.022964223861694336, 0.023112703323364257, 0.022986751556396484, 0.023029760360717775, 0.02313216018676758, 0.022985727310180663, 0.02309734344482422, 0.023134208679199218, 0.023010303497314453, 0.023014400482177736, 0.023133184432983397, 0.0230645751953125, 0.02309427261352539, 0.023197696685791015, 0.023037952423095705, 0.02306355285644531, 0.023129087448120117, 0.022966272354125978, 0.023053312301635744, 0.02304921531677246, 0.02307379150390625, 0.022953983306884765, 0.022965248107910157, 0.023006208419799806, 0.02301644706726074, 0.022978559494018554, 0.02307174491882324, 0.023145471572875977, 0.023088127136230468, 0.023111679077148437, 0.023202816009521485, 0.023205888748168944, 0.02309939193725586, 0.023146495819091797, 0.02309529685974121, 0.02306764793395996, 0.0230645751953125, 0.023093248367309572, 0.023015424728393553, 0.023014400482177736, 0.023211008071899415, 0.023060480117797853, 0.023020544052124024, 0.02306252861022949, 0.023128063201904296, 0.023121919631958008, 0.023009279251098632, 0.023133184432983397, 0.023234560012817384, 0.02311680030822754, 0.02307481575012207, 0.02310553550720215, 0.023038976669311522, 0.023169023513793945, 0.02295091247558594, 0.0231147518157959, 0.023378944396972655, 0.04778598403930664, 0.023315456390380858, 0.023211008071899415, 0.023187456130981447, 0.023003135681152344, 0.02302566337585449, 0.023045120239257814, 0.023137279510498047, 0.023198720932006835, 0.023162879943847657, 0.023112703323364257, 0.02310041618347168, 0.02309836769104004, 0.023113727569580078, 0.023134208679199218, 0.023104511260986327, 0.022922239303588866, 0.023538688659667968, 0.023500799179077148, 0.022957056045532227, 0.02303385543823242, 0.023215103149414062, 0.02331648063659668, 0.023388160705566406, 0.02307792091369629, 0.023046112060546874, 0.02306255912780762, 0.023044063568115235, 0.023022592544555662, 0.02330419158935547, 0.023153663635253906, 0.02329804801940918, 0.02309427261352539, 0.023376895904541017, 0.023163904190063478, 0.02386636734008789, 0.02477568054199219, 0.02368819236755371, 0.02351513671875, 0.02327347183227539, 0.02367692756652832, 0.023191551208496093, 0.023183359146118163, 0.0230328311920166, 0.02329599952697754, 0.023157760620117186, 0.02332876777648926, 0.02324787139892578, 0.02312294387817383, 0.023319551467895508, 0.023126079559326173, 0.023301055908203125, 0.023113727569580078, 0.02309427261352539, 0.023243776321411135, 0.023192575454711914, 0.02329497528076172, 0.023362560272216795, 0.023161855697631836, 0.02304102325439453, 0.023163904190063478, 0.023006208419799806, 0.02313523292541504, 0.048102401733398435, 0.02312499237060547, 0.023173120498657225, 0.023206911087036132, 0.023015424728393553, 0.023002111434936523, 0.02309017562866211, 0.023076864242553712, 0.023197696685791015, 0.023068672180175782, 0.023030784606933592, 0.023178272247314453, 0.023213024139404296, 0.023069696426391603, 0.023129087448120117, 0.023387136459350585, 0.023047168731689452, 0.023225343704223633, 0.023137279510498047, 0.022795263290405272, 0.02310655975341797, 0.023173120498657225, 0.023152639389038086, 0.023052352905273438, 0.02300511932373047, 0.023137279510498047, 0.022951936721801756, 0.023169023513793945, 0.02304921531677246, 0.023113727569580078, 0.02309734344482422, 0.02309836769104004, 0.023068672180175782, 0.023004159927368165, 0.023137279510498047, 0.02309017562866211, 0.023243776321411135, 0.023401472091674806, 0.023205888748168944, 0.023128063201904296, 0.023590911865234376, 0.024308736801147462, 0.02345779228210449, 0.02309939193725586, 0.023096319198608398, 0.023204864501953124, 0.023027711868286133, 0.02315673637390137, 0.023014400482177736, 0.023053312301635744, 0.02310758399963379, 0.023085056304931642, 0.02308608055114746, 0.02304921531677246, 0.023179264068603517, 0.02332262420654297, 0.023142400741577147, 0.02308095932006836, 0.02311577606201172, 0.02311884880065918, 0.023053312301635744, 0.023060480117797853, 0.023150592803955077, 0.04772147369384765, 0.023504896163940428, 0.023121919631958008, 0.02312294387817383, 0.02308095932006836, 0.023126047134399415, 0.023094240188598632, 0.02311577606201172, 0.022972415924072266, 0.02314035224914551, 0.023085056304931642, 0.023187456130981447, 0.023011327743530274, 0.023170047760009766, 0.023138303756713868, 0.023216127395629883, 0.0231147518157959, 0.023013376235961915, 0.023045120239257814, 0.023060480117797853, 0.023187456130981447, 0.02302566337585449, 0.022943744659423827, 0.023356416702270507, 0.023152639389038086, 0.023045120239257814, 0.023197696685791015, 0.02308198356628418, 0.023157760620117186, 0.02326118469238281, 0.02308095932006836, 0.023351295471191406, 0.023059455871582032, 0.02331340789794922, 0.023119871139526366, 0.023155712127685548, 0.023178239822387696, 0.023164928436279295, 0.023225343704223633, 0.02295910453796387, 0.022973440170288087, 0.023171072006225587, 0.02326016044616699, 0.023120895385742187, 0.02306662368774414, 0.02307788848876953, 0.023054399490356446, 0.023076799392700194, 0.023111679077148437, 0.02312396812438965, 0.02307174491882324, 0.023117824554443358, 0.02304614448547363, 0.02312499237060547, 0.02307174491882324, 0.023137279510498047, 0.02310553550720215, 0.023040000915527343, 0.02301644706726074, 0.02307481575012207, 0.023023616790771483, 0.02310348892211914, 0.0233175048828125, 0.04755251312255859, 0.022854656219482423, 0.022975488662719725, 0.023353343963623048, 0.023152639389038086, 0.02314854431152344, 0.02305843162536621, 0.02309836769104004, 0.023037952423095705, 0.023147520065307618, 0.02307891273498535, 0.0230830078125, 0.02305536079406738, 0.023223360061645507, 0.023193599700927735, 0.02304300880432129, 0.02308608055114746, 0.023192575454711914, 0.02326425552368164, 0.022939647674560547, 0.02311680030822754, 0.02394419288635254, 0.023348224639892577, 0.02305740737915039, 0.023009279251098632, 0.023167999267578124, 0.023144447326660156, 0.022809600830078124, 0.022800384521484376, 0.023015424728393553, 0.02313523292541504, 0.023006208419799806, 0.0230645751953125, 0.023166976928710937, 0.023175167083740233, 0.02311577606201172, 0.023161855697631836, 0.02304921531677246, 0.023134208679199218, 0.023093248367309572, 0.02309734344482422, 0.023030784606933592, 0.023034912109375, 0.022987743377685547, 0.023060480117797853, 0.02312396812438965, 0.023242752075195314, 0.023349248886108398, 0.023112703323364257, 0.023111679077148437, 0.023141376495361327, 0.023121919631958008, 0.023111679077148437, 0.023208959579467774, 0.02305023956298828, 0.023026687622070312, 0.023161855697631836, 0.022963199615478515, 0.022746112823486327, 0.023113759994506836, 0.023089120864868164, 0.02312396812438965, 0.023011327743530274, 0.04798668670654297, 0.023198720932006835, 0.023079935073852538, 0.023096319198608398, 0.023208959579467774, 0.02288128089904785, 0.02272051239013672, 0.023003135681152344, 0.022995967864990235, 0.02310860824584961, 0.02305023956298828, 0.022961151123046874, 0.023149568557739256, 0.023373823165893554, 0.023361536026000978, 0.02596659278869629, 0.023402496337890624, 0.023341056823730468, 0.02310553550720215, 0.023129087448120117, 0.023195648193359376, 0.022996992111206056, 0.023002111434936523, 0.022964223861694336, 0.023008256912231444, 0.023119871139526366, 0.02313932800292969, 0.023022592544555662, 0.023030784606933592, 0.023009279251098632, 0.022921215057373046, 0.02306559944152832, 0.022990848541259764, 0.023144447326660156, 0.02306662368774414, 0.022993919372558593, 0.023187456130981447, 0.02310758399963379, 0.022982656478881838, 0.02309222412109375, 0.023020544052124024, 0.023142400741577147, 0.023266304016113282, 0.023081024169921874, 0.023039936065673828, 0.023121919631958008, 0.022969343185424804, 0.023117824554443358, 0.023021568298339845, 0.02302774429321289, 0.023051231384277344, 0.023006208419799806, 0.02306355285644531, 0.02308403205871582, 0.02312704086303711, 0.02306662368774414, 0.023104511260986327, 0.023133184432983397, 0.023158784866333007, 0.023085056304931642, 0.023061504364013673, 0.02309119987487793, 0.02312396812438965, 0.04583731079101563, 0.022203456878662108, 0.02216134452819824, 0.022226943969726562, 0.022037504196166992, 0.022280256271362306, 0.022361024856567383, 0.02212761688232422, 0.022123552322387694, 0.021957599639892578, 0.022017023086547852, 0.02204979133605957, 0.022320127487182616, 0.022146047592163084, 0.02210304069519043, 0.02234880065917969, 0.022205440521240235, 0.02222591972351074, 0.021987327575683592, 0.0224532470703125, 0.022833152770996092, 0.022792192459106447, 0.023018495559692383, 0.023052288055419923, 0.02313523292541504, 0.023206911087036132, 0.023060543060302734, 0.023125951766967773, 0.023023616790771483, 0.02324684715270996, 0.022953983306884765, 0.02307583999633789, 0.023133184432983397, 0.023149568557739256, 0.02301644706726074, 0.02313216018676758, 0.023147520065307618, 0.02290892791748047, 0.02285977554321289, 0.023060480117797853, 0.02306252861022949, 0.02309222412109375, 0.025249792098999024, 0.023840768814086914, 0.023362560272216795, 0.023129087448120117, 0.023014400482177736, 0.022960128784179686, 0.023056415557861327, 0.023097312927246094, 0.023189504623413085, 0.02311680030822754, 0.023061504364013673, 0.023026687622070312, 0.023216127395629883, 0.023191551208496093, 0.023008256912231444, 0.023112703323364257, 0.02313113594055176, 0.02313113594055176, 0.023143423080444335, 0.023120895385742187, 0.023212032318115236, 0.047608833312988284, 0.02329395294189453, 0.023111679077148437, 0.02310655975341797, 0.02308403205871582, 0.02309119987487793, 0.0230830078125, 0.022982656478881838, 0.023038976669311522, 0.023051263809204102, 0.023031808853149413, 0.02308095932006836, 0.023060480117797853, 0.023021568298339845, 0.0231014404296875, 0.023044095993041993, 0.02312396812438965, 0.02310758399963379, 0.023217151641845703, 0.02310758399963379, 0.023192575454711914, 0.023007232666015624, 0.022994943618774414, 0.022928384780883788, 0.023113727569580078, 0.02289254379272461, 0.02278508758544922, 0.022734783172607423, 0.023166976928710937, 0.023109632492065428, 0.023586816787719726, 0.023093248367309572, 0.023017471313476562, 0.023234560012817384, 0.023157760620117186, 0.023044095993041993, 0.02302566337585449, 0.023150592803955077, 0.023169023513793945, 0.02308608055114746, 0.022939680099487304, 0.023059423446655274, 0.02305023956298828, 0.022963199615478515, 0.02302566337585449, 0.022994943618774414, 0.023315456390380858, 0.023192575454711914, 0.0234803524017334, 0.023378911972045897, 0.023433216094970705, 0.02332467269897461, 0.023188480377197264, 0.0230328311920166, 0.023040000915527343, 0.023022592544555662, 0.023121919631958008, 0.02322329521179199, 0.023230464935302734, 0.023113727569580078, 0.022987775802612305, 0.02313113594055176, 0.023158784866333007, 0.047710208892822265, 0.023205888748168944, 0.023052288055419923, 0.023023616790771483, 0.022976512908935546, 0.022992895126342772, 0.022973440170288087, 0.023619583129882812, 0.02326527976989746, 0.023347200393676756, 0.022972415924072266, 0.023111679077148437, 0.023133184432983397, 0.02310758399963379, 0.02323865509033203, 0.02305843162536621, 0.023149568557739256, 0.023197696685791015, 0.023204864501953124, 0.023143423080444335, 0.023742464065551756, 0.024250368118286132, 0.023478271484375, 0.02314854431152344, 0.023423999786376954, 0.023174144744873046, 0.023266304016113282, 0.023119871139526366, 0.023229440689086913, 0.023147520065307618, 0.023191551208496093, 0.023224319458007812, 0.023133184432983397, 0.023096319198608398, 0.023164928436279295, 0.02327654457092285, 0.02330624008178711, 0.023302143096923827, 0.023157760620117186, 0.023134208679199218, 0.02307072067260742, 0.02306662368774414, 0.023096319198608398, 0.023129087448120117, 0.023036928176879884, 0.02313523292541504, 0.023166976928710937, 0.02310553550720215, 0.02305536079406738, 0.02313113594055176, 0.023143423080444335, 0.023166976928710937, 0.0231014404296875, 0.022992895126342772, 0.02305740737915039, 0.023044095993041993, 0.02310758399963379, 0.02325299263000488, 0.023207935333251953, 0.02304204750061035, 0.02314035224914551, 0.02306662368774414, 0.0230328311920166]",tokens/s,42.63272278414554,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen2-beta-72B,Qwen/Qwen2-beta-72B,cuda,0,42,,,True,,,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",qwen2,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-0.5B,,cuda,0,42,,,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,217063f5c507ed7cc255df7e1f64c4333a0b4dfe,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,,MB,1282.88768,2227.699712,0.0,1581.252608,1426.269184,s,10,0.2108043518066406,0.02108043518066406,0.0005729565326272884,0.021317215919494627,0.021508796882629397,0.021691230869293213,0.021837178058624267,"[0.021873664855957032, 0.02032598304748535, 0.02131977653503418, 0.020177055358886718, 0.02143619155883789, 0.021370784759521484, 0.021301311492919923, 0.020216672897338868, 0.021314655303955078, 0.021468255996704103]",tokens/s,12143.961820807897,kWh,2.410668080626161e-07,1.320933590792191e-07,6.492607234898651e-07,1.0224208906317004e-06,tokens/kWh,250386120.18366623,MB,1283.21536,2229.796864,0.0,1583.34976,1426.271744,s,10,12.93885302734375,1.293885302734375,0.010209288984023323,1.2990811767578125,1.303646923828125,1.3039076293945313,1.3041161938476562,"[1.272334716796875, 1.2902384033203125, 1.301608154296875, 1.279857177734375, 1.3041683349609374, 1.2990198974609375, 1.2996243896484374, 1.2892705078125, 1.3035889892578125, 1.2991424560546876]",tokens/s,48.6905600263499,kWh,1.4401469716206699e-05,7.891491556620538e-06,2.5919316086908614e-05,4.821227735973585e-05,tokens/kWh,1306721.0978217348,,s,629,13.106234365463253,0.02083662061281917,0.002556131621854488,0.020626432418823244,0.020793753814697263,0.020964351654052735,0.04031315856933594,"[0.021301248550415038, 0.019644416809082032, 0.019417087554931642, 0.019405824661254883, 0.019315711975097655, 0.019344383239746094, 0.019401727676391603, 0.019475456237792968, 0.01943654441833496, 0.019418111801147463, 0.019305471420288087, 0.019345439910888673, 0.019362783432006835, 0.019405824661254883, 0.0194467830657959, 0.01934854316711426, 0.019460031509399414, 0.019341312408447265, 0.019273727416992188, 0.019339263916015623, 0.019326976776123047, 0.01964031982421875, 0.019559423446655275, 0.019604480743408204, 0.01943756866455078, 0.019504127502441407, 0.01939356803894043, 0.01949897575378418, 0.02085273551940918, 0.021181440353393553, 0.021117952346801756, 0.02079539108276367, 0.02067967987060547, 0.020586496353149415, 0.0206059513092041, 0.0206878719329834, 0.020586496353149415, 0.02062745666503906, 0.020588544845581053, 0.020775936126708985, 0.0206376953125, 0.020703231811523438, 0.020556800842285155, 0.020702207565307617, 0.020603904724121092, 0.020777984619140624, 0.020723712921142577, 0.02065100860595703, 0.02083635139465332, 0.0206878719329834, 0.021196800231933592, 0.02105753517150879, 0.0206376953125, 0.02066534423828125, 0.020572160720825194, 0.020779008865356444, 0.020616191864013672, 0.020626432418823244, 0.02059775924682617, 0.020658176422119142, 0.020684799194335936, 0.02066022491455078, 0.04036505508422852, 0.019338239669799806, 0.01942732810974121, 0.019825664520263672, 0.019357696533203125, 0.019393535614013673, 0.0193832950592041, 0.019372032165527343, 0.019369983673095705, 0.019341312408447265, 0.01942630386352539, 0.02070425605773926, 0.021015552520751952, 0.02067558479309082, 0.020816896438598635, 0.020717567443847656, 0.020588544845581053, 0.02068070411682129, 0.020862016677856445, 0.020742111206054688, 0.020818912506103515, 0.020726783752441406, 0.020706304550170897, 0.020600831985473633, 0.020566015243530272, 0.020692991256713866, 0.020586496353149415, 0.02067046356201172, 0.02071244812011719, 0.02065510368347168, 0.020609024047851563, 0.02058137512207031, 0.020709375381469726, 0.02064896011352539, 0.02068992042541504, 0.0206376953125, 0.020730880737304686, 0.02063052749633789, 0.02065203285217285, 0.02067251205444336, 0.020569087982177735, 0.020708351135253905, 0.0206059513092041, 0.020634624481201173, 0.020799488067626954, 0.020839424133300782, 0.020702207565307617, 0.02082815933227539, 0.02059775924682617, 0.02068889617919922, 0.02061311912536621, 0.020955135345458984, 0.02062745666503906, 0.020685823440551757, 0.02067967987060547, 0.020620288848876952, 0.020566015243530272, 0.020569087982177735, 0.020802560806274413, 0.0206878719329834, 0.0206878719329834, 0.020710399627685547, 0.020745216369628908, 0.04254207992553711, 0.020619264602661135, 0.02067148780822754, 0.02070528030395508, 0.020662303924560546, 0.020623327255249024, 0.020744192123413087, 0.02064793586730957, 0.02063667106628418, 0.020756479263305663, 0.020685823440551757, 0.020699136734008788, 0.02072166442871094, 0.020624383926391602, 0.020793344497680662, 0.020755456924438476, 0.02068377685546875, 0.02063155174255371, 0.020773887634277344, 0.02064691162109375, 0.02069196891784668, 0.020633600234985353, 0.020730880737304686, 0.020685823440551757, 0.02069708824157715, 0.02065715217590332, 0.020619264602661135, 0.020642816543579103, 0.020649984359741212, 0.020633600234985353, 0.02058140754699707, 0.020655071258544922, 0.02065715217590332, 0.020754432678222655, 0.02067967987060547, 0.020964351654052735, 0.020717567443847656, 0.020546560287475587, 0.020544511795043945, 0.020634624481201173, 0.02062848091125488, 0.020538368225097657, 0.020595712661743162, 0.02063155174255371, 0.020641792297363282, 0.020690944671630858, 0.02064691162109375, 0.020610048294067384, 0.020602880477905275, 0.02062950325012207, 0.020547584533691408, 0.020562944412231447, 0.020577280044555665, 0.020676607131958007, 0.02059878349304199, 0.02066739273071289, 0.020588544845581053, 0.020567039489746093, 0.02065510368347168, 0.020573280334472657, 0.02060892868041992, 0.020520959854125977, 0.020600831985473633, 0.0401797103881836, 0.01941606330871582, 0.01946316719055176, 0.01946214485168457, 0.01944268798828125, 0.01940377616882324, 0.01942630386352539, 0.019404800415039062, 0.01943654441833496, 0.019501056671142578, 0.01947648048400879, 0.019392511367797852, 0.01943142318725586, 0.0194703369140625, 0.019556415557861327, 0.01956857681274414, 0.019430400848388672, 0.01941196823120117, 0.019449855804443358, 0.019378175735473634, 0.019579904556274414, 0.021230592727661132, 0.020847616195678712, 0.020840448379516603, 0.020682752609252928, 0.02063974380493164, 0.020615232467651366, 0.020613056182861328, 0.02069708824157715, 0.020593664169311524, 0.020730880737304686, 0.02124799919128418, 0.020761600494384767, 0.020609024047851563, 0.020773887634277344, 0.020554752349853517, 0.020728864669799806, 0.020620256423950194, 0.0206561279296875, 0.02064076805114746, 0.0206561279296875, 0.020620288848876952, 0.020368383407592772, 0.020331520080566406, 0.020682752609252928, 0.02063667106628418, 0.02062131118774414, 0.020676607131958007, 0.020568063735961914, 0.02068172836303711, 0.020764671325683593, 0.02146406364440918, 0.021727231979370116, 0.02103910446166992, 0.020770816802978515, 0.02081279945373535, 0.02067148780822754, 0.020616191864013672, 0.020595712661743162, 0.02066431999206543, 0.020601856231689454, 0.020587520599365236, 0.020538368225097657, 0.04273561477661133, 0.020694015502929687, 0.020609024047851563, 0.02064896011352539, 0.020668415069580077, 0.020699136734008788, 0.02102272033691406, 0.02064896011352539, 0.02069811248779297, 0.02067967987060547, 0.020723712921142577, 0.02069811248779297, 0.02067865562438965, 0.020745216369628908, 0.020731903076171874, 0.020692991256713866, 0.020695039749145508, 0.020671520233154297, 0.020605920791625976, 0.020686847686767578, 0.020686847686767578, 0.020562944412231447, 0.020609024047851563, 0.02068377685546875, 0.020695039749145508, 0.020709375381469726, 0.02061622428894043, 0.020730911254882814, 0.020680639266967775, 0.02069196891784668, 0.020601856231689454, 0.020700159072875975, 0.020634624481201173, 0.02063564872741699, 0.020568063735961914, 0.020788223266601562, 0.020765695571899414, 0.020754432678222655, 0.020587520599365236, 0.02066534423828125, 0.020582399368286132, 0.020718591690063477, 0.02064793586730957, 0.020701183319091796, 0.020947967529296875, 0.020959232330322267, 0.020753408432006838, 0.020944896697998046, 0.0207391357421875, 0.020826047897338867, 0.020717567443847656, 0.02062233543395996, 0.020489215850830078, 0.020753408432006838, 0.02104729652404785, 0.02067148780822754, 0.020554752349853517, 0.020595712661743162, 0.02051481628417969, 0.02067148780822754, 0.020561920166015626, 0.02061516761779785, 0.020602880477905275, 0.04249292755126953, 0.020569087982177735, 0.020571136474609376, 0.020504575729370117, 0.02066329574584961, 0.020603904724121092, 0.020661247253417968, 0.02060492706298828, 0.02059878349304199, 0.020577280044555665, 0.020572160720825194, 0.020762624740600585, 0.02063155174255371, 0.020559871673583984, 0.020633600234985353, 0.020637760162353514, 0.020584447860717774, 0.020634559631347655, 0.020677631378173827, 0.020610048294067384, 0.020584447860717774, 0.020551679611206054, 0.020674560546875, 0.02048307228088379, 0.020551712036132812, 0.020543455123901367, 0.020601856231689454, 0.02061516761779785, 0.020593664169311524, 0.020587520599365236, 0.020612096786499022, 0.020658176422119142, 0.02061516761779785, 0.020715520858764647, 0.020716543197631835, 0.02043801689147949, 0.020229120254516602, 0.020669439315795898, 0.021005311965942384, 0.02065920066833496, 0.020559871673583984, 0.02059775924682617, 0.02066534423828125, 0.020569087982177735, 0.020642816543579103, 0.020658176422119142, 0.02059878349304199, 0.020800512313842775, 0.020571136474609376, 0.020626432418823244, 0.020624383926391602, 0.02061311912536621, 0.020530176162719727, 0.020625408172607423, 0.020519935607910156, 0.020587520599365236, 0.020568063735961914, 0.020616191864013672, 0.02050764846801758, 0.020526079177856444, 0.020692031860351564, 0.020646848678588868, 0.02079539108276367, 0.04242943954467773, 0.020529151916503906, 0.020642816543579103, 0.020582399368286132, 0.020610048294067384, 0.020487167358398437, 0.020666368484497072, 0.02085478401184082, 0.020626432418823244, 0.02065100860595703, 0.020591615676879883, 0.020519935607910156, 0.020583423614501953, 0.020707328796386718, 0.020666368484497072, 0.02066227149963379, 0.02082815933227539, 0.020577280044555665, 0.02063974380493164, 0.020517887115478514, 0.020576255798339844, 0.02067251205444336, 0.020993024826049804, 0.02064384078979492, 0.020570112228393556, 0.020588544845581053, 0.020562944412231447, 0.020580352783203124, 0.020557823181152343, 0.02068070411682129, 0.020452352523803712, 0.020558847427368163, 0.020552703857421875, 0.020478975296020507, 0.02051180839538574, 0.02061612892150879, 0.020669439315795898, 0.020700159072875975, 0.020571136474609376, 0.020512767791748047, 0.020528127670288086, 0.020567039489746093, 0.02063667106628418, 0.020564992904663085, 0.0206376953125, 0.020611072540283205, 0.020552703857421875, 0.020554752349853517, 0.020588544845581053, 0.020588544845581053, 0.02057318305969238, 0.020553728103637696, 0.020692991256713866, 0.020577280044555665, 0.02067148780822754, 0.020559871673583984, 0.020789247512817383, 0.020624383926391602, 0.020583423614501953, 0.0208035831451416, 0.02086297607421875, 0.020791296005249024, 0.02065100860595703, 0.04008652877807617, 0.019397632598876953, 0.01940275192260742, 0.019508224487304687, 0.01946419143676758, 0.019409919738769533, 0.019401727676391603, 0.019390495300292968, 0.01977142333984375, 0.019848127365112305, 0.02065715217590332, 0.020594688415527345, 0.02065510368347168, 0.020702207565307617, 0.020609024047851563, 0.02060492706298828, 0.020541439056396483, 0.020761600494384767, 0.020669439315795898, 0.020602880477905275, 0.020555776596069338, 0.020613183975219728, 0.020602815628051756, 0.02070528030395508, 0.02069196891784668, 0.020555776596069338, 0.020583423614501953, 0.020592639923095703, 0.020616191864013672, 0.020711423873901368, 0.020566015243530272, 0.02059775924682617, 0.020531200408935548, 0.02062233543395996, 0.020555776596069338, 0.020601856231689454, 0.020582399368286132, 0.02062233543395996, 0.02062335968017578, 0.020591615676879883, 0.02063052749633789, 0.020973567962646485, 0.020759552001953126, 0.020701183319091796, 0.0206561279296875, 0.020718591690063477, 0.020602880477905275, 0.02063871955871582, 0.020584447860717774, 0.020708415985107424, 0.02063968086242676, 0.02065715217590332, 0.02067353630065918, 0.02066431999206543, 0.020610048294067384, 0.02062441635131836, 0.02094179153442383, 0.02061311912536621, 0.02023219108581543, 0.020477951049804686, 0.021215232849121093, 0.020817920684814452, 0.020180992126464844, 0.04240793609619141, 0.020537343978881836, 0.02064793586730957, 0.020572160720825194, 0.02057318305969238, 0.020574207305908202, 0.020618240356445314, 0.020594688415527345, 0.02052403259277344, 0.02058137512207031, 0.020686847686767578, 0.020609024047851563, 0.02060492706298828, 0.02069305610656738, 0.020596672058105468, 0.020607999801635742, 0.020583423614501953, 0.020634624481201173, 0.02063155174255371, 0.020580352783203124, 0.02063871955871582, 0.02064896011352539, 0.020576255798339844, 0.020555776596069338, 0.020765695571899414, 0.020567039489746093, 0.020686847686767578, 0.020582399368286132, 0.0206878719329834, 0.02068889617919922, 0.02050764846801758, 0.02068889617919922, 0.02070124816894531, 0.02057414436340332, 0.02066227149963379, 0.020756479263305663, 0.0208035831451416, 0.020590591430664062, 0.0206878719329834, 0.02061414337158203, 0.020969472885131835, 0.020690944671630858, 0.02062848091125488, 0.02087424087524414, 0.020625408172607423, 0.02061414337158203, 0.020964351654052735, 0.020745216369628908, 0.02069606399536133, 0.020664352416992188, 0.021784543991088867, 0.021000255584716798, 0.020958143234252928, 0.020586496353149415, 0.02060492706298828, 0.02066227149963379, 0.021114879608154297, 0.020779008865356444, 0.0206889591217041, 0.020675519943237304, 0.020549631118774413, 0.02062335968017578, 0.020579328536987306, 0.04263628768920898, 0.020562944412231447, 0.020649984359741212, 0.02067148780822754, 0.020601856231689454, 0.020612096786499022, 0.02062848091125488, 0.020600831985473633, 0.020553728103637696, 0.02060492706298828, 0.020587520599365236, 0.02071244812011719, 0.02064588737487793, 0.02061414337158203, 0.020609024047851563, 0.020798463821411133, 0.02065203285217285, 0.02065203285217285, 0.020642816543579103, 0.020579328536987306, 0.020634624481201173, 0.020644863128662108, 0.02064384078979492, 0.020599807739257812, 0.021104639053344726, 0.020543487548828124, 0.02067967987060547, 0.020508672714233397, 0.020585472106933594, 0.020620288848876952, 0.020555776596069338, 0.020644863128662108, 0.02064076805114746, 0.020564992904663085, 0.020633600234985353, 0.020574207305908202, 0.020527103424072265, 0.020590591430664062, 0.02063052749633789, 0.02052403259277344, 0.020710399627685547, 0.020551679611206054, 0.02069708824157715, 0.020596736907958983, 0.020570112228393556, 0.020563968658447264, 0.020569087982177735, 0.020579328536987306, 0.020149248123168945, 0.020340736389160157, 0.02062335968017578, 0.020497407913208008, 0.020587520599365236, 0.02067148780822754, 0.020726783752441406, 0.02061414337158203, 0.02065407943725586, 0.020574207305908202, 0.020716543197631835, 0.020772863388061523, 0.020564992904663085, 0.020609024047851563, 0.020616191864013672]",tokens/s,47.99242730295607,,,main,False,False,,,,,,,,,,,,,,,,,,,,,,, float16-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-1.8B,,cuda,0,42,,,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,217063f5c507ed7cc255df7e1f64c4333a0b4dfe,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,,MB,1298.96448,5222.432768,0.0,4575.985664,4408.404992,s,10,0.23840172576904298,0.023840172576904295,0.0005106541420409152,0.023670864105224607,0.02390000705718994,0.024634579563140867,0.025222237567901613,"[0.025369152069091797, 0.023706623077392578, 0.02363369560241699, 0.023661151885986328, 0.023677087783813475, 0.023684959411621093, 0.0236376953125, 0.02373676872253418, 0.02366464042663574, 0.02362995147705078]",tokens/s,10738.177300277002,kWh,2.8159645578217885e-07,1.5430172882848377e-07,1.3087894068190103e-06,1.7446875914296731e-06,tokens/kWh,146731140.43885785,MB,1298.96448,5222.432768,0.0,4575.985664,4408.407552,s,10,12.470606933593752,1.247060693359375,0.00442238733599079,1.2463931274414062,1.2521589477539063,1.254918768310547,1.2571266247558595,"[1.2576785888671875, 1.241505615234375, 1.2468797607421875, 1.2479893798828126, 1.2451905517578126, 1.24659228515625, 1.2461939697265625, 1.251545654296875, 1.2443409423828125, 1.242690185546875]",tokens/s,50.518792177057904,kWh,1.4861989345342398e-05,8.14409445473497e-06,3.876174264958031e-05,6.176782644965767e-05,tokens/kWh,1019948.468663481,,s,629,12.662182926177977,0.02013065648041014,0.002876504415868315,0.019755008697509766,0.019963495254516603,0.020318616867065428,0.04381483139038086,"[0.02206105613708496, 0.020191232681274415, 0.020123647689819335, 0.02004787254333496, 0.0196945915222168, 0.01988096046447754, 0.019949567794799804, 0.019783679962158202, 0.01986867141723633, 0.020222976684570314, 0.019762176513671875, 0.019762176513671875, 0.019983360290527344, 0.020343807220458983, 0.02008166313171387, 0.019764223098754884, 0.01967616081237793, 0.019738624572753907, 0.01984819221496582, 0.019794944763183595, 0.01970790481567383, 0.020213760375976563, 0.019758079528808595, 0.019673088073730468, 0.019745855331420897, 0.019688383102416992, 0.019740671157836915, 0.02042367935180664, 0.020728832244873048, 0.020775936126708985, 0.02050764846801758, 0.019894271850585937, 0.0198656005859375, 0.019779584884643556, 0.019750911712646483, 0.01986457633972168, 0.01987379264831543, 0.01983283233642578, 0.01985228729248047, 0.01984614372253418, 0.019842048645019532, 0.019961856842041017, 0.020126752853393555, 0.019762144088745118, 0.019746816635131836, 0.019745792388916016, 0.019760128021240234, 0.019728384017944335, 0.019713024139404296, 0.019730432510375977, 0.019769344329833984, 0.019810304641723633, 0.019851264953613282, 0.01984511947631836, 0.019755008697509766, 0.019777536392211914, 0.019784704208374023, 0.019901439666748046, 0.019812351226806642, 0.019801088333129883, 0.019761152267456054, 0.01983692741394043, 0.0439183349609375, 0.019715072631835938, 0.019772415161132813, 0.019745792388916016, 0.019778560638427735, 0.019738624572753907, 0.01979903984069824, 0.019723264694213868, 0.019748863220214845, 0.019666976928710937, 0.01965667152404785, 0.019647487640380858, 0.019709951400756837, 0.019499008178710937, 0.019628032684326172, 0.01962393569946289, 0.0196945915222168, 0.019697664260864257, 0.019697664260864257, 0.019688447952270507, 0.019489791870117186, 0.019494911193847657, 0.019571712493896484, 0.01983590316772461, 0.01980723190307617, 0.01965158462524414, 0.019754047393798827, 0.01972831916809082, 0.019727392196655272, 0.019847135543823242, 0.019727359771728514, 0.01964031982421875, 0.019507200241088866, 0.019784704208374023, 0.019727359771728514, 0.019783679962158202, 0.01981439971923828, 0.01969152069091797, 0.019563520431518554, 0.019773439407348634, 0.019853311538696287, 0.019520511627197267, 0.019566591262817384, 0.01981439971923828, 0.019817472457885742, 0.019767295837402343, 0.019938304901123048, 0.019550207138061524, 0.01965465545654297, 0.019672063827514647, 0.0198656005859375, 0.019718143463134767, 0.019589120864868165, 0.019519487380981446, 0.019512319564819337, 0.019629056930541993, 0.019594240188598632, 0.01984614372253418, 0.01978982353210449, 0.019959808349609375, 0.019817472457885742, 0.019629056930541993, 0.019535871505737306, 0.04380672073364258, 0.019698688507080078, 0.019570688247680663, 0.019580928802490235, 0.019551231384277345, 0.01962700843811035, 0.019545087814331053, 0.019568639755249022, 0.019555328369140625, 0.019495935440063478, 0.019796991348266603, 0.019715072631835938, 0.019802112579345704, 0.019741695404052736, 0.019734527587890623, 0.019791872024536132, 0.019802112579345704, 0.019768320083618163, 0.01984102439880371, 0.019803136825561524, 0.01985945510864258, 0.01984102439880371, 0.019685375213623048, 0.0198287353515625, 0.019680255889892577, 0.019727359771728514, 0.01968639945983887, 0.01985638427734375, 0.019801088333129883, 0.019739648818969727, 0.019767295837402343, 0.019763200759887696, 0.019744768142700195, 0.01977446365356445, 0.019808256149291992, 0.0198604793548584, 0.01968435287475586, 0.019528703689575197, 0.019729408264160156, 0.01980620765686035, 0.019693567276000978, 0.020011007308959963, 0.019931135177612306, 0.019732511520385743, 0.019831775665283203, 0.019861503601074217, 0.019902463912963866, 0.019813375473022463, 0.019811328887939454, 0.019766271591186522, 0.019735551834106444, 0.01986662483215332, 0.019809280395507813, 0.02062745666503906, 0.019810304641723633, 0.019742719650268553, 0.019735551834106444, 0.01979084777832031, 0.019862527847290038, 0.019764223098754884, 0.020163583755493163, 0.02040115165710449, 0.019919872283935547, 0.04389990234375, 0.019764223098754884, 0.01987993621826172, 0.019767295837402343, 0.019763200759887696, 0.019727359771728514, 0.019886112213134764, 0.019773408889770507, 0.019787776947021486, 0.01983283233642578, 0.019786752700805665, 0.019769344329833984, 0.0198287353515625, 0.019768320083618163, 0.01982156753540039, 0.019796991348266603, 0.019800064086914062, 0.019775487899780272, 0.019813375473022463, 0.020162559509277343, 0.02006937599182129, 0.019783679962158202, 0.019533824920654298, 0.01959321594238281, 0.019564544677734375, 0.019582975387573243, 0.019513343811035155, 0.019520511627197267, 0.019607551574707033, 0.019602432250976562, 0.019577856063842772, 0.01990553665161133, 0.01985740852355957, 0.019783679962158202, 0.019753984451293945, 0.019771392822265626, 0.019731456756591798, 0.019819583892822266, 0.01976313591003418, 0.019687423706054686, 0.019522560119628905, 0.019486719131469727, 0.019751935958862304, 0.019704832077026366, 0.019505151748657225, 0.019518463134765626, 0.019580928802490235, 0.01967103958129883, 0.01990665626525879, 0.01981532859802246, 0.019654720306396485, 0.019724224090576174, 0.019743743896484374, 0.02006118392944336, 0.019769344329833984, 0.019703807830810546, 0.019603456497192383, 0.019636224746704102, 0.020520959854125977, 0.022184959411621095, 0.020471807479858398, 0.019868703842163087, 0.019851232528686525, 0.04391628646850586, 0.01960960006713867, 0.019595264434814453, 0.019737663269042968, 0.01975494384765625, 0.019705856323242187, 0.01981439971923828, 0.01985228729248047, 0.019747840881347657, 0.019853311538696287, 0.01977039909362793, 0.01973244857788086, 0.01971609687805176, 0.019715072631835938, 0.01983283233642578, 0.019743743896484374, 0.01969254493713379, 0.019744768142700195, 0.019740671157836915, 0.019718143463134767, 0.019729408264160156, 0.019725311279296876, 0.0196997127532959, 0.019761152267456054, 0.019722240447998047, 0.01968639945983887, 0.019688447952270507, 0.019713024139404296, 0.019581951141357423, 0.01946316719055176, 0.019503103256225587, 0.019606527328491212, 0.01946316719055176, 0.019528703689575197, 0.019540992736816407, 0.01987174415588379, 0.019677183151245118, 0.019728384017944335, 0.02002739143371582, 0.02017791938781738, 0.020374528884887694, 0.020351999282836913, 0.020057088851928712, 0.019695615768432616, 0.01959219169616699, 0.019843072891235353, 0.019719167709350584, 0.019935232162475586, 0.019760128021240234, 0.019736576080322265, 0.019505151748657225, 0.019497983932495116, 0.01967001533508301, 0.019550207138061524, 0.019471359252929688, 0.019495935440063478, 0.019979263305664064, 0.020353023529052734, 0.019886079788208007, 0.020256767272949217, 0.02004582405090332, 0.019729408264160156, 0.019514368057250975, 0.04391219329833984, 0.01970278358459473, 0.019717119216918946, 0.019720191955566405, 0.01985638427734375, 0.0196997127532959, 0.019772415161132813, 0.019772415161132813, 0.019720191955566405, 0.019697664260864257, 0.019753984451293945, 0.019704864501953124, 0.019889184951782227, 0.01972012710571289, 0.019692607879638672, 0.01976108741760254, 0.019752960205078125, 0.019711999893188475, 0.019741695404052736, 0.019731456756591798, 0.019747840881347657, 0.019751935958862304, 0.019565568923950196, 0.019479551315307618, 0.019681280136108398, 0.019494911193847657, 0.019705856323242187, 0.02046566390991211, 0.020164608001708984, 0.020915199279785156, 0.01982361602783203, 0.019758079528808595, 0.019787839889526367, 0.019821504592895507, 0.019761152267456054, 0.019745792388916016, 0.01988096046447754, 0.019703807830810546, 0.019520511627197267, 0.019520511627197267, 0.01974790382385254, 0.01976927947998047, 0.019732479095458985, 0.019677183151245118, 0.01982156753540039, 0.01981439971923828, 0.019752960205078125, 0.01984716796875, 0.019763200759887696, 0.019843072891235353, 0.019745792388916016, 0.019794944763183595, 0.019861503601074217, 0.019775487899780272, 0.019786752700805665, 0.01977446365356445, 0.019786752700805665, 0.019745792388916016, 0.0196997127532959, 0.019764223098754884, 0.019981311798095702, 0.019794944763183595, 0.0196997127532959, 0.04375040054321289, 0.01945702362060547, 0.01962598419189453, 0.01968639945983887, 0.019752960205078125, 0.01984614372253418, 0.019759103775024413, 0.019753984451293945, 0.01978982353210449, 0.01983795166015625, 0.019765247344970704, 0.019766271591186522, 0.019811328887939454, 0.019750911712646483, 0.019800064086914062, 0.020477951049804686, 0.019723264694213868, 0.01978265571594238, 0.019741695404052736, 0.01977446365356445, 0.019734527587890623, 0.019730432510375977, 0.01988403129577637, 0.020558847427368163, 0.02027212715148926, 0.02027008056640625, 0.019736576080322265, 0.019685375213623048, 0.019826688766479493, 0.020303871154785155, 0.019802112579345704, 0.01984102439880371, 0.019784704208374023, 0.019826688766479493, 0.019693567276000978, 0.019728384017944335, 0.019718143463134767, 0.019748863220214845, 0.019795967102050782, 0.019761152267456054, 0.01962700843811035, 0.01985843276977539, 0.019697664260864257, 0.01966182327270508, 0.019721216201782226, 0.019685375213623048, 0.019523584365844726, 0.019569664001464843, 0.019565568923950196, 0.019490816116333007, 0.019519487380981446, 0.01948876762390137, 0.019494976043701172, 0.019692480087280275, 0.019681280136108398, 0.019679231643676756, 0.019794944763183595, 0.019762176513671875, 0.019887168884277343, 0.01977337646484375, 0.019757055282592775, 0.0196997127532959, 0.019862527847290038, 0.044418048858642575, 0.02021990394592285, 0.02042572784423828, 0.02021990394592285, 0.01990553665161133, 0.019703807830810546, 0.01982259178161621, 0.019805183410644533, 0.01974787139892578, 0.01976419258117676, 0.019735551834106444, 0.019725311279296876, 0.01991372871398926, 0.019731456756591798, 0.019723264694213868, 0.019760128021240234, 0.019748863220214845, 0.019842048645019532, 0.01977446365356445, 0.01988198471069336, 0.019740671157836915, 0.019801088333129883, 0.019711999893188475, 0.019755008697509766, 0.019788799285888673, 0.019817472457885742, 0.019810304641723633, 0.019862527847290038, 0.019787776947021486, 0.019767295837402343, 0.019728384017944335, 0.019753984451293945, 0.019752960205078125, 0.019851264953613282, 0.019742719650268553, 0.019745824813842773, 0.019810272216796876, 0.019779584884643556, 0.019730432510375977, 0.019811328887939454, 0.019796991348266603, 0.019715072631835938, 0.019896320343017578, 0.019756032943725587, 0.019714048385620117, 0.019970048904418947, 0.019750911712646483, 0.019726335525512697, 0.019721216201782226, 0.01986867141723633, 0.019719167709350584, 0.019544063568115236, 0.019589120864868165, 0.01980620765686035, 0.019752960205078125, 0.020616191864013672, 0.01997209548950195, 0.020328447341918944, 0.020497407913208008, 0.020485120773315428, 0.02007961654663086, 0.019791872024536132, 0.019809280395507813, 0.043817985534667966, 0.02004991912841797, 0.01978982353210449, 0.019769344329833984, 0.019760128021240234, 0.019837984085083006, 0.01976419258117676, 0.019744768142700195, 0.01970790481567383, 0.019810304641723633, 0.019776512145996093, 0.01968230438232422, 0.01978265571594238, 0.019786752700805665, 0.019725311279296876, 0.019739648818969727, 0.019696640014648437, 0.019751935958862304, 0.019725311279296876, 0.019778560638427735, 0.019750911712646483, 0.01975606346130371, 0.019735519409179687, 0.0196997127532959, 0.019723264694213868, 0.019729408264160156, 0.019705856323242187, 0.01977446365356445, 0.01985536003112793, 0.01966694450378418, 0.019719167709350584, 0.01982054328918457, 0.01971609687805176, 0.019769344329833984, 0.019793920516967774, 0.019743743896484374, 0.01964851188659668, 0.019719167709350584, 0.019745792388916016, 0.019904512405395508, 0.019500032424926757, 0.019572736740112305, 0.019711999893188475, 0.019742719650268553, 0.019556352615356445, 0.019518463134765626, 0.019697696685791015, 0.01971299171447754, 0.019696640014648437, 0.019722240447998047, 0.01978265571594238, 0.019701759338378907, 0.019765247344970704, 0.019751935958862304, 0.01979084777832031, 0.019688512802124022, 0.019884992599487304, 0.01987276840209961, 0.019760128021240234, 0.01979084777832031, 0.019791872024536132, 0.019788799285888673, 0.019767295837402343, 0.04391321563720703, 0.019793920516967774, 0.019719167709350584, 0.019770368576049805, 0.019741695404052736, 0.01966182327270508, 0.01981439971923828, 0.019695680618286134, 0.01975187110900879, 0.019819520950317384, 0.019687423706054686, 0.0196997127532959, 0.019762176513671875, 0.019732479095458985, 0.019711999893188475, 0.019758079528808595, 0.019532800674438477, 0.019589120864868165, 0.019728384017944335, 0.019759103775024413, 0.019717119216918946, 0.019791872024536132, 0.01975196838378906, 0.019744735717773437, 0.019779584884643556, 0.01982361602783203, 0.019785728454589844, 0.019736576080322265, 0.019768320083618163, 0.019513343811035155, 0.019506175994873046, 0.01945907211303711, 0.019478527069091797, 0.019495935440063478, 0.019704832077026366, 0.01969152069091797, 0.019751935958862304, 0.01966592025756836, 0.01994041633605957, 0.01954092788696289, 0.019628095626831054, 0.019602432250976562, 0.020024255752563478, 0.019668991088867188, 0.01967411231994629, 0.019763200759887696, 0.019711999893188475, 0.019768320083618163, 0.019742719650268553, 0.01983692741394043, 0.019767295837402343, 0.019885120391845704, 0.019687360763549804, 0.01968230438232422, 0.019698688507080078, 0.019709951400756837, 0.019777536392211914, 0.019770368576049805, 0.0196997127532959, 0.019791872024536132, 0.01970790481567383, 0.01970790481567383, 0.019741695404052736]",tokens/s,49.675478838612925,,,main,False,False,,,,,,,,,,,,,,,,,,,,,,, float16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-70m,EleutherAI/pythia-70m,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.42.1,,0.31.0,,,,1.20.0,,,,0.11.1,,,,MB,1297.87904,885.522432,0.0,239.075328,215.592448,s,33,0.1848514566421509,0.0056015592921863895,0.00021439998069873264,0.005567264080047607,0.005707366561889649,0.005759743976593018,0.006384801197052001,"[0.006664127826690674, 0.005791232109069824, 0.005659647941589355, 0.005701375961303711, 0.005738751888275146, 0.005683263778686523, 0.005654911994934082, 0.00559497594833374, 0.005614496231079101, 0.005670976161956787, 0.005687039852142334, 0.005611936092376709, 0.0055324797630310054, 0.005441664218902588, 0.005356416225433349, 0.0056295042037963865, 0.0054960641860961916, 0.005474207878112793, 0.005486144065856934, 0.005554719924926758, 0.005509759902954102, 0.005392640113830566, 0.005503744125366211, 0.0054124479293823245, 0.005502624034881592, 0.0055697917938232425, 0.005491392135620117, 0.005454207897186279, 0.0055658559799194336, 0.005514848232269287, 0.00561407995223999, 0.005567264080047607, 0.005708864212036133]",tokens/s,45701.56034179522,kWh,6.726581921036773e-08,3.685800050730806e-08,1.3602469120387301e-07,2.401485109215488e-07,tokens/kWh,1066007026.308939,MB,1297.87904,885.522432,0.0,239.075328,215.595008,s,33,10.110901977539061,0.3063909690163352,0.005235888562605174,0.3046666564941406,0.313100341796875,0.3138336364746094,0.31839340332031246,"[0.3201356201171875, 0.3146911926269531, 0.3130406494140625, 0.3123949584960938, 0.3132619323730469, 0.3130270690917969, 0.31013104248046874, 0.310192138671875, 0.31257464599609375, 0.3131152648925781, 0.30663223266601564, 0.302794677734375, 0.30060760498046873, 0.29929910278320315, 0.30552670288085937, 0.3015677490234375, 0.30291119384765625, 0.3032377624511719, 0.3068295593261719, 0.30429931640625, 0.2991128234863281, 0.30101226806640624, 0.29974032592773436, 0.3015511169433594, 0.3022117004394531, 0.3025533752441406, 0.3035852355957031, 0.30491546630859373, 0.30410919189453123, 0.306804931640625, 0.3046666564941406, 0.3044737854003906, 0.30989468383789065]",tokens/s,205.6196375574019,kWh,3.6551806857699583e-06,2.0027464723641347e-06,6.3880517086086495e-06,1.2045978866742746e-05,tokens/kWh,5229961.026574116,,s,2078,10.28417228698729,0.0049490723229005315,0.0007523572084945051,0.004816895961761475,0.004998451280593872,0.005034132957458496,0.01077602294921875,"[0.005623807907104492, 0.005470208168029785, 0.005545983791351319, 0.005537792205810547, 0.005620800018310547, 0.0054148478507995606, 0.005500927925109863, 0.005940224170684814, 0.005064703941345215, 0.004893695831298828, 0.004842495918273926, 0.0048230400085449215, 0.004808703899383545, 0.004810751914978028, 0.0047923197746276855, 0.004811776161193848, 0.005092351913452148, 0.00522547197341919, 0.005258240222930908, 0.005348351955413819, 0.005089280128479004, 0.004986879825592041, 0.004996096134185791, 0.0050206718444824215, 0.004994048118591309, 0.005037055969238281, 0.004965375900268554, 0.004980735778808594, 0.0050032639503479, 0.004951039791107178, 0.0050094079971313476, 0.004988927841186524, 0.005039103984832764, 0.0049725441932678225, 0.004978687763214112, 0.005014527797698975, 0.004992000102996826, 0.005022719860076904, 0.004965375900268554, 0.004993023872375488, 0.00499507188796997, 0.0049500160217285155, 0.004999167919158935, 0.0049725441932678225, 0.004971519947052002, 0.00501145601272583, 0.004959231853485108, 0.005054463863372802, 0.004960256099700928, 0.005032959938049316, 0.004964352130889893, 0.0049489917755126955, 0.005017600059509277, 0.004966400146484375, 0.005008384227752686, 0.0050503678321838375, 0.0049500160217285155, 0.005005311965942383, 0.004951039791107178, 0.005005311965942383, 0.004975615978240967, 0.004997119903564453, 0.011374591827392578, 0.005004288196563721, 0.0049725441932678225, 0.005021696090698242, 0.0050032639503479, 0.005007359981536865, 0.0050032639503479, 0.0050032639503479, 0.0050124797821044925, 0.004960256099700928, 0.004976640224456787, 0.004986879825592041, 0.004964352130889893, 0.004999167919158935, 0.004959231853485108, 0.004984831809997559, 0.005019711971282959, 0.004943808078765869, 0.00499507188796997, 0.004959231853485108, 0.0049910078048706055, 0.004991968154907226, 0.004957183837890625, 0.005099520206451416, 0.0049387521743774416, 0.0050227842330932615, 0.004957119941711426, 0.004984831809997559, 0.0049909758567810054, 0.0049530878067016604, 0.0050462718009948735, 0.00495411205291748, 0.004964352130889893, 0.004999199867248535, 0.004956128120422364, 0.005031936168670655, 0.004926464080810547, 0.005000192165374756, 0.004974592208862305, 0.004957183837890625, 0.005004288196563721, 0.004960256099700928, 0.00520908784866333, 0.0051404800415039064, 0.004966400146484375, 0.0050206718444824215, 0.0050124797821044925, 0.005048319816589355, 0.005007359981536865, 0.0050032639503479, 0.004987904071807861, 0.004988927841186524, 0.005059584140777588, 0.00496230411529541, 0.004976640224456787, 0.00491212797164917, 0.004890624046325683, 0.0049500160217285155, 0.004881408214569092, 0.004965375900268554, 0.004900864124298096, 0.004931583881378174, 0.004897791862487793, 0.011135999679565429, 0.004960256099700928, 0.004988927841186524, 0.004984831809997559, 0.004958208084106445, 0.005001215934753418, 0.00496230411529541, 0.005000192165374756, 0.004973567962646484, 0.004964352130889893, 0.004992000102996826, 0.0049459199905395506, 0.00496230411529541, 0.004890624046325683, 0.00491212797164917, 0.004908031940460205, 0.0049489917755126955, 0.005090303897857666, 0.005172224044799805, 0.005033984184265137, 0.005014527797698975, 0.00496127986907959, 0.005005311965942383, 0.004974592208862305, 0.005000192165374756, 0.004929599761962891, 0.004895679950714112, 0.0050135040283203125, 0.004899839878082276, 0.004964352130889893, 0.004882431983947754, 0.004926464080810547, 0.0049192957878112795, 0.00491315221786499, 0.00495411205291748, 0.005017600059509277, 0.004999167919158935, 0.005073919773101807, 0.004978720188140869, 0.0050349760055541995, 0.004947968006134033, 0.005022719860076904, 0.004975615978240967, 0.004888576030731201, 0.004925471782684326, 0.004888544082641602, 0.004899839878082276, 0.004936704158782959, 0.004876287937164306, 0.004935679912567138, 0.004981760025024414, 0.004975615978240967, 0.004960256099700928, 0.004892672061920166, 0.004965375900268554, 0.0048578882217407224, 0.004902880191802978, 0.004910079956054687, 0.005000192165374756, 0.004984831809997559, 0.004888576030731201, 0.005004288196563721, 0.004956160068511963, 0.011132927894592285, 0.004924416065216064, 0.004958208084106445, 0.004863999843597412, 0.004928512096405029, 0.004893695831298828, 0.004888576030731201, 0.0049203200340270994, 0.004872191905975342, 0.004971519947052002, 0.004905983924865722, 0.004901887893676758, 0.005030911922454834, 0.004959231853485108, 0.005018623828887939, 0.004964352130889893, 0.004893695831298828, 0.004987904071807861, 0.004960256099700928, 0.0050094079971313476, 0.004978687763214112, 0.005031968116760254, 0.005039072036743164, 0.004941823959350586, 0.0050104641914367675, 0.004946911811828613, 0.004996096134185791, 0.004983808040618896, 0.00495411205291748, 0.0050135040283203125, 0.004899839878082276, 0.005000192165374756, 0.004898816108703613, 0.0048895998001098635, 0.004955135822296143, 0.004914175987243652, 0.004980735778808594, 0.004896768093109131, 0.0049192957878112795, 0.004959231853485108, 0.0048865280151367185, 0.004951039791107178, 0.004905983924865722, 0.0049203200340270994, 0.00496127986907959, 0.004869120121002197, 0.004947968006134033, 0.004877312183380127, 0.004935679912567138, 0.004908031940460205, 0.00486300802230835, 0.004929535865783692, 0.004879327774047851, 0.004925439834594727, 0.005093376159667969, 0.005058559894561767, 0.005112832069396973, 0.004967455863952637, 0.004994016170501709, 0.004981760025024414, 0.004983808040618896, 0.0050135040283203125, 0.004958208084106445, 0.01113804817199707, 0.004965375900268554, 0.005028863906860351, 0.004966400146484375, 0.0050206718444824215, 0.004959231853485108, 0.004997119903564453, 0.0049725441932678225, 0.004997119903564453, 0.004986879825592041, 0.004975615978240967, 0.005022719860076904, 0.004947968006134033, 0.004956160068511963, 0.004985856056213379, 0.004974592208862305, 0.005023744106292725, 0.0049500160217285155, 0.004984831809997559, 0.00499507188796997, 0.0049530878067016604, 0.0050094079971313476, 0.0049797120094299315, 0.004976640224456787, 0.005000192165374756, 0.004944896221160889, 0.005029888153076172, 0.004959231853485108, 0.0050657281875610355, 0.0050360321998596195, 0.0050289278030395505, 0.005016511917114258, 0.004960256099700928, 0.00501145601272583, 0.004973567962646484, 0.004966400146484375, 0.0050206718444824215, 0.004951039791107178, 0.005062655925750732, 0.004882431983947754, 0.004933631896972656, 0.004921343803405762, 0.004868095874786377, 0.004971519947052002, 0.004985856056213379, 0.005024767875671386, 0.0049797120094299315, 0.004936704158782959, 0.004963327884674072, 0.004904960155487061, 0.004923391819000244, 0.004878335952758789, 0.004925439834594727, 0.00491315221786499, 0.0048855037689208985, 0.004917247772216797, 0.0048568320274353025, 0.004915200233459473, 0.004901887893676758, 0.004878335952758789, 0.0049192957878112795, 0.004921343803405762, 0.0050432000160217285, 0.011096063613891602, 0.004985856056213379, 0.004944896221160889, 0.00501043176651001, 0.004983808040618896, 0.0049909758567810054, 0.004987904071807861, 0.004965375900268554, 0.004998144149780274, 0.004959231853485108, 0.004952064037322998, 0.004986879825592041, 0.004986879825592041, 0.004976640224456787, 0.004931583881378174, 0.004992000102996826, 0.004975615978240967, 0.0049725441932678225, 0.0049909758567810054, 0.004937727928161621, 0.004897791862487793, 0.004929535865783692, 0.004868095874786377, 0.004936704158782959, 0.00486297607421875, 0.004966400146484375, 0.004907008171081543, 0.005125120162963868, 0.005058559894561767, 0.004973567962646484, 0.004987904071807861, 0.0049459199905395506, 0.00501145601272583, 0.00501043176651001, 0.004908031940460205, 0.0050135040283203125, 0.004907008171081543, 0.004963327884674072, 0.00501043176651001, 0.0049500160217285155, 0.004951039791107178, 0.004963327884674072, 0.004999167919158935, 0.004960256099700928, 0.00496230411529541, 0.005018623828887939, 0.004955135822296143, 0.00501964807510376, 0.004969471931457519, 0.005018623828887939, 0.005024767875671386, 0.004952064037322998, 0.0049797120094299315, 0.0048895998001098635, 0.004929535865783692, 0.0049192957878112795, 0.0048947839736938475, 0.004932544231414795, 0.004907008171081543, 0.004918272018432617, 0.004892672061920166, 0.0048865280151367185, 0.004921343803405762, 0.011053055763244628, 0.0048527359962463375, 0.004893695831298828, 0.004897791862487793, 0.004878335952758789, 0.0049162240028381345, 0.0048865280151367185, 0.004890624046325683, 0.004904960155487061, 0.00490499210357666, 0.004901855945587158, 0.0049909758567810054, 0.004975615978240967, 0.004984831809997559, 0.004944896221160889, 0.00501145601272583, 0.0049090561866760255, 0.004929535865783692, 0.004881408214569092, 0.0048793601989746095, 0.0049090561866760255, 0.0048568320274353025, 0.0049090561866760255, 0.004907008171081543, 0.00486297607421875, 0.0049203200340270994, 0.0048527359962463375, 0.004907008171081543, 0.004898848056793213, 0.004869152069091797, 0.004904895782470703, 0.0048752641677856446, 0.004927487850189209, 0.0048752641677856446, 0.004770815849304199, 0.005024767875671386, 0.004926464080810547, 0.004998144149780274, 0.0049725441932678225, 0.004940800189971924, 0.004993023872375488, 0.00496230411529541, 0.004999167919158935, 0.004965375900268554, 0.004969503879547119, 0.005016608238220215, 0.004958144187927246, 0.0050124797821044925, 0.00491315221786499, 0.004926464080810547, 0.004839424133300781, 0.004738048076629638, 0.004944896221160889, 0.0048855037689208985, 0.004927487850189209, 0.004894720077514648, 0.004846591949462891, 0.004927487850189209, 0.004890624046325683, 0.004878335952758789, 0.004918272018432617, 0.0048793601989746095, 0.004937727928161621, 0.011006976127624512, 0.004927487850189209, 0.004974656105041504, 0.0049069437980651855, 0.004868095874786377, 0.004918272018432617, 0.004887551784515381, 0.004876287937164306, 0.004867072105407715, 0.004888576030731201, 0.004906015872955322, 0.004866015911102295, 0.005021696090698242, 0.004966400146484375, 0.004935679912567138, 0.005023744106292725, 0.004941823959350586, 0.004984831809997559, 0.004973567962646484, 0.0049500160217285155, 0.005001215934753418, 0.004922431945800781, 0.004902847766876221, 0.00491315221786499, 0.004860928058624267, 0.004905983924865722, 0.00486195182800293, 0.004940800189971924, 0.004996096134185791, 0.004943871974945068, 0.005053440093994141, 0.004874239921569825, 0.0049459199905395506, 0.004858880043029785, 0.004836351871490479, 0.0049162240028381345, 0.004917247772216797, 0.004917247772216797, 0.004851712226867676, 0.004878335952758789, 0.0049203200340270994, 0.004892672061920166, 0.004915232181549072, 0.004865024089813232, 0.004889567852020263, 0.004927487850189209, 0.00486195182800293, 0.004900864124298096, 0.004910079956054687, 0.0048855037689208985, 0.00491212797164917, 0.0048568320274353025, 0.004946944236755371, 0.004937727928161621, 0.004935679912567138, 0.004917247772216797, 0.004899839878082276, 0.00495411205291748, 0.004895743846893311, 0.004900864124298096, 0.004911104202270508, 0.00487116813659668, 0.0049213762283325194, 0.011032544136047363, 0.004929535865783692, 0.0049827837944030765, 0.0049459199905395506, 0.004969471931457519, 0.0050135040283203125, 0.0049387521743774416, 0.004882431983947754, 0.004903935909271241, 0.004894720077514648, 0.004965375900268554, 0.004869120121002197, 0.004947968006134033, 0.004975615978240967, 0.005038080215454102, 0.005024767875671386, 0.004948031902313232, 0.004997056007385254, 0.004946944236755371, 0.004944896221160889, 0.0049838399887084965, 0.004959199905395508, 0.004952064037322998, 0.004986879825592041, 0.004925439834594727, 0.00499507188796997, 0.0048865280151367185, 0.004941855907440186, 0.004872159957885742, 0.004882431983947754, 0.004915200233459473, 0.004926464080810547, 0.005000192165374756, 0.00496127986907959, 0.004989952087402344, 0.0050022401809692385, 0.0049909758567810054, 0.0050769920349121095, 0.005222400188446045, 0.0050769920349121095, 0.004968448162078858, 0.004959231853485108, 0.0050063362121582035, 0.005004288196563721, 0.0050032639503479, 0.004997119903564453, 0.004918272018432617, 0.004959231853485108, 0.0048895998001098635, 0.004951039791107178, 0.004939775943756103, 0.004890624046325683, 0.004936704158782959, 0.004873216152191162, 0.004976640224456787, 0.004898816108703613, 0.0049459199905395506, 0.004931583881378174, 0.004869120121002197, 0.004933631896972656, 0.004874239921569825, 0.004911104202270508, 0.004896768093109131, 0.011125760078430176, 0.004944896221160889, 0.0050094079971313476, 0.00496230411529541, 0.005015552043914795, 0.004978687763214112, 0.004971519947052002, 0.005005311965942383, 0.004988927841186524, 0.00496230411529541, 0.004976640224456787, 0.004971519947052002, 0.0050124797821044925, 0.0049428482055664065, 0.004985856056213379, 0.00496127986907959, 0.00496230411529541, 0.004981760025024414, 0.0049725441932678225, 0.004997119903564453, 0.004977663993835449, 0.00499507188796997, 0.005005311965942383, 0.0049459199905395506, 0.005015552043914795, 0.0049428482055664065, 0.004966400146484375, 0.004960256099700928, 0.004976704120635987, 0.004954048156738281, 0.004884543895721435, 0.00492742395401001, 0.004996096134185791, 0.004952064037322998, 0.004984831809997559, 0.004964352130889893, 0.004964352130889893, 0.004993023872375488, 0.004967423915863037, 0.00501964807510376, 0.004935679912567138, 0.0050135040283203125, 0.004924416065216064, 0.004924416065216064, 0.004996096134185791, 0.004960256099700928, 0.005047327995300293, 0.004955103874206543, 0.004936704158782959, 0.004964352130889893, 0.004911104202270508, 0.005000192165374756, 0.004926464080810547, 0.004994048118591309, 0.004952064037322998, 0.0048895998001098635, 0.004964352130889893, 0.004899839878082276, 0.004908031940460205, 0.004878335952758789, 0.004865024089813232, 0.004894720077514648, 0.0049459199905395506, 0.011126784324645997, 0.00497049617767334, 0.004968448162078858, 0.004956160068511963, 0.004985856056213379, 0.004943871974945068, 0.004967423915863037, 0.004992000102996826, 0.0049530878067016604, 0.005243904113769531, 0.004866047859191895, 0.0048056321144104, 0.004783103942871094, 0.004825088024139404, 0.004788224220275879, 0.004787199974060059, 0.0047820801734924315, 0.004788224220275879, 0.00481382417678833, 0.004780032157897949, 0.004776959896087647, 0.004776959896087647, 0.004779007911682129, 0.004760575771331787, 0.004800511837005615, 0.004779007911682129, 0.004774911880493164, 0.004783103942871094, 0.004767744064331054, 0.004806655883789063, 0.004770815849304199, 0.004776959896087647, 0.0047923197746276855, 0.004830207824707031, 0.004789247989654541, 0.004780032157897949, 0.004781055927276612, 0.004789247989654541, 0.004789247989654541, 0.004800511837005615, 0.004770815849304199, 0.004780032157897949, 0.0051066880226135255, 0.0051333122253417966, 0.004968448162078858, 0.0049459199905395506, 0.00497049617767334, 0.004996096134185791, 0.004967423915863037, 0.004937727928161621, 0.0050462718009948735, 0.004994048118591309, 0.004857855796813965, 0.004791296005249023, 0.004784128189086914, 0.0047861762046813965, 0.004803584098815918, 0.004783167839050293, 0.00477894401550293, 0.004850687980651855, 0.004781055927276612, 0.004921343803405762, 0.004773888111114502, 0.010884096145629883, 0.004779007911682129, 0.004790272235870361, 0.004784128189086914, 0.004807680130004883, 0.004790272235870361, 0.004801536083221435, 0.004789247989654541, 0.0047933759689331055, 0.004783071994781494, 0.004781055927276612, 0.004798463821411133, 0.004800511837005615, 0.0047861762046813965, 0.0047861762046813965, 0.004776959896087647, 0.004794367790222168, 0.004789247989654541, 0.004785183906555176, 0.004793312072753906, 0.004781055927276612, 0.0048261117935180665, 0.004799488067626953, 0.0047820801734924315, 0.004829184055328369, 0.004775936126708984, 0.00481382417678833, 0.004769792079925537, 0.004770815849304199, 0.0047964158058166504, 0.004781055927276612, 0.004800511837005615, 0.004779007911682129, 0.0047933440208435055, 0.004795455932617187, 0.004799424171447754, 0.004817920207977295, 0.0047851519584655765, 0.00480460786819458, 0.00479744005203247, 0.004811776161193848, 0.004810751914978028, 0.004787199974060059, 0.0047933440208435055, 0.00479744005203247, 0.004943871974945068, 0.004802559852600098, 0.0047964158058166504, 0.004798463821411133, 0.0047851519584655765, 0.004933631896972656, 0.0048261117935180665, 0.004788224220275879, 0.004788224220275879, 0.0047933440208435055, 0.004799488067626953, 0.004795392036437988, 0.004799488067626953, 0.004790272235870361, 0.004790272235870361, 0.0047851519584655765, 0.004824063777923584, 0.004806655883789063, 0.010777600288391113, 0.004772863864898682, 0.004790272235870361, 0.004784160137176514, 0.004769760131835938, 0.0048189439773559575, 0.004768767833709717, 0.004764671802520752, 0.004770815849304199, 0.004789311885833741, 0.004794303894042969, 0.004781055927276612, 0.004801536083221435, 0.004791296005249023, 0.004926464080810547, 0.004811776161193848, 0.004803584098815918, 0.004816895961761475, 0.004790272235870361, 0.0047933440208435055, 0.0047851519584655765, 0.0047288317680358885, 0.004715519905090332, 0.0047216639518737795, 0.0047226881980895995, 0.004726784229278564, 0.004741119861602783, 0.004724736213684082, 0.004750336170196533, 0.004737023830413818, 0.004716544151306152, 0.004743167877197266, 0.004771840095520019, 0.004783103942871094, 0.004808703899383545, 0.004825088024139404, 0.004798463821411133, 0.004794367790222168, 0.0047861762046813965, 0.004798463821411133, 0.004777984142303467, 0.00481382417678833, 0.004736000061035156, 0.0047185921669006346, 0.004728896141052246, 0.004741055965423584, 0.004717567920684815, 0.004746240139007568, 0.004765696048736572, 0.004729856014251709, 0.0047257599830627444, 0.004774911880493164, 0.004737023830413818, 0.004729856014251709, 0.004702208042144776, 0.004724736213684082, 0.004733952045440673, 0.004740096092224121, 0.004747263908386231, 0.0047257599830627444, 0.0047329277992248535, 0.004750336170196533, 0.0047636480331420894, 0.010660863876342774, 0.004765696048736572, 0.004759552001953125, 0.0047820801734924315, 0.004773888111114502, 0.004765696048736572, 0.004775936126708984, 0.004780032157897949, 0.004794367790222168, 0.0047626237869262695, 0.0047185921669006346, 0.0047391037940979, 0.00471343994140625, 0.004745215892791748, 0.004726784229278564, 0.004709375858306885, 0.004726784229278564, 0.004716544151306152, 0.004700160026550293, 0.0047513599395751956, 0.004711423873901367, 0.0047185921669006346, 0.004731904029846192, 0.004724736213684082, 0.004738048076629638, 0.0047257599830627444, 0.0047288317680358885, 0.004717567920684815, 0.004731904029846192, 0.00470630407333374, 0.004730879783630371, 0.004764671802520752, 0.004723711967468262, 0.004720640182495117, 0.0047257599830627444, 0.004741119861602783, 0.004716544151306152, 0.004780032157897949, 0.004734975814819336, 0.004720672130584717, 0.004723680019378662, 0.004757503986358643, 0.004737023830413818, 0.004711423873901367, 0.0047185921669006346, 0.0047216639518737795, 0.004760575771331787, 0.004720704078674316, 0.0047103362083435055, 0.004716544151306152, 0.004720640182495117, 0.0047216639518737795, 0.0047554559707641605, 0.004791296005249023, 0.004771840095520019, 0.0047923197746276855, 0.004772863864898682, 0.0048148479461669925, 0.004760575771331787, 0.004781055927276612, 0.004779007911682129, 0.004777984142303467, 0.004809792041778564, 0.010501055717468262, 0.00470630407333374, 0.004709375858306885, 0.0047288317680358885, 0.004768767833709717, 0.004777984142303467, 0.004777984142303467, 0.004800511837005615, 0.004782144069671631, 0.004775872230529785, 0.004769792079925537, 0.004801536083221435, 0.0047861762046813965, 0.004777984142303467, 0.004767744064331054, 0.004770815849304199, 0.004770815849304199, 0.004772863864898682, 0.004798495769500732, 0.004791264057159424, 0.004772863864898682, 0.0047851519584655765, 0.004767744064331054, 0.004789247989654541, 0.004790272235870361, 0.0047851519584655765, 0.0047861762046813965, 0.004783103942871094, 0.0047964158058166504, 0.004798463821411133, 0.004787199974060059, 0.004798463821411133, 0.004848639965057373, 0.0048148479461669925, 0.004780032157897949, 0.004799488067626953, 0.004799488067626953, 0.004787199974060059, 0.004801536083221435, 0.004779007911682129, 0.004776959896087647, 0.004799488067626953, 0.004781055927276612, 0.004801536083221435, 0.004790272235870361, 0.004779007911682129, 0.004795392036437988, 0.004787199974060059, 0.004830207824707031, 0.004794367790222168, 0.0050360321998596195, 0.005294079780578613, 0.005037055969238281, 0.004976640224456787, 0.005018623828887939, 0.004975615978240967, 0.0051138558387756345, 0.0049827837944030765, 0.004957183837890625, 0.005004288196563721, 0.005796864032745362, 0.004934656143188477, 0.004836351871490479, 0.010897407531738281, 0.004807680130004883, 0.004776959896087647, 0.004777984142303467, 0.004810751914978028, 0.0047513599395751956, 0.004758528232574463, 0.004738048076629638, 0.004781055927276612, 0.0047329277992248535, 0.0047933440208435055, 0.004759552001953125, 0.004726784229278564, 0.0047523841857910155, 0.004736000061035156, 0.004761600017547608, 0.004746240139007568, 0.004746240139007568, 0.004746240139007568, 0.0047636480331420894, 0.0047554559707641605, 0.004743167877197266, 0.004759552001953125, 0.004765696048736572, 0.004905983924865722, 0.004829184055328369, 0.004761600017547608, 0.004760575771331787, 0.004788224220275879, 0.004753407955169678, 0.004726784229278564, 0.004741119861602783, 0.004750336170196533, 0.004880383968353271, 0.004736000061035156, 0.004727807998657227, 0.004726784229278564, 0.004758528232574463, 0.004748288154602051, 0.004760575771331787, 0.004740096092224121, 0.0047626237869262695, 0.004729856014251709, 0.004747263908386231, 0.004730879783630371, 0.004760575771331787, 0.004723711967468262, 0.0047933440208435055, 0.0047964158058166504, 0.004808703899383545, 0.00502784013748169, 0.004914175987243652, 0.00487014389038086, 0.004986879825592041, 0.00487014389038086, 0.004825088024139404, 0.0047861762046813965, 0.004743167877197266, 0.004727807998657227, 0.004866047859191895, 0.004817920207977295, 0.004743167877197266, 0.0047381119728088375, 0.010719167709350585, 0.004738048076629638, 0.004742144107818603, 0.004726784229278564, 0.004754432201385498, 0.004730879783630371, 0.004726784229278564, 0.004726784229278564, 0.0047257599830627444, 0.004733952045440673, 0.004749311923980713, 0.0047329277992248535, 0.0048855037689208985, 0.004788224220275879, 0.004874239921569825, 0.004867072105407715, 0.0048455681800842285, 0.00475551986694336, 0.004759488105773926, 0.004858880043029785, 0.0047851519584655765, 0.004779007911682129, 0.004819968223571777, 0.00480460786819458, 0.004837376117706299, 0.004788224220275879, 0.0047861762046813965, 0.004823103904724121, 0.0047840638160705565, 0.004836351871490479, 0.0047923197746276855, 0.004787199974060059, 0.004829184055328369, 0.004831232070922851, 0.004830207824707031, 0.004791296005249023, 0.004791296005249023, 0.004836351871490479, 0.004789247989654541, 0.004816895961761475, 0.004801536083221435, 0.004788224220275879, 0.004816895961761475, 0.004806655883789063, 0.004801536083221435, 0.004897791862487793, 0.004850687980651855, 0.004828159809112549, 0.0047964158058166504, 0.004863999843597412, 0.004809728145599365, 0.004800511837005615, 0.004848671913146973, 0.004793312072753906, 0.0048220157623291016, 0.004808703899383545, 0.004802559852600098, 0.004820991992950439, 0.004794367790222168, 0.0047964158058166504, 0.004866047859191895, 0.004816895961761475, 0.0048261117935180665, 0.010636320114135742, 0.004810719966888428, 0.004887551784515381, 0.0049387521743774416, 0.004801536083221435, 0.004831232070922851, 0.004835328102111816, 0.0047933440208435055, 0.004902912139892578, 0.004795392036437988, 0.0048148479461669925, 0.004833280086517334, 0.004833280086517334, 0.004911136150360107, 0.004828127861022949, 0.004839424133300781, 0.004839424133300781, 0.005054463863372802, 0.004985856056213379, 0.004857855796813965, 0.004824063777923584, 0.0048220157623291016, 0.004866047859191895, 0.0048527359962463375, 0.004770815849304199, 0.004749311923980713, 0.004761600017547608, 0.004775936126708984, 0.004772863864898682, 0.004734975814819336, 0.004741119861602783, 0.0047554559707641605, 0.00475648021697998, 0.004749311923980713, 0.004750336170196533, 0.004739071846008301, 0.004766719818115234, 0.004720640182495117, 0.004759552001953125, 0.004748288154602051, 0.004734975814819336, 0.004772863864898682, 0.004747263908386231, 0.004739071846008301, 0.004742144107818603, 0.004729919910430909, 0.004759488105773926, 0.004737023830413818, 0.0047554559707641605, 0.004779007911682129, 0.004734975814819336, 0.004771840095520019, 0.0047257599830627444, 0.004731904029846192, 0.004775936126708984, 0.004911104202270508, 0.0049909758567810054, 0.004924416065216064, 0.0049572157859802245, 0.004843488216400147, 0.004801536083221435, 0.004808703899383545, 0.0047820801734924315, 0.010829824447631836, 0.004784128189086914, 0.004839424133300781, 0.004779039859771728, 0.004792287826538086, 0.0050165758132934574, 0.004898816108703613, 0.0048527359962463375, 0.004789247989654541, 0.0047964158058166504, 0.004811776161193848, 0.004790272235870361, 0.004847616195678711, 0.004959231853485108, 0.004896768093109131, 0.004927487850189209, 0.004914175987243652, 0.004922368049621582, 0.004820991992950439, 0.004788224220275879, 0.0048158721923828125, 0.004802559852600098, 0.004905983924865722, 0.004907008171081543, 0.004940800189971924, 0.005244927883148193, 0.005496831893920898, 0.004983808040618896, 0.004928512096405029, 0.00499507188796997, 0.004939775943756103, 0.004929535865783692, 0.004957183837890625, 0.0049827837944030765, 0.004931583881378174, 0.0048568320274353025, 0.004809728145599365, 0.004863999843597412, 0.0047288317680358885, 0.004767744064331054, 0.004781055927276612, 0.0047851519584655765, 0.004918272018432617, 0.0047861762046813965, 0.00480460786819458, 0.004765696048736572, 0.004774911880493164, 0.004790272235870361, 0.0047933440208435055, 0.004801536083221435, 0.004780032157897949, 0.004794367790222168, 0.004819968223571777, 0.004835328102111816, 0.004792352199554443, 0.004809696197509766, 0.004801536083221435, 0.004832255840301514, 0.004798463821411133, 0.004783103942871094, 0.004810751914978028, 0.0048220157623291016, 0.004842495918273926, 0.010793984413146973, 0.004781055927276612, 0.004831232070922851, 0.004924416065216064, 0.004900864124298096, 0.004935679912567138, 0.004918272018432617, 0.004897791862487793, 0.004894720077514648, 0.004924416065216064, 0.004905983924865722, 0.004897791862487793, 0.004742144107818603, 0.004743167877197266, 0.0047298879623413085, 0.004753376007080078, 0.004745215892791748, 0.004741119861602783, 0.00475648021697998, 0.004775936126708984, 0.004772863864898682, 0.004810751914978028, 0.0047861762046813965, 0.0048056321144104, 0.0049192957878112795, 0.004916255950927735, 0.0049387202262878415, 0.004887551784515381, 0.004841472148895264, 0.004890624046325683, 0.004863999843597412, 0.004934656143188477, 0.004896768093109131, 0.0048793601989746095, 0.00481382417678833, 0.004863999843597412, 0.004795392036437988, 0.004780032157897949, 0.004809728145599365, 0.0047820801734924315, 0.00481279993057251, 0.004808703899383545, 0.004773888111114502, 0.004806655883789063, 0.004774911880493164, 0.004828159809112549, 0.0047933440208435055, 0.004771840095520019, 0.004858880043029785, 0.0047964158058166504, 0.004788224220275879, 0.004787199974060059, 0.004772863864898682, 0.004803584098815918, 0.004780032157897949, 0.004810751914978028, 0.004808703899383545, 0.00481279993057251, 0.004820991992950439, 0.004780032157897949, 0.0048189439773559575, 0.004784128189086914, 0.0047216639518737795, 0.01075814437866211, 0.004740096092224121, 0.004743167877197266, 0.004769792079925537, 0.004757503986358643, 0.00475648021697998, 0.004726784229278564, 0.004740096092224121, 0.0047523841857910155, 0.0047554559707641605, 0.0047513599395751956, 0.004724736213684082, 0.004739071846008301, 0.004724736213684082, 0.0047329277992248535, 0.004724736213684082, 0.0047288317680358885, 0.004736000061035156, 0.004720640182495117, 0.004736000061035156, 0.004768767833709717, 0.0047329277992248535, 0.004727871894836426, 0.004714431762695312, 0.004765696048736572, 0.004745215892791748, 0.004740096092224121, 0.004753407955169678, 0.0047329277992248535, 0.0047257599830627444, 0.004750336170196533, 0.004741119861602783, 0.004789247989654541, 0.004726784229278564, 0.004729856014251709, 0.004772863864898682, 0.0047288317680358885, 0.0047554559707641605, 0.004750336170196533, 0.004740096092224121, 0.004761600017547608, 0.004727807998657227, 0.0047216639518737795, 0.004757503986358643, 0.004727807998657227, 0.004749311923980713, 0.0047329277992248535, 0.004737023830413818, 0.004781055927276612, 0.004724736213684082, 0.004748288154602051, 0.004727807998657227, 0.004715519905090332, 0.004789247989654541, 0.004729856014251709, 0.004776959896087647, 0.004730879783630371, 0.004719615936279297, 0.004723711967468262, 0.004736000061035156, 0.0046991357803344725, 0.004716544151306152, 0.004727807998657227, 0.01084928035736084, 0.004754432201385498, 0.004760575771331787, 0.004720640182495117, 0.004783135890960693, 0.004722655773162842, 0.00471449613571167, 0.004820991992950439, 0.004794367790222168, 0.004817920207977295, 0.004771840095520019, 0.004789247989654541, 0.004801536083221435, 0.004775936126708984, 0.004827136039733886, 0.004767744064331054, 0.004768767833709717, 0.0047820801734924315, 0.004769792079925537, 0.004767744064331054, 0.0047216639518737795, 0.0047216639518737795, 0.004754432201385498, 0.004717567920684815, 0.004740096092224121, 0.0048056321144104, 0.004726784229278564, 0.004803584098815918, 0.004783135890960693, 0.004786143779754639, 0.004827136039733886, 0.004792352199554443, 0.004830175876617431, 0.004795392036437988, 0.004799488067626953, 0.004825088024139404, 0.004776959896087647, 0.004801536083221435, 0.004743167877197266, 0.004724736213684082, 0.004767744064331054, 0.004720640182495117, 0.004799488067626953, 0.004834303855895996, 0.004780032157897949, 0.00487116813659668, 0.004773888111114502, 0.0047851519584655765, 0.004753407955169678, 0.004749311923980713, 0.004810751914978028, 0.0047185921669006346, 0.004747263908386231, 0.0047329277992248535, 0.004702208042144776, 0.004758592128753662, 0.004712384223937988, 0.004715519905090332, 0.004726784229278564, 0.004730879783630371, 0.004743167877197266, 0.0047329277992248535, 0.0047226881980895995, 0.010652671813964844, 0.0047636480331420894, 0.004710400104522705, 0.004724736213684082, 0.004753407955169678, 0.004779007911682129, 0.00481279993057251, 0.0047626237869262695, 0.004769792079925537, 0.004808703899383545, 0.004775936126708984, 0.004834303855895996, 0.004789247989654541, 0.004764671802520752, 0.004766719818115234, 0.004730879783630371, 0.004750336170196533, 0.004727807998657227, 0.0047226881980895995, 0.004809728145599365, 0.004734975814819336, 0.004729856014251709, 0.0047288317680358885, 0.004720640182495117, 0.0047513599395751956, 0.0047257599830627444, 0.0047257599830627444, 0.0047554559707641605, 0.004770815849304199, 0.004811776161193848, 0.004766719818115234, 0.004771840095520019, 0.0047523841857910155, 0.004721695899963379, 0.004741119861602783, 0.004719583988189697, 0.004736000061035156, 0.004743167877197266, 0.004754432201385498, 0.004741119861602783, 0.0047257599830627444, 0.004740096092224121, 0.004794367790222168, 0.00479744005203247, 0.004794367790222168, 0.004754432201385498, 0.0047329277992248535, 0.004734975814819336, 0.004745215892791748, 0.004726784229278564, 0.004749311923980713, 0.004729856014251709, 0.004768767833709717, 0.0047185921669006346, 0.004748288154602051, 0.0047554559707641605, 0.004707327842712402, 0.004758528232574463, 0.004716544151306152, 0.0047513599395751956, 0.004737023830413818, 0.004736000061035156, 0.004737023830413818, 0.010554368019104005, 0.00481382417678833, 0.004773888111114502, 0.004794367790222168, 0.004745215892791748, 0.0047257599830627444, 0.0047513599395751956, 0.004711423873901367, 0.004790272235870361, 0.004781055927276612, 0.00481385612487793, 0.00480457592010498, 0.0047820801734924315, 0.004810815811157227, 0.004775872230529785, 0.004798463821411133, 0.00480460786819458, 0.004777984142303467, 0.00479744005203247, 0.0047964158058166504, 0.004791296005249023, 0.004801536083221435, 0.004767744064331054, 0.0047964158058166504, 0.004810751914978028, 0.004819968223571777, 0.0047964158058166504, 0.004820991992950439, 0.00479744005203247, 0.0047964158058166504, 0.004766719818115234, 0.004758528232574463, 0.004765727996826172, 0.004747231960296631, 0.004766719818115234, 0.004744192123413086, 0.004744192123413086, 0.004741119861602783, 0.004740096092224121, 0.004757503986358643, 0.004749375820159912, 0.004748223781585693, 0.0047257599830627444, 0.0047226881980895995, 0.004764671802520752, 0.004726784229278564, 0.004746240139007568, 0.0047523841857910155, 0.004727807998657227, 0.004941823959350586, 0.0048568320274353025, 0.004803584098815918, 0.004799488067626953, 0.004778016090393067, 0.004821983814239502, 0.004784192085266113, 0.004873151779174805, 0.004801536083221435, 0.004774911880493164, 0.0047933440208435055, 0.004768767833709717, 0.004800511837005615, 0.00479744005203247, 0.01080016040802002, 0.004760543823242188, 0.004765696048736572, 0.004789247989654541, 0.004773888111114502, 0.004829184055328369, 0.004776959896087647, 0.004773888111114502, 0.0048056640625, 0.004767712116241455, 0.004816895961761475, 0.004773888111114502, 0.004749311923980713, 0.0047861762046813965, 0.004771840095520019, 0.004783103942871094, 0.004775936126708984, 0.004767744064331054, 0.0048189439773559575, 0.004794367790222168, 0.0048230400085449215, 0.004741119861602783, 0.004727807998657227, 0.004768767833709717, 0.004964352130889893, 0.0050165758132934574, 0.004851712226867676, 0.004832287788391113, 0.004891615867614746, 0.004744192123413086, 0.004780032157897949, 0.004737023830413818, 0.0047227201461791995, 0.004772831916809082, 0.0047216639518737795, 0.004731904029846192, 0.004816895961761475, 0.004790272235870361, 0.004802559852600098, 0.004746240139007568, 0.004709375858306885, 0.004816895961761475, 0.004767744064331054, 0.004808703899383545, 0.004869120121002197, 0.004781055927276612, 0.004820032119750977, 0.004774847984313965, 0.0047933440208435055, 0.004787199974060059, 0.004761600017547608, 0.004833280086517334, 0.0047226881980895995, 0.00475648021697998, 0.0047636480331420894, 0.004783103942871094, 0.004784128189086914, 0.004749311923980713, 0.004759552001953125, 0.0047933440208435055, 0.004773888111114502, 0.004847616195678711, 0.004817920207977295, 0.010781696319580078, 0.004803584098815918, 0.004802559852600098, 0.0047861762046813965, 0.00486195182800293, 0.004974592208862305, 0.004905983924865722, 0.004863999843597412, 0.004783103942871094, 0.0048158721923828125, 0.004832255840301514, 0.004734975814819336, 0.004733952045440673, 0.0047288317680358885, 0.00475654411315918, 0.004816832065582276, 0.004737023830413818, 0.004748288154602051, 0.0047227201461791995, 0.0047441601753234865, 0.004759552001953125, 0.00499507188796997, 0.004841472148895264, 0.004881408214569092, 0.004811776161193848, 0.004746240139007568, 0.004726784229278564, 0.004727807998657227, 0.004711423873901367, 0.004734975814819336, 0.004790272235870361, 0.004838399887084961, 0.0048056321144104, 0.004788224220275879, 0.0047861762046813965, 0.004806655883789063, 0.00481382417678833, 0.00480460786819458, 0.004773888111114502, 0.004833280086517334, 0.004793407917022705, 0.004825024127960205, 0.004794367790222168, 0.004764671802520752, 0.004790272235870361, 0.004799488067626953, 0.0048220157623291016, 0.004819968223571777, 0.004776959896087647, 0.004787199974060059, 0.0047964158058166504, 0.004801536083221435, 0.0048148479461669925, 0.0047820801734924315, 0.004772863864898682, 0.004784128189086914, 0.004809728145599365, 0.004798463821411133, 0.004795392036437988, 0.004780032157897949, 0.00481382417678833, 0.004788224220275879, 0.0047257599830627444, 0.010705920219421386, 0.004827136039733886, 0.004776959896087647, 0.004798463821411133, 0.004800511837005615, 0.004968448162078858, 0.0048158721923828125, 0.004784128189086914, 0.004810751914978028, 0.00479744005203247, 0.004798463821411133, 0.004811776161193848, 0.004787199974060059, 0.004819007873535156, 0.004776927947998047, 0.00481276798248291, 0.0047964158058166504, 0.004820991992950439, 0.004802559852600098, 0.004800511837005615, 0.004803584098815918, 0.004811776161193848, 0.004783103942871094, 0.0047964158058166504, 0.0048056321144104, 0.004809728145599365, 0.004808703899383545, 0.004802559852600098, 0.004799488067626953, 0.0048261117935180665, 0.0048148479461669925, 0.004817920207977295, 0.004779007911682129, 0.004795392036437988, 0.00480460786819458, 0.0048148479461669925, 0.004803584098815918, 0.004790272235870361, 0.004931583881378174, 0.004800576210021973, 0.004804543972015381, 0.004801536083221435, 0.004768767833709717, 0.004812831878662109, 0.004852704048156738, 0.004860928058624267, 0.004840447902679444, 0.004803647994995117, 0.00478713607788086, 0.00480460786819458, 0.0048046398162841794, 0.004799456119537354, 0.004788224220275879, 0.0047923197746276855, 0.004806655883789063, 0.0048158721923828125, 0.004800511837005615, 0.004918272018432617, 0.004937727928161621, 0.004842495918273926, 0.004777984142303467, 0.004801536083221435, 0.004771840095520019, 0.010771455764770508, 0.0047606401443481446, 0.004822976112365722, 0.004775936126708984, 0.004850687980651855, 0.0048158721923828125, 0.004787199974060059, 0.0048261117935180665, 0.0047944321632385255, 0.004804543972015381, 0.004808703899383545, 0.004817920207977295, 0.0048384637832641604, 0.004774847984313965, 0.004802559852600098, 0.004829184055328369, 0.004787199974060059, 0.004801536083221435, 0.004750336170196533, 0.004761600017547608, 0.004746240139007568, 0.0047257599830627444, 0.004744192123413086, 0.004742144107818603, 0.004759552001953125, 0.004758528232574463, 0.004746240139007568, 0.005061632156372071, 0.004873216152191162, 0.004840447902679444, 0.004901887893676758, 0.005198847770690918, 0.005697535991668701, 0.004992000102996826, 0.004941823959350586, 0.004918272018432617, 0.004807680130004883, 0.00485478401184082, 0.004827136039733886, 0.004799488067626953, 0.0048455681800842285, 0.004806655883789063, 0.0048793601989746095, 0.004765696048736572, 0.004780032157897949, 0.004904960155487061, 0.004779007911682129, 0.00481279993057251, 0.004775936126708984, 0.004776959896087647, 0.0048189439773559575, 0.0047964158058166504, 0.004798463821411133, 0.004785215854644776, 0.004786111831665039, 0.004819968223571777, 0.0047861762046813965, 0.0047851519584655765, 0.004783103942871094, 0.004777984142303467, 0.004816895961761475, 0.0047933440208435055, 0.004809728145599365, 0.01083289623260498, 0.004846591949462891, 0.005007359981536865, 0.004957183837890625, 0.004921343803405762, 0.004831232070922851, 0.0048230400085449215, 0.0047933440208435055, 0.004902912139892578, 0.004860928058624267, 0.004859903812408447, 0.004910079956054687, 0.005074944019317627, 0.0049489917755126955, 0.004853759765625, 0.004790272235870361, 0.004834303855895996, 0.004836351871490479, 0.004810751914978028, 0.00481382417678833, 0.004774911880493164, 0.004810751914978028, 0.004825088024139404, 0.004832255840301514, 0.00481382417678833, 0.0048455681800842285, 0.004807680130004883, 0.004774911880493164, 0.004820991992950439, 0.0048148479461669925, 0.0048148479461669925, 0.004808735847473144, 0.004778975963592529, 0.004794367790222168, 0.004803679943084717, 0.004797344207763672, 0.004798463821411133, 0.004789247989654541, 0.00481279993057251, 0.004832255840301514, 0.0047820801734924315, 0.0047820801734924315, 0.0047964158058166504, 0.004768832206726074, 0.004876224040985108, 0.0048865280151367185, 0.004844543933868409, 0.004785183906555176, 0.0047738561630249025, 0.004827136039733886, 0.004768767833709717, 0.004802559852600098, 0.004806655883789063, 0.004769792079925537, 0.004765696048736572, 0.004727807998657227, 0.004743167877197266, 0.0047636480331420894, 0.004734975814819336, 0.004779007911682129, 0.004719615936279297, 0.004770815849304199, 0.004790272235870361, 0.010760191917419434, 0.0048261117935180665, 0.004832255840301514, 0.0048148479461669925, 0.004872191905975342, 0.004928512096405029, 0.004794367790222168, 0.0048189439773559575, 0.004780032157897949, 0.00480460786819458, 0.004891647815704346, 0.004922368049621582, 0.0049500160217285155, 0.004859903812408447, 0.004808703899383545, 0.004843520164489746, 0.004746240139007568, 0.0047554559707641605, 0.004759552001953125, 0.004729856014251709, 0.004758528232574463, 0.004727807998657227, 0.004746240139007568, 0.0047216639518737795, 0.004746240139007568, 0.0048056321144104, 0.004795392036437988, 0.004762656211853027, 0.004726751804351807, 0.0047554559707641605, 0.0047554559707641605, 0.004724736213684082, 0.004743167877197266, 0.0047606401443481446, 0.004812736034393311, 0.004776959896087647, 0.004787199974060059, 0.004810751914978028, 0.004808703899383545, 0.00511078405380249, 0.004893695831298828, 0.00479744005203247, 0.0048496642112731934, 0.004947968006134033, 0.004894720077514648, 0.0048046717643737795, 0.005298175811767578, 0.004834239959716797, 0.004881408214569092, 0.004820991992950439, 0.00487014389038086, 0.0048261117935180665, 0.005287936210632324, 0.005557248115539551, 0.004988927841186524, 0.005138432025909424, 0.004943871974945068, 0.004951039791107178, 0.004944896221160889, 0.004973567962646484, 0.005033984184265137, 0.0048230400085449215, 0.004842495918273926, 0.010894335746765137, 0.00481279993057251, 0.004798463821411133, 0.004811776161193848, 0.004784128189086914, 0.004788224220275879, 0.004791296005249023, 0.004769792079925537, 0.004827136039733886, 0.004811776161193848, 0.004800511837005615, 0.004899839878082276, 0.004775936126708984, 0.004878335952758789, 0.004784128189086914, 0.005044223785400391, 0.004817920207977295, 0.004784128189086914, 0.004834303855895996, 0.00479744005203247, 0.004790272235870361, 0.004820991992950439, 0.0047923197746276855, 0.004846591949462891, 0.004777984142303467, 0.0047820801734924315, 0.004809728145599365, 0.0047861762046813965, 0.004840447902679444, 0.00479750394821167, 0.004815807819366455, 0.004832255840301514, 0.004789247989654541, 0.0048158721923828125, 0.004787199974060059, 0.004791327953338623, 0.0048414402008056644, 0.00479744005203247, 0.0048527359962463375, 0.004806655883789063, 0.004788224220275879, 0.0048261117935180665, 0.004794367790222168, 0.004843520164489746, 0.004838399887084961, 0.0047933759689331055, 0.0048178877830505375, 0.004899839878082276, 0.0048148479461669925, 0.004809728145599365, 0.00479744005203247, 0.004828159809112549, 0.004809728145599365, 0.004803584098815918, 0.004788224220275879, 0.0048158721923828125, 0.0048220157623291016, 0.004809728145599365, 0.005634047985076904, 0.004869120121002197, 0.004803584098815918, 0.004841472148895264, 0.004791296005249023, 0.010775551795959473, 0.00481382417678833, 0.00480460786819458, 0.0047861762046813965, 0.004883456230163574, 0.004810751914978028, 0.004790272235870361, 0.004799488067626953, 0.004808703899383545, 0.00481382417678833, 0.0047933440208435055, 0.004794367790222168, 0.004807680130004883, 0.0047933440208435055, 0.004857855796813965, 0.004827136039733886, 0.004780032157897949, 0.004985856056213379, 0.004807680130004883, 0.004824063777923584, 0.004848639965057373, 0.004777023792266846, 0.004817855834960937, 0.004777984142303467, 0.004760575771331787, 0.004753407955169678, 0.004734975814819336, 0.004803616046905518, 0.004826079845428466, 0.00481382417678833, 0.004783103942871094, 0.004773888111114502, 0.0048148479461669925, 0.004833280086517334, 0.0048056321144104, 0.0048220157623291016, 0.004803584098815918, 0.004843520164489746, 0.004807680130004883, 0.004819968223571777, 0.00480460786819458, 0.004798463821411133, 0.004819968223571777, 0.004800511837005615, 0.004798463821411133, 0.004824063777923584, 0.0047861762046813965, 0.004817920207977295, 0.00479744005203247, 0.004799488067626953, 0.004966400146484375, 0.004911104202270508, 0.004978687763214112, 0.00494704008102417, 0.004903903961181641, 0.004818880081176758, 0.004819968223571777, 0.004868095874786377, 0.0048230400085449215, 0.004883456230163574, 0.004872191905975342, 0.004902912139892578, 0.004946944236755371, 0.011116607666015625, 0.004925375938415528, 0.004923391819000244, 0.004969471931457519, 0.004887551784515381, 0.00495411205291748, 0.0048230400085449215, 0.0049090561866760255, 0.004929535865783692, 0.004901887893676758, 0.004925439834594727, 0.004841472148895264, 0.004915200233459473, 0.0048558077812194825, 0.004897791862487793, 0.004921343803405762, 0.004867072105407715, 0.004865024089813232, 0.004876287937164306, 0.0048455681800842285, 0.0048558077812194825, 0.004798463821411133, 0.004936704158782959, 0.004868095874786377, 0.0048230400085449215, 0.004892672061920166, 0.0048189439773559575, 0.004820991992950439, 0.004863999843597412, 0.005096447944641113, 0.006543360233306885, 0.005668863773345947, 0.0056945281028747555, 0.00513427209854126, 0.005017600059509277, 0.004941823959350586, 0.0049162240028381345, 0.004930560111999512, 0.004859903812408447, 0.0048895998001098635, 0.0048230400085449215, 0.004808703899383545, 0.0047329277992248535, 0.004727807998657227, 0.004761600017547608, 0.004720640182495117, 0.0047329277992248535, 0.004765696048736572, 0.0047933440208435055, 0.004806655883789063, 0.004776959896087647, 0.004776959896087647, 0.004789247989654541, 0.00479744005203247, 0.004809728145599365, 0.004803616046905518, 0.004897759914398193, 0.004811776161193848, 0.004802559852600098, 0.0048220157623291016, 0.004783103942871094, 0.004768767833709717, 0.00479744005203247]",tokens/s,202.0580696250412,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,google/gemma-2b,google/gemma-2b,cuda,0,42,,,True,,,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.223-212.873.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.0,,,,1.21.4,,,,0.12.0,,,gemma,MB,1549.574144,6797.39392,0.0,6150.946816,5419.87328,s,10,0.3060660495758057,0.03060660495758057,7.193886942037245e-05,0.030592047691345214,0.03068368377685547,0.030694354248046875,0.030702890625,"[0.03044371223449707, 0.030681312561035155, 0.030584735870361326, 0.03056051254272461, 0.03066582489013672, 0.030581600189208986, 0.030588735580444337, 0.030595359802246095, 0.030659231185913086, 0.03070502471923828]",tokens/s,8364.207671997758,kWh,3.6120757368543264e-07,1.9792357128510668e-07,1.984262545614669e-06,2.543393690585208e-06,tokens/kWh,100652919.34458527,MB,1611.931648,6797.39392,0.0,6150.946816,5419.87584,s,10,13.032300781250001,1.3032300781250001,0.006094333936100913,1.3046716918945314,1.3112125854492187,1.311342498779297,1.3114464294433594,"[1.30115283203125, 1.3111837158203126, 1.3055372314453124, 1.311472412109375, 1.2925338134765625, 1.304407958984375, 1.306199462890625, 1.2931976318359375, 1.3049354248046876, 1.3016802978515625]",tokens/s,48.341425706380384,kWh,1.5460242567356417e-05,8.469340465277226e-06,5.28743091189857e-05,7.680389215161933e-05,tokens/kWh,820270.9294423657,,s,630,13.030602731704711,0.02068349639953129,0.00048116754412136625,0.020403712272644044,0.021182567024230957,0.021269554710388185,0.021722215194702153,"[0.020330495834350586, 0.020290559768676757, 0.020311040878295897, 0.02036735916137695, 0.020118528366088868, 0.020130815505981444, 0.020126720428466797, 0.020330495834350586, 0.021097471237182617, 0.020601856231689454, 0.020315135955810547, 0.020323328018188477, 0.02029363250732422, 0.020299776077270508, 0.020360191345214843, 0.02032640075683594, 0.02032537651062012, 0.020358144760131838, 0.020323328018188477, 0.02031001663208008, 0.020296703338623046, 0.020328447341918944, 0.02030080032348633, 0.020372480392456056, 0.020288511276245116, 0.02088140869140625, 0.02105958366394043, 0.021155839920043946, 0.021161983489990235, 0.021182464599609374, 0.02119987106323242, 0.020332544326782227, 0.020319232940673827, 0.020313087463378905, 0.020347904205322266, 0.020360191345214843, 0.02026393508911133, 0.020559871673583984, 0.021206016540527343, 0.021105663299560547, 0.021061632156372072, 0.021154815673828126, 0.020595712661743162, 0.021110784530639647, 0.021021696090698243, 0.02107494354248047, 0.02107904052734375, 0.021097471237182617, 0.02107494354248047, 0.0210882568359375, 0.021126144409179686, 0.020316160202026368, 0.020348928451538087, 0.020343807220458983, 0.02042470359802246, 0.020357120513916017, 0.020307968139648438, 0.02106060791015625, 0.021754880905151368, 0.0212992000579834, 0.021121023178100586, 0.02109951972961426, 0.02109132766723633, 0.020372480392456056, 0.020883455276489257, 0.021132287979125978, 0.021132287979125978, 0.021061632156372072, 0.021203968048095705, 0.021164031982421876, 0.021124095916748048, 0.02106879997253418, 0.021146623611450196, 0.021130239486694336, 0.02041548728942871, 0.02032537651062012, 0.020338687896728515, 0.020395008087158203, 0.020386816024780274, 0.020342784881591795, 0.02039091110229492, 0.020393983840942383, 0.020412416458129884, 0.020354047775268554, 0.020587520599365236, 0.02112512016296387, 0.021130239486694336, 0.02112512016296387, 0.021112831115722656, 0.021109760284423826, 0.021130239486694336, 0.020343807220458983, 0.020303871154785155, 0.020319232940673827, 0.020610048294067384, 0.021081087112426757, 0.021114879608154297, 0.02038374328613281, 0.020335615158081053, 0.020249599456787108, 0.020280319213867186, 0.020280319213867186, 0.02029465675354004, 0.020257791519165038, 0.02146918487548828, 0.0211015682220459, 0.02123980712890625, 0.021445632934570313, 0.021190656661987304, 0.021140480041503908, 0.021141504287719725, 0.02101043128967285, 0.02110771179199219, 0.02108006477355957, 0.02103398323059082, 0.021133312225341795, 0.021164031982421876, 0.020999168395996092, 0.02106777572631836, 0.021078016281127928, 0.020314111709594726, 0.02033459281921387, 0.02040729522705078, 0.02104422378540039, 0.02105855941772461, 0.02109951972961426, 0.02041651153564453, 0.02033459281921387, 0.02031820869445801, 0.020377599716186523, 0.020319232940673827, 0.02086195182800293, 0.021124095916748048, 0.02108415985107422, 0.020344831466674804, 0.020380672454833985, 0.02109644889831543, 0.021104639053344726, 0.021185535430908203, 0.021121023178100586, 0.020345855712890625, 0.02043289566040039, 0.021142528533935546, 0.02128281593322754, 0.021217279434204102, 0.021517311096191406, 0.021341184616088867, 0.021416959762573243, 0.021200895309448242, 0.02125312042236328, 0.021097471237182617, 0.021440511703491212, 0.021219327926635743, 0.021127168655395507, 0.02051481628417969, 0.020726783752441406, 0.020497407913208008, 0.020411392211914063, 0.020344831466674804, 0.02039910316467285, 0.020321279525756835, 0.020333568572998048, 0.02030182456970215, 0.020365312576293947, 0.020330495834350586, 0.020350976943969725, 0.020343807220458983, 0.020303871154785155, 0.020338687896728515, 0.020350976943969725, 0.020338687896728515, 0.020373504638671876, 0.020337663650512695, 0.020349952697753908, 0.020345855712890625, 0.020295679092407228, 0.02028646469116211, 0.020340736389160157, 0.020338687896728515, 0.020728832244873048, 0.021178367614746094, 0.021126144409179686, 0.02101862335205078, 0.021170175552368165, 0.021139455795288087, 0.02109542465209961, 0.02108518409729004, 0.021110784530639647, 0.020358144760131838, 0.02042780876159668, 0.020455392837524414, 0.020410367965698242, 0.02038582420349121, 0.02031100845336914, 0.020331520080566406, 0.02025164794921875, 0.020165632247924805, 0.020323328018188477, 0.020246528625488282, 0.020196352005004883, 0.020338687896728515, 0.020361215591430663, 0.020668415069580077, 0.020329471588134765, 0.020709375381469726, 0.02043801689147949, 0.020360191345214843, 0.020370431900024414, 0.020350976943969725, 0.020393983840942383, 0.022389759063720704, 0.021610496520996093, 0.02126233673095703, 0.021219327926635743, 0.021158912658691405, 0.021193727493286133, 0.021152767181396484, 0.02105241584777832, 0.021149696350097655, 0.021161983489990235, 0.02109235191345215, 0.02304204750061035, 0.021432319641113282, 0.021156864166259767, 0.02111692810058594, 0.021117952346801756, 0.021120000839233398, 0.021061632156372072, 0.021090303421020508, 0.021094400405883788, 0.02106368064880371, 0.021169151306152344, 0.021131263732910157, 0.021153791427612305, 0.021127168655395507, 0.021147647857666017, 0.021179391860961915, 0.021156864166259767, 0.021180416107177736, 0.02105753517150879, 0.020346879959106445, 0.020274175643920898, 0.02038374328613281, 0.020321279525756835, 0.020320255279541014, 0.02064588737487793, 0.021328895568847657, 0.020358144760131838, 0.020349952697753908, 0.020378623962402344, 0.020365312576293947, 0.020358144760131838, 0.02126335906982422, 0.020435968399047853, 0.020330495834350586, 0.020361215591430663, 0.02030899238586426, 0.020385791778564453, 0.020328447341918944, 0.020296703338623046, 0.020321279525756835, 0.020323328018188477, 0.020307968139648438, 0.020333568572998048, 0.02030284881591797, 0.02031718444824219, 0.021143552780151367, 0.02032537651062012, 0.020335615158081053, 0.020331520080566406, 0.020265983581542968, 0.020345855712890625, 0.020327423095703127, 0.02108415985107422, 0.021089279174804687, 0.020380672454833985, 0.020341760635375978, 0.020327423095703127, 0.02047283172607422, 0.02025984001159668, 0.020358144760131838, 0.020397056579589845, 0.020353023529052734, 0.02037555122375488, 0.020360191345214843, 0.020389888763427736, 0.02044211196899414, 0.020384767532348632, 0.02042265510559082, 0.020357120513916017, 0.020379648208618165, 0.020410367965698242, 0.020512767791748047, 0.02042265510559082, 0.020386816024780274, 0.020376575469970702, 0.020404224395751954, 0.02053222465515137, 0.020406272888183592, 0.020366336822509764, 0.02039091110229492, 0.020328447341918944, 0.02068992042541504, 0.020674560546875, 0.02043903923034668, 0.020742143630981445, 0.02129408073425293, 0.02111692810058594, 0.021149696350097655, 0.021183488845825195, 0.02145280075073242, 0.021147647857666017, 0.020384767532348632, 0.020370431900024414, 0.020313087463378905, 0.020563968658447264, 0.020305919647216796, 0.020304895401000975, 0.020303871154785155, 0.020322303771972656, 0.021081087112426757, 0.021082111358642578, 0.020711423873901368, 0.020288511276245116, 0.020419584274291993, 0.02031820869445801, 0.020336639404296874, 0.020362239837646484, 0.020669439315795898, 0.020372480392456056, 0.020387840270996094, 0.021169151306152344, 0.021143552780151367, 0.021123071670532227, 0.02109542465209961, 0.02106265640258789, 0.02107187271118164, 0.021089279174804687, 0.021149696350097655, 0.021129215240478515, 0.021809152603149414, 0.021217279434204102, 0.021016576766967773, 0.02110361671447754, 0.02105753517150879, 0.020320255279541014, 0.02026803207397461, 0.020315135955810547, 0.020297727584838866, 0.02041651153564453, 0.020345855712890625, 0.021425151824951173, 0.021526527404785157, 0.021311487197875977, 0.021061632156372072, 0.021227519989013673, 0.020295679092407228, 0.020274175643920898, 0.020247552871704103, 0.02028441619873047, 0.020297727584838866, 0.020291584014892578, 0.020311040878295897, 0.020298751831054687, 0.020297727584838866, 0.020320255279541014, 0.020271104812622072, 0.020311040878295897, 0.020980735778808594, 0.02089574432373047, 0.0210565128326416, 0.021113855361938477, 0.021004287719726563, 0.02106879997253418, 0.021117952346801756, 0.020321279525756835, 0.020304895401000975, 0.02029465675354004, 0.020379648208618165, 0.020347904205322266, 0.020321279525756835, 0.020419584274291993, 0.020297727584838866, 0.02044927978515625, 0.02187775993347168, 0.021246976852416992, 0.021211135864257814, 0.02125823974609375, 0.021174272537231444, 0.02108723258972168, 0.02107699203491211, 0.021146623611450196, 0.021188608169555666, 0.021185535430908203, 0.021135360717773437, 0.02107904052734375, 0.020332544326782227, 0.02028544044494629, 0.020346879959106445, 0.02066022491455078, 0.020563968658447264, 0.020314111709594726, 0.02031820869445801, 0.020313087463378905, 0.02082611274719238, 0.021144575119018554, 0.021072895050048827, 0.021053440093994142, 0.020321279525756835, 0.020322303771972656, 0.020312063217163084, 0.020303871154785155, 0.020305919647216796, 0.020337663650512695, 0.02030284881591797, 0.021157888412475585, 0.021171199798583985, 0.020792320251464845, 0.021823488235473632, 0.021296127319335938, 0.021113855361938477, 0.020388864517211915, 0.02032748794555664, 0.02033964729309082, 0.02028544044494629, 0.020602880477905275, 0.021136383056640624, 0.021111808776855468, 0.02105855941772461, 0.02107187271118164, 0.0210831356048584, 0.020299776077270508, 0.020348928451538087, 0.020299776077270508, 0.02027212715148926, 0.02149990463256836, 0.021111808776855468, 0.02028339195251465, 0.020290559768676757, 0.020297727584838866, 0.020246528625488282, 0.020295679092407228, 0.02025267219543457, 0.020305919647216796, 0.020288511276245116, 0.020281343460083007, 0.020298751831054687, 0.020858879089355468, 0.021642240524291992, 0.021465087890625, 0.0212807674407959, 0.021122047424316406, 0.02104012870788574, 0.021195775985717775, 0.02106572723388672, 0.02105548858642578, 0.020327423095703127, 0.02031820869445801, 0.020339712142944336, 0.020280319213867186, 0.020362239837646484, 0.020337663650512695, 0.021179391860961915, 0.020389888763427736, 0.020377599716186523, 0.02028339195251465, 0.020329471588134765, 0.02041651153564453, 0.020419584274291993, 0.020404224395751954, 0.020349952697753908, 0.02040934371948242, 0.020411392211914063, 0.020159488677978517, 0.02019430351257324, 0.020184064865112306, 0.020379648208618165, 0.020346879959106445, 0.02036735916137695, 0.020346879959106445, 0.02031718444824219, 0.02036735916137695, 0.020356096267700196, 0.020350976943969725, 0.020355072021484375, 0.02040012741088867, 0.020346879959106445, 0.020387840270996094, 0.020355072021484375, 0.020348928451538087, 0.020323328018188477, 0.020328447341918944, 0.020345855712890625, 0.020350976943969725, 0.020356096267700196, 0.020305919647216796, 0.020328447341918944, 0.020287488937377928, 0.02041548728942871, 0.020775936126708985, 0.021081087112426757, 0.021081087112426757, 0.021073919296264648, 0.021020671844482423, 0.020385791778564453, 0.020331520080566406, 0.020271104812622072, 0.020934656143188478, 0.021130239486694336, 0.02104934310913086, 0.02041753578186035, 0.020370431900024414, 0.020805631637573242, 0.021126144409179686, 0.021113855361938477, 0.02110054397583008, 0.02109132766723633, 0.021098495483398438, 0.021090303421020508, 0.0210882568359375, 0.021139455795288087, 0.021144575119018554, 0.021113855361938477, 0.02030284881591797, 0.020329471588134765, 0.020335615158081053, 0.02048204803466797, 0.020319232940673827, 0.020295679092407228, 0.021619712829589844, 0.021183488845825195, 0.021304319381713867, 0.02112512016296387, 0.021123071670532227, 0.020939775466918945, 0.021191680908203125, 0.021188608169555666, 0.021134336471557616, 0.021200895309448242, 0.021178367614746094, 0.021104639053344726, 0.021102592468261717, 0.021129215240478515, 0.021113855361938477, 0.02027519989013672, 0.020298751831054687, 0.02045747184753418, 0.020748287200927733, 0.020403200149536133, 0.020360191345214843, 0.020353023529052734, 0.02038374328613281, 0.02030182456970215, 0.020319232940673827, 0.020327423095703127, 0.020290559768676757, 0.02026905632019043, 0.020553728103637696, 0.020273151397705077, 0.02030080032348633, 0.020271104812622072, 0.020404224395751954, 0.02042982482910156, 0.020288511276245116, 0.020315135955810547, 0.020343807220458983, 0.0202926082611084, 0.020378623962402344, 0.020344831466674804, 0.020313087463378905, 0.020381696701049806, 0.020331520080566406, 0.02032640075683594, 0.020358144760131838, 0.02030182456970215, 0.02067148780822754, 0.021142528533935546, 0.021190656661987304, 0.021186559677124024, 0.021167104721069335, 0.021158912658691405, 0.020355072021484375, 0.020299776077270508, 0.02031718444824219, 0.020265983581542968, 0.020374528884887694, 0.020335615158081053, 0.020329471588134765, 0.020322303771972656, 0.02025574493408203, 0.020350976943969725, 0.020340736389160157, 0.02028441619873047, 0.020346879959106445, 0.02030899238586426, 0.020389888763427736, 0.020707328796386718, 0.02043903923034668, 0.020805631637573242, 0.02109542465209961, 0.021194751739501954, 0.021136383056640624, 0.020372480392456056, 0.020323328018188477, 0.020373504638671876, 0.02070425605773926, 0.020323328018188477, 0.020237312316894532, 0.02030284881591797, 0.020265983581542968, 0.020262912750244142, 0.02012774467468262, 0.020082687377929686, 0.020262912750244142, 0.020402175903320312, 0.020405248641967775, 0.02025062370300293, 0.020303871154785155, 0.020273151397705077, 0.020463615417480468, 0.026468351364135743, 0.021398527145385742, 0.02121625518798828, 0.02127462387084961, 0.02107187271118164, 0.021165056228637694, 0.02106777572631836, 0.021111808776855468, 0.021181440353393553, 0.020305919647216796]",tokens/s,48.34772519517837,,,,,,MB,933.96992,6768.033792,0.0,6138.363904,6060.931072,s,1,6.89289501953125,6.89289501953125,0.0,6.89289501953125,6.89289501953125,6.89289501953125,6.89289501953125,[6.89289501953125],,kWh,1.619802311103058e-06,8.71413727314462e-07,4.542225856003779e-06,7.033441894421299e-06, float16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,s,s,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/s/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66947363-08e5e89f68e1b8012d1b76bd;662573c1-9152-46c5-a4d9-5143f0981568) Repository Not Found for url: https://huggingface.co/s/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: s is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-40b,tiiuae/falcon-40b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ self.load_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 250, in load_model_with_no_weights self.pretrained_model = self.automodel_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 559, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3704, in from_pretrained config = cls._autoset_attn_implementation( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1490, in _autoset_attn_implementation config = cls._check_and_enable_sdpa( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1656, in _check_and_enable_sdpa raise ValueError( ValueError: FalconForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,google/gemma-7b,google/gemma-7b,cuda,0,42,,,True,,,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.223-212.873.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.0,,,,1.21.4,,,,0.12.0,,,gemma,MB,1534.046208,19300.614144,0.0,18654.16704,17596.665856,s,10,0.9373033828735352,0.09373033828735353,9.173135946059763e-05,0.09374342346191407,0.09385008773803712,0.09385560493469239,0.0938600186920166,"[0.09356540679931641, 0.0937131805419922, 0.09384886169433594, 0.09375711822509766, 0.09372972869873047, 0.0936060791015625, 0.093771484375, 0.09366358184814454, 0.0937868194580078, 0.09386112213134766]",tokens/s,2731.23947568789,kWh,1.1077207424259952e-06,6.069757903606582e-07,6.454976607383203e-06,8.169673140169857e-06,tokens/kWh,31335402.97239817,MB,1587.408896,19302.711296,0.0,18656.264192,17596.668416,s,10,27.46353515625,2.746353515625,0.0009310654054011403,2.74612890625,2.7475124755859373,2.7477498901367188,2.7479398217773436,"[2.7479873046875, 2.74597119140625, 2.74626025390625, 2.747459716796875, 2.744705810546875, 2.745867919921875, 2.7473837890625, 2.746109619140625, 2.745641357421875, 2.746148193359375]",tokens/s,22.93950856711278,kWh,3.242917121111551e-05,1.777257424103388e-05,0.00014130841938101796,0.00019151016483316734,tokens/kWh,328964.2617919628,,s,630,27.46173646545408,0.04359005788167318,0.0001391294541353059,0.04356355094909668,0.04368506889343262,0.04379427890777588,0.04421626968383789,"[0.044229633331298826, 0.04356710433959961, 0.043480064392089846, 0.04348211288452149, 0.04346777725219726, 0.04345446395874023, 0.04347187042236328, 0.0434442253112793, 0.043469825744628904, 0.0434524154663086, 0.04369715118408203, 0.04356915283203125, 0.04357324981689453, 0.043514881134033206, 0.04350361633300781, 0.04351078414916992, 0.04396851348876953, 0.04350054550170898, 0.043514881134033206, 0.043529216766357424, 0.04356915283203125, 0.043512832641601565, 0.043551742553710936, 0.04351795196533203, 0.043671550750732424, 0.04352102279663086, 0.04358553695678711, 0.04353638458251953, 0.04362137603759766, 0.04353126525878906, 0.04356710433959961, 0.043551742553710936, 0.043641857147216793, 0.04349542236328125, 0.044349441528320314, 0.04362547302246094, 0.043617279052734374, 0.0436234245300293, 0.04381388854980469, 0.043679744720458984, 0.04363468933105469, 0.04357529449462891, 0.04360908889770508, 0.04363468933105469, 0.043641857147216793, 0.04362444686889649, 0.04361318588256836, 0.04368281555175781, 0.04361318588256836, 0.04358860778808594, 0.04365107345581055, 0.04360806274414063, 0.04368998336791992, 0.04361830520629883, 0.043641857147216793, 0.04358758544921875, 0.043686912536621096, 0.04362137603759766, 0.04388147354125976, 0.04363776016235352, 0.04370431900024414, 0.04360704040527344, 0.043668479919433595, 0.043646015167236325, 0.04347180938720703, 0.04346060943603516, 0.04345651245117187, 0.04346777725219726, 0.04363673782348633, 0.04345548629760742, 0.04344934463500977, 0.043450366973876955, 0.04345139312744141, 0.043437057495117185, 0.04347596740722656, 0.04357734298706055, 0.04372275161743164, 0.04347903823852539, 0.04348723220825195, 0.04352102279663086, 0.04357734298706055, 0.04350156784057617, 0.04349235153198242, 0.04354764938354492, 0.043499519348144534, 0.04350361633300781, 0.04348723220825195, 0.043524097442626954, 0.043507713317871094, 0.04355379104614258, 0.04349235153198242, 0.04355686569213867, 0.043551742553710936, 0.0435599365234375, 0.04352204895019531, 0.043524097442626954, 0.04354662322998047, 0.04354048156738281, 0.043535358428955076, 0.04356095886230469, 0.043561985015869144, 0.0436234245300293, 0.04356095886230469, 0.044183551788330076, 0.043632640838623046, 0.04371148681640625, 0.04354048156738281, 0.043594753265380856, 0.043568126678466795, 0.043561985015869144, 0.04358041763305664, 0.04366745758056641, 0.043684864044189455, 0.043703296661376956, 0.04405759811401367, 0.04363776016235352, 0.0439552001953125, 0.043643905639648435, 0.04360704040527344, 0.043649024963378906, 0.04359884643554687, 0.04367257690429688, 0.043614208221435545, 0.04362854385375976, 0.043633663177490234, 0.043668479919433595, 0.04408729553222656, 0.043407360076904294, 0.04341964721679688, 0.04343296051025391, 0.04349747085571289, 0.04348928070068359, 0.04349030303955078, 0.0434595832824707, 0.04356095886230469, 0.04350156784057617, 0.04349542236328125, 0.04343910217285156, 0.04353945541381836, 0.04348825454711914, 0.04354048156738281, 0.04346572875976563, 0.04352614212036133, 0.043619327545166016, 0.04353843307495117, 0.04349030303955078, 0.04351078414916992, 0.04351078414916992, 0.04368076705932617, 0.04364595031738281, 0.04353126525878906, 0.04352204895019531, 0.043566078186035154, 0.043518974304199216, 0.04357222366333008, 0.04356095886230469, 0.043600894927978515, 0.04354457473754883, 0.04369919967651367, 0.04357632064819336, 0.04358553695678711, 0.04352204895019531, 0.043579391479492184, 0.043515903472900394, 0.04355276870727539, 0.043561985015869144, 0.043595775604248044, 0.043584510803222655, 0.04361011123657227, 0.04355583953857422, 0.04361523056030273, 0.04355583953857422, 0.04416204833984375, 0.04362444686889649, 0.04360192108154297, 0.04361523056030273, 0.04362547302246094, 0.04361830520629883, 0.043993087768554685, 0.04363776016235352, 0.04372172927856445, 0.04359065628051758, 0.043650047302246094, 0.04355583953857422, 0.04361830520629883, 0.043584510803222655, 0.043643905639648435, 0.043663360595703124, 0.043807743072509765, 0.04372377777099609, 0.043440128326416014, 0.043463680267333986, 0.04342988967895508, 0.04357017517089844, 0.04363468933105469, 0.04349849700927735, 0.04348928070068359, 0.043483135223388675, 0.0435333137512207, 0.04352819061279297, 0.04356099319458008, 0.04400944137573242, 0.04362854385375976, 0.04354355239868164, 0.04348518371582031, 0.04352716827392578, 0.04352716827392578, 0.043535358428955076, 0.04347187042236328, 0.04352204895019531, 0.043507713317871094, 0.043548671722412106, 0.043493377685546876, 0.04355379104614258, 0.0435599365234375, 0.04355276870727539, 0.043600894927978515, 0.04357017517089844, 0.04362547302246094, 0.04357734298706055, 0.043548671722412106, 0.043581439971923826, 0.043652095794677735, 0.04371353530883789, 0.04360704040527344, 0.043622398376464845, 0.04355276870727539, 0.04354457473754883, 0.043535358428955076, 0.04355072021484375, 0.04359167861938477, 0.043589630126953126, 0.04357017517089844, 0.04360908889770508, 0.0436121597290039, 0.043606014251708985, 0.0436121597290039, 0.043622398376464845, 0.0435968017578125, 0.04412211227416992, 0.04359372711181641, 0.04415078353881836, 0.04370943832397461, 0.043757568359375, 0.043684864044189455, 0.04364287948608398, 0.043641857147216793, 0.04381081771850586, 0.04366643142700195, 0.04365619277954102, 0.04363980865478516, 0.04369510269165039, 0.04369305419921875, 0.04349542236328125, 0.04343603134155274, 0.04340224075317383, 0.0434442253112793, 0.04342476654052734, 0.04347084808349609, 0.04348518371582031, 0.043499519348144534, 0.043466751098632815, 0.04351391983032227, 0.04355987167358399, 0.0434595832824707, 0.043480064392089846, 0.043494400024414064, 0.04348723220825195, 0.04351692962646484, 0.04363161468505859, 0.043483135223388675, 0.04353228759765625, 0.04350668716430664, 0.043480064392089846, 0.043529216766357424, 0.04352000045776367, 0.04351385498046875, 0.04351078414916992, 0.04352102279663086, 0.04350361633300781, 0.04352819061279297, 0.04351078414916992, 0.04354764938354492, 0.0435865592956543, 0.043565055847167966, 0.04352102279663086, 0.04359884643554687, 0.04361523056030273, 0.04354662322998047, 0.043524097442626954, 0.04355379104614258, 0.043652095794677735, 0.04357427215576172, 0.04356095886230469, 0.043564094543457034, 0.04359468841552734, 0.0435599365234375, 0.04357222366333008, 0.04355891036987305, 0.043565055847167966, 0.043617279052734374, 0.0435968017578125, 0.043641857147216793, 0.043597824096679685, 0.04393881607055664, 0.0436234245300293, 0.043589630126953126, 0.04358041763305664, 0.04359987258911133, 0.04357734298706055, 0.04423680114746094, 0.04367769622802734, 0.043635711669921876, 0.04361523056030273, 0.04363776016235352, 0.04369715118408203, 0.043425792694091796, 0.04345446395874023, 0.0434411506652832, 0.04354662322998047, 0.04349030303955078, 0.04348620986938476, 0.043437057495117185, 0.0435599365234375, 0.04344319915771484, 0.043483135223388675, 0.04348928070068359, 0.043504638671875, 0.04347187042236328, 0.043514881134033206, 0.04345651245117187, 0.043507713317871094, 0.04400537490844727, 0.04359884643554687, 0.043499519348144534, 0.043606014251708985, 0.04349849700927735, 0.04353228759765625, 0.04347903823852539, 0.0435333137512207, 0.04347187042236328, 0.04354355239868164, 0.04350566482543945, 0.04356710433959961, 0.043548671722412106, 0.043687934875488284, 0.04350566482543945, 0.043578369140625, 0.04353126525878906, 0.043551742553710936, 0.043512832641601565, 0.043579391479492184, 0.043515903472900394, 0.0436049919128418, 0.043514881134033206, 0.04365926361083984, 0.043757568359375, 0.04373299026489258, 0.04360806274414063, 0.04361011123657227, 0.04355583953857422, 0.043597824096679685, 0.04356915283203125, 0.04375142288208008, 0.04362137603759766, 0.04364492797851562, 0.043638782501220705, 0.04366233444213867, 0.04362547302246094, 0.04423987197875977, 0.04359987258911133, 0.043632640838623046, 0.04359987258911133, 0.04363468933105469, 0.043594753265380856, 0.043684864044189455, 0.0436121597290039, 0.0436756477355957, 0.044298240661621094, 0.04348518371582031, 0.0434524154663086, 0.04342476654052734, 0.0434595832824707, 0.04350054550170898, 0.04346060943603516, 0.043476993560791016, 0.043494400024414064, 0.043548671722412106, 0.043535358428955076, 0.0435230712890625, 0.04356095886230469, 0.043529216766357424, 0.04349747085571289, 0.04345446395874023, 0.04353433609008789, 0.043466751098632815, 0.04350156784057617, 0.04350361633300781, 0.04354662322998047, 0.04354048156738281, 0.043529216766357424, 0.044042240142822264, 0.04359884643554687, 0.04362137603759766, 0.04353126525878906, 0.04354355239868164, 0.0435333137512207, 0.04354764938354492, 0.043641857147216793, 0.04357120132446289, 0.04358860778808594, 0.04360192108154297, 0.04356915283203125, 0.04352102279663086, 0.04363161468505859, 0.04357632064819336, 0.04358553695678711, 0.04359065628051758, 0.043602943420410156, 0.04360192108154297, 0.04359372711181641, 0.04360704040527344, 0.043643905639648435, 0.0436049919128418, 0.043878398895263675, 0.04398899078369141, 0.04366438293457031, 0.04361625671386719, 0.043614208221435545, 0.04367462539672851, 0.04369510269165039, 0.04395827102661133, 0.043698177337646485, 0.043660289764404295, 0.043650047302246094, 0.04363468933105469, 0.04364799880981445, 0.04361318588256836, 0.04365107345581055, 0.04364492797851562, 0.04362956619262695, 0.043638782501220705, 0.043453441619873044, 0.04357529449462891, 0.04350259017944336, 0.043469825744628904, 0.04469452667236328, 0.04346572875976563, 0.04354662322998047, 0.04348416137695312, 0.043478015899658204, 0.04346777725219726, 0.04344934463500977, 0.043480064392089846, 0.04379033660888672, 0.04353945541381836, 0.04350361633300781, 0.043518974304199216, 0.04350566482543945, 0.04365619277954102, 0.043641857147216793, 0.04353638458251953, 0.043507713317871094, 0.04349747085571289, 0.043480064392089846, 0.043514881134033206, 0.0435333137512207, 0.04353740692138672, 0.04362444686889649, 0.04371865463256836, 0.04377395248413086, 0.04354969787597656, 0.04354048156738281, 0.04354048156738281, 0.043535358428955076, 0.04354355239868164, 0.043548671722412106, 0.043524097442626954, 0.04354355239868164, 0.043512832641601565, 0.043499519348144534, 0.04353945541381836, 0.04353638458251953, 0.04357427215576172, 0.043565055847167966, 0.04357120132446289, 0.043565055847167966, 0.04359167861938477, 0.043551742553710936, 0.04357120132446289, 0.04355686569213867, 0.04363980865478516, 0.043643905639648435, 0.04369305419921875, 0.043769855499267575, 0.04379750442504883, 0.04360396957397461, 0.04361318588256836, 0.04360908889770508, 0.04360806274414063, 0.0435865592956543, 0.04360806274414063, 0.043606014251708985, 0.04360806274414063, 0.04368384170532227, 0.04342988967895508, 0.04343603134155274, 0.04342988967895508, 0.043496448516845705, 0.04349747085571289, 0.04350156784057617, 0.0434524154663086, 0.04354252624511719, 0.04343807983398437, 0.04350566482543945, 0.0444846076965332, 0.04354662322998047, 0.043453441619873044, 0.04354048156738281, 0.043491329193115234, 0.0435333137512207, 0.04351180648803711, 0.04349849700927735, 0.043483135223388675, 0.04350361633300781, 0.04346572875976563, 0.04352204895019531, 0.04348928070068359, 0.04353126525878906, 0.043504638671875, 0.04353228759765625, 0.04354252624511719, 0.04353945541381836, 0.04352102279663086, 0.04357017517089844, 0.043529216766357424, 0.0435599365234375, 0.043535358428955076, 0.04377395248413086, 0.04371148681640625, 0.04358860778808594, 0.04351078414916992, 0.043561985015869144, 0.04352511978149414, 0.04358246231079101, 0.04358041763305664, 0.04362649536132813, 0.04356300735473633, 0.043617279052734374, 0.04355583953857422, 0.0435865592956543, 0.04353433609008789, 0.04361011123657227, 0.04357017517089844, 0.04362444686889649, 0.04377702331542969, 0.04385996627807617, 0.043714561462402345, 0.04367257690429688, 0.04360908889770508, 0.043614208221435545, 0.04356915283203125, 0.043660289764404295, 0.043679744720458984, 0.04365311813354492, 0.04357120132446289, 0.043657215118408206, 0.043660289764404295, 0.04357222366333008, 0.04344319915771484, 0.043428863525390625, 0.04347289657592773, 0.04349542236328125, 0.0434411506652832, 0.043466751098632815, 0.04350668716430664, 0.043512832641601565, 0.043478015899658204, 0.04345446395874023, 0.04354969787597656, 0.04350668716430664, 0.04349849700927735, 0.04346265411376953, 0.04349747085571289, 0.044165119171142575, 0.04358553695678711, 0.04349235153198242, 0.0435230712890625, 0.0435333137512207, 0.04353433609008789, 0.043496448516845705, 0.04354355239868164, 0.04355276870727539, 0.04353638458251953, 0.043561985015869144, 0.043545600891113284, 0.043548671722412106, 0.04356710433959961, 0.04353945541381836, 0.04356300735473633, 0.04355276870727539, 0.043681793212890625, 0.04359167861938477, 0.04357222366333008, 0.04354764938354492, 0.043551742553710936, 0.04357427215576172, 0.043600894927978515, 0.04409446334838867, 0.04362649536132813, 0.04360192108154297, 0.04359987258911133, 0.0435968017578125, 0.04359270477294922, 0.043617279052734374, 0.0436234245300293, 0.04359270477294922, 0.043632640838623046, 0.04361625671386719, 0.04398489761352539, 0.043663360595703124, 0.04363673782348633, 0.043635711669921876, 0.04362956619262695, 0.0436234245300293, 0.043640830993652346, 0.04360704040527344, 0.04365107345581055, 0.043622398376464845, 0.043673599243164066]",tokens/s,22.941011060699594,,,,,,MB,935.088128,19279.642624,0.0,18649.972736,18648.241152,s,1,6.940919921875,6.940919921875,0.0,6.940919921875,6.940919921875,6.940919921875,6.940919921875,[6.940919921875],,kWh,2.297889397916227e-06,1.2431957099768545e-06,4.692781532006363e-06,8.233866639899446e-06, float16-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-125m,facebook/opt-125m,cuda,0,42,,,True,,,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",opt,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,databricks/dbrx-base,databricks/dbrx-base,cuda,0,42,,,True,,,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",dbrx,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-350m,facebook/opt-350m,cuda,0,42,,,True,,,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",opt,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,-,-,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 106, in _inner_fn validate_repo_id(arg_value) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 160, in validate_repo_id raise HFValidationError( huggingface_hub.errors.HFValidationError: Repo id must use alphanumeric chars or '-', '_', '.', '--' and '..' are forbidden, '-' and '.' cannot start or end the name, max length is 96: '-'. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 466, in cached_file raise EnvironmentError( OSError: Incorrect path_or_model_id: '-'. Please provide either the path to a local folder or the repo_id of a model on the Hub. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,stabilityai/stablelm-2-12b,stabilityai/stablelm-2-12b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ self.load_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 250, in load_model_with_no_weights self.pretrained_model = self.automodel_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3710, in from_pretrained model = cls(config, *model_args, **model_kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/stablelm/modeling_stablelm.py"", line 1188, in __init__ self.model = StableLmModel(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/stablelm/modeling_stablelm.py"", line 966, in __init__ [StableLmDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/stablelm/modeling_stablelm.py"", line 966, in [StableLmDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/stablelm/modeling_stablelm.py"", line 748, in __init__ self.mlp = StableLmMLP(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/stablelm/modeling_stablelm.py"", line 206, in __init__ self.down_proj = nn.Linear(self.intermediate_size, self.hidden_size, bias=False) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/linear.py"", line 98, in __init__ self.weight = Parameter(torch.empty((out_features, in_features), **factory_kwargs)) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_device.py"", line 78, in __torch_function__ return func(*args, **kwargs) torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 136.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,v,v,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/v/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6693256e-61513b6f4f18904048764bfa;5668ed51-fc89-4fe5-a8ae-8d2f4917e0ad) Repository Not Found for url: https://huggingface.co/v/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: v is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,openai-community/gpt2,openai-community/gpt2,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.42.3,,0.31.0,,,,1.20.0,,,,0.11.1,,,,MB,1204.30592,1032.323072,0.0,385.875968,319.013888,s,24,0.17006591987609862,0.007086079994837444,0.00021354295745949786,0.006997040033340454,0.007281977653503418,0.007317865538597107,0.007787319173812866,"[0.007926688194274903, 0.007320735931396484, 0.006998047828674316, 0.0071829757690429685, 0.006926943778991699, 0.007201151847839355, 0.006962495803833008, 0.007301599979400635, 0.006996032238006592, 0.007236192226409912, 0.006959616184234619, 0.007013951778411865, 0.006959968090057373, 0.006940447807312012, 0.006995488166809082, 0.007037951946258545, 0.006972640037536621, 0.007229184150695801, 0.006959136009216309, 0.006918943881988525, 0.00703711986541748, 0.006952640056610108, 0.007084896087646484, 0.0069510722160339355]",tokens/s,36127.16765637822,kWh,8.244359399437341e-08,4.517496273690451e-08,2.1530780192739605e-07,3.42926358658674e-07,tokens/kWh,746515960.4567035,MB,1204.6336,1032.323072,0.0,385.875968,319.016448,s,24,9.87300927734375,0.41137538655598954,0.005250170554912943,0.4099345245361328,0.4181420715332031,0.4233946075439453,0.4242666848754883,"[0.42407754516601565, 0.42432318115234374, 0.40719000244140624, 0.41228839111328125, 0.4101876220703125, 0.414212158203125, 0.4090625305175781, 0.4143986206054687, 0.40732171630859376, 0.4050887145996094, 0.40830712890625, 0.41137936401367187, 0.40607290649414063, 0.4059980163574219, 0.4146722412109375, 0.4087408447265625, 0.4195246276855469, 0.41491610717773436, 0.40676226806640625, 0.406918212890625, 0.4065990295410156, 0.41191845703125, 0.40968142700195315, 0.4133681640625]",tokens/s,153.14479684220362,kWh,4.779430648629273e-06,2.6186744362525065e-06,8.932299516153335e-06,1.6330404601035114e-05,tokens/kWh,3857834.606008886,,s,1511,10.028284921169284,0.00663685302526094,0.0009178240455832627,0.006451200008392334,0.006739967823028564,0.006828543901443482,0.013709311771392823,"[0.007173120021820068, 0.007184383869171143, 0.007200767993927002, 0.0071495680809020995, 0.007126016139984131, 0.007072768211364746, 0.007037951946258545, 0.007685120105743408, 0.0066744318008422855, 0.006437888145446777, 0.006458367824554443, 0.006509568214416504, 0.006421599864959717, 0.006458271980285644, 0.006421504020690918, 0.006463488101959228, 0.006482944011688232, 0.006401023864746094, 0.00643071985244751, 0.0063836159706115725, 0.006421504020690918, 0.006469632148742676, 0.006384640216827392, 0.006552576065063476, 0.006398975849151611, 0.0064133119583129885, 0.006455296039581298, 0.00643174409866333, 0.0064542717933654785, 0.006463488101959228, 0.006492159843444824, 0.006782976150512696, 0.006812672138214112, 0.0068392958641052244, 0.006806528091430664, 0.006804480075836182, 0.006789120197296142, 0.006772736072540283, 0.00678604793548584, 0.00679423999786377, 0.006766592025756836, 0.006853663921356201, 0.006771679878234863, 0.006726655960083008, 0.006781951904296875, 0.006816768169403077, 0.006843391895294189, 0.006807551860809326, 0.006804480075836182, 0.006761472225189209, 0.006737919807434082, 0.006719488143920899, 0.006792191982269287, 0.006758463859558105, 0.006839231967926025, 0.006819839954376221, 0.0067983360290527345, 0.006739967823028564, 0.0067645440101623535, 0.006743040084838867, 0.006761472225189209, 0.006751232147216797, 0.01458073616027832, 0.00679423999786377, 0.006840320110321045, 0.0067420158386230465, 0.006801407814025879, 0.006770688056945801, 0.006746111869812011, 0.006738944053649902, 0.006813695907592773, 0.006779903888702392, 0.006773759841918945, 0.006851583957672119, 0.006763519763946534, 0.006756351947784424, 0.006847487926483154, 0.00681276798248291, 0.0078048319816589355, 0.00801587200164795, 0.007895040035247802, 0.006912000179290771, 0.0070563840866088865, 0.006979584217071533, 0.006883327960968018, 0.006642687797546387, 0.006708223819732666, 0.00676966381072998, 0.006684671878814697, 0.006696959972381592, 0.006726655960083008, 0.006725632190704346, 0.006759424209594727, 0.0067348480224609375, 0.006650879859924317, 0.006633471965789795, 0.0066375679969787596, 0.006654975891113281, 0.006619135856628418, 0.006665215969085693, 0.006609920024871826, 0.006623231887817383, 0.006766592025756836, 0.006639616012573242, 0.006661119937896728, 0.006687839984893799, 0.006606751918792724, 0.006617087841033936, 0.006676479816436768, 0.006639616012573242, 0.006601727962493896, 0.0066304001808166506, 0.006602911949157715, 0.006428512096405029, 0.006429696083068847, 0.006405119895935059, 0.006415359973907471, 0.006401023864746094, 0.006434815883636475, 0.006471680164337158, 0.006407167911529541, 0.006416384220123291, 0.0063907837867736815, 0.006420512199401856, 0.00643887996673584, 0.013719552040100098, 0.006394879817962646, 0.006371327877044678, 0.006469632148742676, 0.00643891191482544, 0.006406144142150879, 0.006416512012481689, 0.006423488140106202, 0.006403007984161377, 0.006440959930419922, 0.006412288188934326, 0.00642252779006958, 0.006396927833557129, 0.006418432235717773, 0.0064849920272827145, 0.0064000959396362305, 0.006432672023773193, 0.006384640216827392, 0.006396927833557129, 0.006493184089660644, 0.006319136142730713, 0.0063723201751708985, 0.006371327877044678, 0.0067123198509216305, 0.006756351947784424, 0.006391808032989502, 0.006519807815551758, 0.00653107213973999, 0.006464511871337891, 0.006558720111846924, 0.006474751949310303, 0.006491136074066162, 0.006494207859039307, 0.006419456005096436, 0.0064174079895019534, 0.006441984176635742, 0.006432767868041992, 0.006453248023986816, 0.00652288007736206, 0.0064542717933654785, 0.006419456005096436, 0.006427648067474365, 0.00643071985244751, 0.006405119895935059, 0.006451200008392334, 0.006452223777770996, 0.006426623821258545, 0.006467584133148193, 0.006418432235717773, 0.006432767868041992, 0.0064133119583129885, 0.0064174079895019534, 0.006426655769348145, 0.006408160209655762, 0.006457344055175781, 0.006435840129852295, 0.006421504020690918, 0.0064245758056640625, 0.006448128223419189, 0.006411263942718506, 0.006748159885406494, 0.006774784088134766, 0.006755328178405762, 0.014206975936889648, 0.006623231887817383, 0.006616064071655273, 0.006623231887817383, 0.0066007041931152345, 0.006676479816436768, 0.006643712043762207, 0.006635519981384277, 0.006613024234771728, 0.006632415771484375, 0.006651904106140137, 0.006681600093841553, 0.0067041277885437015, 0.006656000137329102, 0.0068577280044555666, 0.00674508810043335, 0.006650879859924317, 0.006619135856628418, 0.0066447358131408694, 0.006625279903411865, 0.006618112087249756, 0.006729728221893311, 0.006626304149627686, 0.006639616012573242, 0.006707200050354004, 0.006656064033508301, 0.006653984069824218, 0.006672287940979004, 0.0066375679969787596, 0.006814720153808594, 0.006761472225189209, 0.006612991809844971, 0.006346752166748047, 0.006346752166748047, 0.006326272010803223, 0.006411263942718506, 0.006392831802368164, 0.00633241605758667, 0.006384640216827392, 0.00637337589263916, 0.006388735771179199, 0.006427648067474365, 0.006479872226715088, 0.006437888145446777, 0.00643993616104126, 0.00643891191482544, 0.006395904064178467, 0.006462463855743408, 0.006402112007141113, 0.006379456043243408, 0.006403071880340576, 0.006472703933715821, 0.0064143362045288085, 0.0064102401733398436, 0.006423552036285401, 0.006469632148742676, 0.0064204797744750975, 0.006411263942718506, 0.006510591983795166, 0.006403071880340576, 0.006392831802368164, 0.00638156795501709, 0.0064133119583129885, 0.013722623825073242, 0.006378496170043945, 0.006416543960571289, 0.0064182720184326175, 0.0064031357765197755, 0.006423488140106202, 0.0063201279640197755, 0.006363135814666748, 0.006467584133148193, 0.006576128005981445, 0.006481919765472412, 0.006396927833557129, 0.006433792114257812, 0.006451200008392334, 0.006415359973907471, 0.006425600051879882, 0.0064328317642211915, 0.00648799991607666, 0.006403071880340576, 0.006392831802368164, 0.0064471039772033695, 0.006405280113220215, 0.006493023872375488, 0.0064245758056640625, 0.006403071880340576, 0.006486015796661377, 0.006457376003265381, 0.006411231994628906, 0.006376448154449463, 0.006392831802368164, 0.0064102401733398436, 0.00637440013885498, 0.006402048110961914, 0.006457344055175781, 0.00642252779006958, 0.006458367824554443, 0.006375423908233643, 0.0064133119583129885, 0.006470655918121338, 0.006396927833557129, 0.006423552036285401, 0.00672870397567749, 0.006721536159515381, 0.006689792156219483, 0.0066344962120056155, 0.006776832103729248, 0.006642687797546387, 0.006650879859924317, 0.006753312110900879, 0.006780896186828613, 0.006736000061035156, 0.006607744216918945, 0.006724639892578125, 0.006701119899749756, 0.006610847949981689, 0.006631487846374511, 0.0066620798110961915, 0.006633471965789795, 0.00643891191482544, 0.006427648067474365, 0.006568960189819336, 0.006662144184112549, 0.006627327919006347, 0.014199808120727539, 0.006609920024871826, 0.006598656177520752, 0.006542367935180664, 0.006734816074371338, 0.006602752208709717, 0.006428671836853027, 0.006432767868041992, 0.006486015796661377, 0.006448128223419189, 0.006435840129852295, 0.006497280120849609, 0.0064102401733398436, 0.006510591983795166, 0.006433792114257812, 0.006416384220123291, 0.0065146880149841305, 0.006419456005096436, 0.006397952079772949, 0.006409215927124024, 0.006407167911529541, 0.006505472183227539, 0.006616064071655273, 0.006609920024871826, 0.006726655960083008, 0.006670400142669678, 0.0065914239883422855, 0.006462463855743408, 0.006453248023986816, 0.006401023864746094, 0.006496255874633789, 0.006906879901885986, 0.0065136637687683106, 0.006446080207824707, 0.006412288188934326, 0.007236608028411865, 0.006837247848510742, 0.007045119762420654, 0.0073471999168395995, 0.0068280320167541505, 0.006746111869812011, 0.0067717118263244626, 0.006653952121734619, 0.006532095909118653, 0.006455296039581298, 0.006585343837738037, 0.006719488143920899, 0.006652927875518798, 0.006707327842712402, 0.006686592102050781, 0.006592512130737305, 0.006423552036285401, 0.006504447937011719, 0.0064133119583129885, 0.0064471039772033695, 0.0064102401733398436, 0.006459392070770263, 0.006436863899230957, 0.0065136637687683106, 0.006450175762176514, 0.006433792114257812, 0.00637440013885498, 0.006446080207824707, 0.013710335731506347, 0.006427648067474365, 0.006406144142150879, 0.00643891191482544, 0.00648089599609375, 0.006393856048583985, 0.006419456005096436, 0.00640719985961914, 0.006459392070770263, 0.006511583805084229, 0.006419456005096436, 0.006436863899230957, 0.006510591983795166, 0.006462463855743408, 0.006440959930419922, 0.006409215927124024, 0.006463488101959228, 0.006440959930419922, 0.006394976139068603, 0.006412191867828369, 0.0064143362045288085, 0.0064245758056640625, 0.0063907837867736815, 0.0064542717933654785, 0.006481919765472412, 0.006411263942718506, 0.006429696083068847, 0.006421504020690918, 0.006401023864746094, 0.006446080207824707, 0.0064102401733398436, 0.006378496170043945, 0.006426623821258545, 0.006402048110961914, 0.006404096126556396, 0.006427648067474365, 0.006445055961608887, 0.0064430079460144046, 0.006419456005096436, 0.006460415840148926, 0.006416416168212891, 0.006411231994628906, 0.006427648067474365, 0.006406144142150879, 0.006645760059356689, 0.0064471039772033695, 0.00638976001739502, 0.006425600051879882, 0.006391808032989502, 0.006448128223419189, 0.00638976001739502, 0.006392831802368164, 0.006429696083068847, 0.006409215927124024, 0.006387712001800537, 0.007012351989746094, 0.006702079772949219, 0.007081984043121338, 0.007332863807678222, 0.0068392958641052244, 0.006673408031463623, 0.00669593620300293, 0.006746111869812011, 0.014347264289855957, 0.006690815925598144, 0.006688767910003662, 0.0066499199867248535, 0.0067255678176879885, 0.006691840171813965, 0.006635519981384277, 0.006830080032348633, 0.006700032234191895, 0.0066795840263366695, 0.006716383934020996, 0.006725632190704346, 0.007526400089263916, 0.006790143966674805, 0.0067964158058166505, 0.0067736320495605466, 0.006964223861694336, 0.007197696208953858, 0.006743040084838867, 0.00669593620300293, 0.0067420158386230465, 0.006739967823028564, 0.006729728221893311, 0.006668288230895996, 0.006675456047058105, 0.0066416640281677245, 0.0064880638122558594, 0.006417471885681153, 0.00639686393737793, 0.006412288188934326, 0.00643993616104126, 0.006412288188934326, 0.006421504020690918, 0.006482944011688232, 0.006499328136444092, 0.006452223777770996, 0.006403071880340576, 0.006406144142150879, 0.006553664207458496, 0.006404160022735595, 0.006429567813873291, 0.006405119895935059, 0.006402048110961914, 0.0064143362045288085, 0.006427648067474365, 0.006432767868041992, 0.006427648067474365, 0.0064430079460144046, 0.006467584133148193, 0.006375423908233643, 0.006375423908233643, 0.00638259220123291, 0.006509568214416504, 0.00643174409866333, 0.006591487884521485, 0.006404096126556396, 0.006448128223419189, 0.006379519939422608, 0.00643993616104126, 0.00638976001739502, 0.0064430079460144046, 0.006402048110961914, 0.00637337589263916, 0.013760512351989745, 0.006391808032989502, 0.00643071985244751, 0.006475776195526123, 0.00643891191482544, 0.006441984176635742, 0.006386688232421875, 0.006433824062347412, 0.006426591873168946, 0.00646457576751709, 0.006425536155700683, 0.006429696083068847, 0.006431903839111328, 0.006438752174377441, 0.006412288188934326, 0.006509568214416504, 0.006440959930419922, 0.006423552036285401, 0.006490240097045899, 0.006375296115875244, 0.006433792114257812, 0.006461440086364746, 0.006454304218292236, 0.006518752098083496, 0.006444191932678223, 0.006492000102996826, 0.006445055961608887, 0.0064471039772033695, 0.006440959930419922, 0.006432767868041992, 0.006416384220123291, 0.006459392070770263, 0.006432767868041992, 0.006445055961608887, 0.006435840129852295, 0.006449151992797852, 0.006423552036285401, 0.0064542717933654785, 0.006446080207824707, 0.006426623821258545, 0.0064174079895019534, 0.006578239917755127, 0.006444992065429688, 0.006425600051879882, 0.00648089599609375, 0.006401023864746094, 0.0064471039772033695, 0.006394879817962646, 0.006445055961608887, 0.006421504020690918, 0.006375423908233643, 0.006411263942718506, 0.006403071880340576, 0.006423583984375, 0.006423520088195801, 0.0063907837867736815, 0.006458399772644043, 0.006437856197357178, 0.006845439910888672, 0.006688767910003662, 0.006692863941192627, 0.006648831844329834, 0.0066938881874084475, 0.01416806411743164, 0.006428671836853027, 0.006426623821258545, 0.00643174409866333, 0.006416384220123291, 0.006405151844024658, 0.006429664134979248, 0.0064245758056640625, 0.006423552036285401, 0.00643891191482544, 0.006395904064178467, 0.00643071985244751, 0.006415359973907471, 0.006406144142150879, 0.0064174079895019534, 0.006394879817962646, 0.006428671836853027, 0.00643071985244751, 0.006426623821258545, 0.006397952079772949, 0.0063580479621887205, 0.006450143814086914, 0.006505472183227539, 0.0063805441856384275, 0.006448128223419189, 0.006396927833557129, 0.006411263942718506, 0.00643071985244751, 0.006387712001800537, 0.006441984176635742, 0.006409215927124024, 0.00643996810913086, 0.00657097578048706, 0.006396927833557129, 0.006386688232421875, 0.006423552036285401, 0.006392831802368164, 0.006461440086364746, 0.006385663986206055, 0.006451200008392334, 0.006456319808959961, 0.00638156795501709, 0.006411263942718506, 0.00638259220123291, 0.0064031357765197755, 0.00643884801864624, 0.006520832061767578, 0.006445055961608887, 0.006456319808959961, 0.006396992206573486, 0.006405119895935059, 0.006384575843811035, 0.006475776195526123, 0.006407167911529541, 0.0064102401733398436, 0.006416384220123291, 0.006419456005096436, 0.0064143362045288085, 0.006406144142150879, 0.0064102401733398436, 0.0064430079460144046, 0.0064143681526184085, 0.006405087947845459, 0.013634655952453613, 0.006385568141937256, 0.0063714561462402345, 0.006417280197143555, 0.0064849920272827145, 0.006402048110961914, 0.006416416168212891, 0.006418399810791016, 0.006423552036285401, 0.006460415840148926, 0.006400000095367431, 0.006426623821258545, 0.006428671836853027, 0.006419456005096436, 0.006449151992797852, 0.006413407802581787, 0.006458271980285644, 0.006489088058471679, 0.006473855972290039, 0.006438784122467041, 0.006384640216827392, 0.0064204797744750975, 0.006436863899230957, 0.0064174079895019534, 0.0064440321922302245, 0.006433792114257812, 0.00643993616104126, 0.006433792114257812, 0.006643712043762207, 0.0064767999649047855, 0.007583744049072265, 0.006734879970550537, 0.006631392002105713, 0.0064440321922302245, 0.006394879817962646, 0.0064133119583129885, 0.006497280120849609, 0.006426623821258545, 0.0064225602149963375, 0.00639689588546753, 0.0065781760215759275, 0.006474751949310303, 0.006523903846740723, 0.006650879859924317, 0.006568960189819336, 0.006505472183227539, 0.0065474557876586915, 0.006477888107299805, 0.006470592021942139, 0.006462624073028564, 0.006430560111999511, 0.006412288188934326, 0.006434815883636475, 0.00642252779006958, 0.006441984176635742, 0.006396927833557129, 0.006546432018280029, 0.006408192157745361, 0.00638259220123291, 0.006455296039581298, 0.006456319808959961, 0.006442048072814941, 0.006507455825805664, 0.013747200012207032, 0.006405119895935059, 0.00638156795501709, 0.0063990077972412105, 0.006451200008392334, 0.00648905611038208, 0.006964223861694336, 0.007663616180419922, 0.006829055786132812, 0.006708223819732666, 0.006716415882110595, 0.006722559928894043, 0.006669312000274658, 0.006625279903411865, 0.0064737281799316405, 0.006497280120849609, 0.006542335987091064, 0.006440959930419922, 0.0064133119583129885, 0.006427648067474365, 0.006394879817962646, 0.006472703933715821, 0.006425600051879882, 0.006421504020690918, 0.00643891191482544, 0.006540287971496582, 0.006429696083068847, 0.0064430079460144046, 0.0064133119583129885, 0.0065413122177124024, 0.006426623821258545, 0.006436863899230957, 0.006468607902526856, 0.006378496170043945, 0.006623231887817383, 0.0066641921997070315, 0.006667263984680176, 0.006648831844329834, 0.006625279903411865, 0.006626304149627686, 0.006653952121734619, 0.0066416640281677245, 0.006636544227600098, 0.0064963197708129885, 0.006428607940673828, 0.00643071985244751, 0.006470655918121338, 0.006392831802368164, 0.006401023864746094, 0.006418432235717773, 0.006505472183227539, 0.006416384220123291, 0.006428671836853027, 0.006487040042877197, 0.006395999908447266, 0.0064174079895019534, 0.0064306240081787105, 0.0063907837867736815, 0.006453248023986816, 0.006429696083068847, 0.006451200008392334, 0.006582272052764892, 0.006408192157745361, 0.01368883228302002, 0.0063610877990722655, 0.006458367824554443, 0.006494207859039307, 0.006384640216827392, 0.006466559886932373, 0.006391808032989502, 0.006429696083068847, 0.006468607902526856, 0.006419456005096436, 0.006432767868041992, 0.006415359973907471, 0.006426623821258545, 0.0064440321922302245, 0.006450175762176514, 0.006453248023986816, 0.006465536117553711, 0.006403071880340576, 0.006437888145446777, 0.006401023864746094, 0.006433792114257812, 0.0064143362045288085, 0.006400000095367431, 0.006496255874633789, 0.006404096126556396, 0.006398975849151611, 0.006516736030578613, 0.0064880638122558594, 0.006448128223419189, 0.00642252779006958, 0.006551551818847656, 0.0064134721755981445, 0.006405983924865723, 0.006437888145446777, 0.006441984176635742, 0.0064204797744750975, 0.0064174079895019534, 0.006395904064178467, 0.006436863899230957, 0.006402048110961914, 0.00643174409866333, 0.00642252779006958, 0.006440959930419922, 0.006445055961608887, 0.006427648067474365, 0.006549503803253174, 0.006415520191192627, 0.006500192165374756, 0.0064102401733398436, 0.0064113597869873045, 0.006375328063964843, 0.006426623821258545, 0.006469632148742676, 0.006440959930419922, 0.006776832103729248, 0.006424672126770019, 0.0064347200393676755, 0.0064430079460144046, 0.006391903877258301, 0.00643779182434082, 0.006440000057220459, 0.006461408138275147, 0.006457376003265381, 0.013682623863220215, 0.006412288188934326, 0.0064204797744750975, 0.006428671836853027, 0.006419583797454834, 0.006430592060089111, 0.0064440321922302245, 0.006394879817962646, 0.006449151992797852, 0.006472703933715821, 0.006419456005096436, 0.0064174079895019534, 0.006378496170043945, 0.00643891191482544, 0.006386688232421875, 0.006423552036285401, 0.006495232105255127, 0.006367231845855713, 0.006502399921417237, 0.00643891191482544, 0.0063703680038452145, 0.006432703971862793, 0.006518784046173095, 0.0064143362045288085, 0.006451200008392334, 0.006425600051879882, 0.006415359973907471, 0.006400000095367431, 0.006408192157745361, 0.006457503795623779, 0.0064203200340271, 0.0064430079460144046, 0.006395904064178467, 0.006375423908233643, 0.0064471039772033695, 0.0064102401733398436, 0.006497280120849609, 0.006393856048583985, 0.006551551818847656, 0.006429696083068847, 0.006371327877044678, 0.006436863899230957, 0.00643071985244751, 0.006441984176635742, 0.006455296039581298, 0.0064204797744750975, 0.0065669121742248536, 0.006487040042877197, 0.00638259220123291, 0.0064245758056640625, 0.006405119895935059, 0.006463488101959228, 0.0064542717933654785, 0.006393856048583985, 0.006499328136444092, 0.006756351947784424, 0.0064440321922302245, 0.006441984176635742, 0.00643174409866333, 0.006435840129852295, 0.0065146880149841305, 0.006392831802368164, 0.006436863899230957, 0.013700096130371094, 0.006440959930419922, 0.006405119895935059, 0.006459487915039063, 0.006629280090332031, 0.0066406397819519045, 0.006717504024505615, 0.006809535980224609, 0.006686719894409179, 0.006616064071655273, 0.006617087841033936, 0.006618112087249756, 0.006681600093841553, 0.006665215969085693, 0.006667263984680176, 0.0066979842185974124, 0.006660096168518067, 0.006597631931304931, 0.006623231887817383, 0.007445536136627197, 0.006737887859344482, 0.00674508810043335, 0.006829055786132812, 0.006707200050354004, 0.006773759841918945, 0.006636544227600098, 0.006680575847625733, 0.006651904106140137, 0.006589439868927002, 0.006715392112731934, 0.006813695907592773, 0.006620160102844238, 0.006411424160003662, 0.006331232070922852, 0.006395904064178467, 0.006339583873748779, 0.006392831802368164, 0.007366655826568603, 0.006966303825378418, 0.00666428804397583, 0.006453120231628418, 0.006385663986206055, 0.006405119895935059, 0.006465536117553711, 0.006389791965484619, 0.006422495841979981, 0.006564864158630371, 0.0063907837867736815, 0.006569983959197998, 0.006409215927124024, 0.006406144142150879, 0.0064245758056640625, 0.006400000095367431, 0.006432767868041992, 0.006387712001800537, 0.0064204797744750975, 0.00643071985244751, 0.0064245758056640625, 0.0064542717933654785, 0.00644924783706665, 0.006412191867828369, 0.006437888145446777, 0.006398975849151611, 0.013867008209228515, 0.006391808032989502, 0.006415359973907471, 0.006393856048583985, 0.0063907837867736815, 0.006462463855743408, 0.006408192157745361, 0.0064174079895019534, 0.006441088199615479, 0.006565760135650635, 0.0064839677810668945, 0.006452223777770996, 0.006394879817962646, 0.006437888145446777, 0.006408192157745361, 0.006427648067474365, 0.006849535942077637, 0.006635615825653076, 0.00643881607055664, 0.006378496170043945, 0.006426623821258545, 0.006585343837738037, 0.00638976001739502, 0.006415359973907471, 0.006395904064178467, 0.006561791896820069, 0.006456319808959961, 0.006452223777770996, 0.006441984176635742, 0.006449151992797852, 0.0064849920272827145, 0.006423552036285401, 0.006524928092956543, 0.006425600051879882, 0.006526976108551025, 0.006657023906707763, 0.006650879859924317, 0.006650879859924317, 0.006602752208709717, 0.006463488101959228, 0.006470655918121338, 0.0064440321922302245, 0.006419456005096436, 0.0064143362045288085, 0.006432767868041992, 0.006496255874633789, 0.0064245758056640625, 0.006449151992797852, 0.006395904064178467, 0.006475776195526123, 0.006472703933715821, 0.006723584175109864, 0.006670335769653321, 0.006673408031463623, 0.006668288230895996, 0.006455296039581298, 0.006446080207824707, 0.006401023864746094, 0.00658025598526001, 0.0064020161628723146, 0.006465536117553711, 0.006405119895935059, 0.006394879817962646, 0.013710335731506347, 0.006375423908233643, 0.0066447358131408694, 0.006649856090545654, 0.006614016056060791, 0.006676479816436768, 0.0066416640281677245, 0.006605823993682861, 0.0066406397819519045, 0.006654975891113281, 0.006645760059356689, 0.006659071922302246, 0.0066406397819519045, 0.006660096168518067, 0.006676479816436768, 0.006619135856628418, 0.006633471965789795, 0.006722559928894043, 0.006628352165222168, 0.0066119680404663084, 0.006626304149627686, 0.006656000137329102, 0.0066416640281677245, 0.006628511905670166, 0.006628191947937011, 0.006661119937896728, 0.0066416640281677245, 0.006625279903411865, 0.006746111869812011, 0.006700032234191895, 0.006636544227600098, 0.006642687797546387, 0.00672870397567749, 0.006677504062652588, 0.0066304001808166506, 0.006676479816436768, 0.0066406397819519045, 0.006654975891113281, 0.006666240215301514, 0.006619135856628418, 0.006690815925598144, 0.006654975891113281, 0.006638591766357422, 0.006743040084838867, 0.006672383785247803, 0.006698016166687012, 0.006738912105560302, 0.006744063854217529, 0.006609920024871826, 0.006676479816436768, 0.006626431941986084, 0.006617055892944336, 0.006672287940979004, 0.006648831844329834, 0.00658739185333252, 0.006794303894042969, 0.006714303970336914, 0.0066979842185974124, 0.006661119937896728, 0.006723584175109864, 0.006648831844329834, 0.006661119937896728, 0.0067573761940002445, 0.014223360061645507, 0.006657023906707763, 0.006656000137329102, 0.0066416640281677245, 0.006658080101013183, 0.006655968189239502, 0.006487040042877197, 0.0066416640281677245, 0.006670335769653321, 0.006642687797546387, 0.0064839677810668945, 0.006432767868041992, 0.0064328317642211915, 0.006581183910369873, 0.006636544227600098, 0.006635519981384277, 0.0066713600158691405, 0.0065710082054138185, 0.006599679946899414, 0.006643712043762207, 0.0066356477737426756, 0.006671232223510742, 0.006651904106140137, 0.006614048004150391, 0.0067901120185852055, 0.006722559928894043, 0.006474751949310303, 0.006632448196411133, 0.007143424034118652, 0.0067041277885437015, 0.006716415882110595, 0.006658048152923584, 0.0066078720092773435, 0.006636544227600098, 0.006650879859924317, 0.006635519981384277, 0.006789120197296142, 0.006669312000274658, 0.006706175804138184, 0.006967296123504638, 0.006647808074951172, 0.006688767910003662, 0.006608895778656006, 0.006455296039581298, 0.0064204797744750975, 0.006435840129852295, 0.006423552036285401, 0.006519807815551758, 0.006415359973907471, 0.006378496170043945, 0.0064204797744750975, 0.006437888145446777, 0.00638156795501709, 0.006435840129852295, 0.0064133119583129885, 0.0064174079895019534, 0.006437920093536377, 0.006390751838684082, 0.006427648067474365, 0.006437888145446777, 0.006423552036285401, 0.006455296039581298, 0.006416384220123291, 0.013747200012207032, 0.006400000095367431, 0.006401023864746094, 0.00643993616104126, 0.006393856048583985, 0.006451200008392334, 0.0063907837867736815, 0.006351871967315674, 0.006372352123260498, 0.006328320026397705, 0.00637337589263916, 0.006336512088775635, 0.006468607902526856, 0.006369279861450195, 0.006329343795776367, 0.006366208076477051, 0.006329343795776367, 0.0063272957801818845, 0.006407167911529541, 0.006407264232635498, 0.006440927982330322, 0.006480832099914551, 0.006373504161834717, 0.006345600128173828, 0.006398975849151611, 0.0064880638122558594, 0.00648089599609375, 0.006409215927124024, 0.006371327877044678, 0.006329343795776367, 0.00638259220123291, 0.00658739185333252, 0.00653004789352417, 0.006500351905822754, 0.006386688232421875, 0.006372352123260498, 0.0063201279640197755, 0.006452223777770996, 0.006351935863494873, 0.006332352161407471, 0.0064471039772033695, 0.006337535858154297, 0.006559743881225586, 0.006628352165222168, 0.006624320030212402, 0.00664569616317749, 0.0067041277885437015, 0.0066447358131408694, 0.0066416959762573245, 0.006820831775665283, 0.006543424129486084, 0.006567872047424316, 0.006837247848510742, 0.0066416640281677245, 0.006516928195953369, 0.006547264099121094, 0.006498303890228272, 0.006373439788818359, 0.006352831840515136, 0.00644924783706665, 0.006367136001586914, 0.006322175979614258, 0.0063498239517211915, 0.013651968002319336, 0.006401023864746094, 0.006409215927124024, 0.006391808032989502, 0.006780928134918213, 0.006815743923187256, 0.0066744318008422855, 0.006676479816436768, 0.006650879859924317, 0.006763519763946534, 0.006653952121734619, 0.006518847942352295, 0.006337471961975098, 0.006369279861450195, 0.006313983917236328, 0.006354944229125976, 0.006386688232421875, 0.0065259838104248045, 0.006422495841979981, 0.006315008163452148, 0.0063201279640197755, 0.006335519790649414, 0.006337503910064698, 0.006342656135559082, 0.006359039783477783, 0.006413407802581787, 0.0064234561920166015, 0.006416384220123291, 0.006434815883636475, 0.007346176147460938, 0.006546432018280029, 0.0065136637687683106, 0.006407296180725098, 0.006365056037902832, 0.0064430079460144046, 0.006381631851196289, 0.0063159999847412105, 0.006361055850982666, 0.006386688232421875, 0.006296576023101807, 0.006391903877258301, 0.0064306240081787105, 0.006387712001800537, 0.0064542717933654785, 0.0064174079895019534, 0.006419456005096436, 0.006408192157745361, 0.006402048110961914, 0.006466559886932373, 0.006378496170043945, 0.006367231845855713, 0.0064174079895019534, 0.006415487766265869, 0.006441855907440185, 0.006429728031158447, 0.006436831951141357, 0.006419456005096436, 0.0063805441856384275, 0.006466559886932373, 0.006464511871337891, 0.0064542717933654785, 0.006434815883636475, 0.006416384220123291, 0.013765631675720215, 0.006334464073181153, 0.006585343837738037, 0.006478847980499268, 0.006404096126556396, 0.006468607902526856, 0.006396927833557129, 0.00643174409866333, 0.006520832061767578, 0.006384640216827392, 0.006407167911529541, 0.006416384220123291, 0.00648089599609375, 0.006516736030578613, 0.00687820816040039, 0.006523903846740723, 0.006461440086364746, 0.0064143362045288085, 0.006463488101959228, 0.006428671836853027, 0.006452223777770996, 0.0066641921997070315, 0.006406144142150879, 0.006441984176635742, 0.0065771517753601075, 0.006406144142150879, 0.006396927833557129, 0.0064143362045288085, 0.0063805441856384275, 0.006449151992797852, 0.0064440321922302245, 0.006455423831939697, 0.00639680004119873, 0.006428671836853027, 0.0064143362045288085, 0.006403071880340576, 0.006598656177520752, 0.006401023864746094, 0.006441984176635742, 0.006421504020690918, 0.00643174409866333, 0.00642252779006958, 0.006421504020690918, 0.0063805441856384275, 0.006407167911529541, 0.006407167911529541, 0.00643993616104126, 0.00652288007736206, 0.00638976001739502, 0.006452223777770996, 0.006369279861450195, 0.006404096126556396, 0.006462463855743408, 0.0064471039772033695, 0.006445055961608887, 0.006377471923828125, 0.006423552036285401, 0.006479872226715088, 0.006402048110961914, 0.006512703895568847, 0.0064347519874572755, 0.006446080207824707, 0.006435872077941895, 0.013683679580688477, 0.006388735771179199, 0.0064839677810668945, 0.006436863899230957, 0.006421504020690918, 0.00642252779006958, 0.00643891191482544, 0.006425727844238281, 0.006451072216033936, 0.006471680164337158, 0.006406144142150879, 0.006468607902526856, 0.006393856048583985, 0.00653004789352417, 0.0065075201988220215, 0.00643993616104126, 0.00652185583114624, 0.006450175762176514, 0.006509568214416504, 0.006676479816436768, 0.006632448196411133, 0.0068351998329162595, 0.006440959930419922, 0.006402143955230713, 0.0064174079895019534, 0.00644598388671875, 0.006446080207824707, 0.006652927875518798, 0.006681600093841553, 0.006709248065948486, 0.006656000137329102, 0.006688767910003662, 0.006627327919006347, 0.006434815883636475, 0.0064102401733398436, 0.006680575847625733, 0.0066806402206420895, 0.006641600131988525, 0.006511616230010986, 0.006407167911529541, 0.00637440013885498, 0.006534143924713135, 0.006397952079772949, 0.006451200008392334, 0.006492159843444824, 0.006648831844329834, 0.006635519981384277, 0.006657023906707763, 0.0066078720092773435, 0.006626304149627686, 0.006654975891113281, 0.006645760059356689, 0.006643712043762207, 0.006677504062652588, 0.006595583915710449, 0.006626304149627686, 0.006652927875518798, 0.0066273918151855465, 0.0064460158348083494, 0.006465536117553711, 0.006511616230010986, 0.006430784225463867, 0.006543295860290527, 0.014124032020568847, 0.0066581439971923825, 0.006596511840820313, 0.006700032234191895, 0.006686719894409179, 0.006642687797546387, 0.00667955207824707, 0.006622208118438721, 0.006626304149627686, 0.006654975891113281, 0.006623360157012939, 0.00660262393951416, 0.006436863899230957, 0.006651904106140137, 0.006649856090545654, 0.0067358717918396, 0.006621183872222901, 0.006739967823028564, 0.006423552036285401, 0.006520832061767578, 0.006412288188934326, 0.00643174409866333, 0.006398975849151611, 0.006523903846740723, 0.006459487915039063, 0.006413216114044189, 0.006402080059051514, 0.006484960079193115, 0.006574079990386963, 0.006466559886932373, 0.006418432235717773, 0.006406144142150879, 0.006449151992797852, 0.006414400100708008, 0.006428607940673828, 0.006395904064178467, 0.0064245758056640625, 0.006459392070770263, 0.006463488101959228, 0.006453248023986816, 0.006425600051879882, 0.006446112155914307, 0.006427616119384766, 0.006455296039581298, 0.006455296039581298, 0.0064143362045288085, 0.006573056221008301, 0.006455296039581298, 0.006425600051879882, 0.0064849920272827145, 0.006395967960357666, 0.006416319847106933, 0.006434815883636475, 0.006391808032989502, 0.006427648067474365, 0.006437888145446777, 0.006396992206573486, 0.006404032230377197, 0.006405119895935059, 0.00642252779006958, 0.006398047924041748, 0.006421408176422119, 0.006433792114257812, 0.01367142391204834, 0.0064102401733398436, 0.006398975849151611, 0.0066119680404663084, 0.006398975849151611, 0.006427648067474365, 0.006416384220123291, 0.006440959930419922, 0.006456319808959961, 0.006474751949310303, 0.006407167911529541, 0.006451200008392334, 0.006450175762176514, 0.0064430079460144046, 0.006427648067474365, 0.006391808032989502, 0.0064471039772033695, 0.0064430079460144046, 0.006407167911529541, 0.007112703800201416, 0.006710271835327148, 0.006666240215301514, 0.0066713600158691405, 0.006612991809844971, 0.006780928134918213, 0.006696959972381592, 0.006680575847625733, 0.006638591766357422, 0.0066744318008422855, 0.006651904106140137, 0.00662937593460083, 0.006762495994567871, 0.0066078720092773435, 0.006619200229644775, 0.006675392150878906, 0.006619135856628418, 0.0066344962120056155, 0.006677504062652588, 0.006730751991271973, 0.006673408031463623, 0.006651904106140137, 0.006609920024871826, 0.006625279903411865, 0.006492159843444824, 0.006575104236602783, 0.006667263984680176, 0.0067051520347595215, 0.006708223819732666, 0.006703104019165039, 0.0064778242111206055, 0.006404096126556396, 0.006426623821258545, 0.006429696083068847, 0.006387712001800537, 0.006464511871337891, 0.0067123198509216305, 0.006523903846740723, 0.0064430079460144046, 0.006453248023986816, 0.006553599834442139, 0.006370336055755615, 0.006380512237548828, 0.006481919765472412]",tokens/s,150.67382028709054,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Salesforce/codegen-6B-nl,Salesforce/codegen-6B-nl,cuda,0,42,,,True,,,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",codegen,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-410m,EleutherAI/pythia-410m,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.42.1,,0.31.0,,,,1.20.0,,,,0.11.1,,,,MB,1296.359424,1711.80032,0.0,1065.353216,1018.332672,s,10,0.17361059188842773,0.017361059188842773,0.0007148956448061991,0.017053423881530763,0.018308728981018065,0.01876489248275757,0.01912982328414917,"[0.01922105598449707, 0.017073280334472658, 0.017059328079223633, 0.01706777572631836, 0.018207359313964843, 0.0170446720123291, 0.016947423934936524, 0.016997695922851563, 0.01704751968383789, 0.016944480895996095]",tokens/s,14745.644100131893,kWh,2.0313776713739417e-07,1.1131044567898915e-07,5.680631692268027e-07,8.82511382043186e-07,tokens/kWh,290081244.5130283,MB,1296.687104,1711.80032,0.0,1065.353216,1033.284608,s,10,10.197119323730469,1.019711932373047,0.005646894322773547,1.0177832641601563,1.0269054931640624,1.0294218017578125,1.0314348486328127,"[1.0263463134765625, 1.0244886474609376, 1.0172501220703125, 1.01831640625, 1.0319381103515626, 1.0170787353515625, 1.014833740234375, 1.01376708984375, 1.0144899291992187, 1.0186102294921875]",tokens/s,61.78215435156088,kWh,1.2358225380227328e-05,6.771808589250536e-06,2.4178373294573337e-05,4.3308407264051204e-05,tokens/kWh,1454682.9121627407,,s,629,10.333282286643977,0.01642811174347215,0.0021056879571517374,0.01614028739929199,0.016374578857421877,0.01677640953063965,0.033515683593750005,"[0.01800704002380371, 0.018256895065307616, 0.016587776184082033, 0.01617305564880371, 0.016244735717773438, 0.01619865608215332, 0.01620889663696289, 0.016166912078857423, 0.016103424072265626, 0.016124927520751953, 0.016102399826049805, 0.016130048751831053, 0.016134143829345703, 0.016105472564697267, 0.016301055908203126, 0.01643110466003418, 0.016063488006591797, 0.016304128646850585, 0.01621196746826172, 0.0160819206237793, 0.01596723175048828, 0.016165887832641602, 0.01613926315307617, 0.01604915237426758, 0.0160819206237793, 0.016159744262695314, 0.0161474552154541, 0.016145408630371092, 0.016082944869995116, 0.015964159965515135, 0.016122880935668944, 0.016973823547363282, 0.017116159439086915, 0.01697996711730957, 0.016530431747436524, 0.016464895248413085, 0.016306175231933593, 0.016301055908203126, 0.016251903533935547, 0.01617203140258789, 0.016123903274536132, 0.01605632019042969, 0.015987711906433104, 0.01602560043334961, 0.015988736152648925, 0.016176128387451173, 0.01616486358642578, 0.01604198455810547, 0.016105472564697267, 0.015960063934326172, 0.015960063934326172, 0.015988736152648925, 0.015987711906433104, 0.015987711906433104, 0.015980544090270995, 0.016133119583129883, 0.01618022346496582, 0.016216064453125, 0.016274431228637695, 0.016149503707885742, 0.016102399826049805, 0.01617407989501953, 0.03362815856933594, 0.016052223205566405, 0.016124927520751953, 0.016165887832641602, 0.016145408630371092, 0.016151552200317384, 0.01619660758972168, 0.01617715263366699, 0.016243711471557617, 0.0161976318359375, 0.01620992088317871, 0.016126976013183594, 0.016160768508911134, 0.01617817687988281, 0.01616383934020996, 0.01617407989501953, 0.01618739128112793, 0.01619660758972168, 0.016248832702636717, 0.016183296203613282, 0.016199680328369142, 0.016142335891723633, 0.016210943222045898, 0.01618739128112793, 0.01619558334350586, 0.01615564727783203, 0.01619046401977539, 0.01619046401977539, 0.016777215957641603, 0.018033664703369142, 0.01705881690979004, 0.016735231399536133, 0.01679974365234375, 0.01660825538635254, 0.0160501766204834, 0.01641062355041504, 0.016693311691284178, 0.016474048614501954, 0.016727039337158203, 0.016266239166259765, 0.016244735717773438, 0.01622425651550293, 0.016070655822753906, 0.016002048492431642, 0.016100351333618163, 0.016145408630371092, 0.016292863845825196, 0.016067583084106444, 0.015972352027893065, 0.016203807830810546, 0.016218080520629882, 0.01603993606567383, 0.016022527694702148, 0.016068607330322265, 0.015963135719299318, 0.016026624679565428, 0.01602457618713379, 0.01605734443664551, 0.016275455474853515, 0.016004095077514647, 0.016219135284423827, 0.01618227195739746, 0.016142335891723633, 0.03356671905517578, 0.01599180793762207, 0.016035839080810545, 0.015977472305297852, 0.01602457618713379, 0.016005119323730468, 0.016013311386108398, 0.01601945686340332, 0.016062463760375977, 0.016021503448486327, 0.01598361587524414, 0.01604812812805176, 0.016110591888427735, 0.01605632019042969, 0.01605632019042969, 0.01601740837097168, 0.01600102424621582, 0.015977472305297852, 0.01600921630859375, 0.016130048751831053, 0.016165887832641602, 0.01601126480102539, 0.015956992149353026, 0.015994879722595216, 0.015988736152648925, 0.015979519844055178, 0.01603993606567383, 0.01658982467651367, 0.016021503448486327, 0.01599078369140625, 0.016002048492431642, 0.01600716781616211, 0.016076799392700195, 0.016022527694702148, 0.01618124771118164, 0.01603276824951172, 0.016012287139892577, 0.01597542381286621, 0.01599283218383789, 0.01599078369140625, 0.015989760398864745, 0.01604812812805176, 0.01600307273864746, 0.015951871871948242, 0.015944704055786133, 0.015963135719299318, 0.0160501766204834, 0.01677519989013672, 0.016592863082885743, 0.01627136039733887, 0.01721958351135254, 0.01660723114013672, 0.016326656341552736, 0.01621504020690918, 0.016226303100585936, 0.01641267204284668, 0.016661504745483398, 0.01678335952758789, 0.01661849594116211, 0.016150527954101563, 0.016037887573242187, 0.01620377540588379, 0.016154655456542967, 0.03355849456787109, 0.016138240814208983, 0.016342016220092775, 0.01640140724182129, 0.016312320709228514, 0.016161792755126952, 0.016112640380859376, 0.016183296203613282, 0.016192512512207033, 0.01616486358642578, 0.016151552200317384, 0.01619865608215332, 0.01598361587524414, 0.01601638412475586, 0.016075775146484374, 0.01620889663696289, 0.01616998481750488, 0.016142335891723633, 0.01620479965209961, 0.016118783950805664, 0.016144384384155275, 0.016133119583129883, 0.01616896057128906, 0.016159744262695314, 0.016112640380859376, 0.016150527954101563, 0.01613107109069824, 0.016115711212158202, 0.016158720016479493, 0.016274431228637695, 0.016058399200439454, 0.01625596809387207, 0.016143360137939454, 0.01620992088317871, 0.01615667152404785, 0.01620070457458496, 0.016259071350097656, 0.016533504486083983, 0.016113664627075194, 0.016105472564697267, 0.01621401596069336, 0.01621811294555664, 0.01614028739929199, 0.01616486358642578, 0.016136255264282227, 0.016198591232299806, 0.01600921630859375, 0.01599078369140625, 0.01594777584075928, 0.015964159965515135, 0.01625395202636719, 0.016160768508911134, 0.016094207763671875, 0.016123903274536132, 0.016093183517456054, 0.016116735458374023, 0.01620992088317871, 0.016109567642211914, 0.016210943222045898, 0.01616383934020996, 0.016183296203613282, 0.016067583084106444, 0.016061439514160156, 0.0355788803100586, 0.01622220802307129, 0.01615667152404785, 0.01616998481750488, 0.01616281509399414, 0.016153600692749022, 0.01618534469604492, 0.01617919921875, 0.016220159530639648, 0.016550912857055664, 0.017925119400024413, 0.01699839973449707, 0.016487424850463867, 0.016307199478149414, 0.016133119583129883, 0.016108543395996093, 0.016184320449829103, 0.01613209533691406, 0.01616896057128906, 0.01621504020690918, 0.016165887832641602, 0.0161474552154541, 0.01618124771118164, 0.016098304748535155, 0.016157695770263672, 0.016154624938964843, 0.016114688873291014, 0.016087039947509766, 0.01616281509399414, 0.01614028739929199, 0.01613107109069824, 0.01614028739929199, 0.017575935363769533, 0.01721446418762207, 0.01685196876525879, 0.016686080932617187, 0.016086015701293945, 0.016137216567993166, 0.016138240814208983, 0.01615564727783203, 0.016144384384155275, 0.016133119583129883, 0.01620684814453125, 0.01616998481750488, 0.01615667152404785, 0.016297983169555663, 0.016133119583129883, 0.016301088333129883, 0.016157663345336915, 0.01615667152404785, 0.01633791923522949, 0.016322559356689453, 0.016858112335205077, 0.0169932804107666, 0.01738035202026367, 0.016785408020019533, 0.01681510353088379, 0.016639999389648438, 0.016623615264892578, 0.01619660758972168, 0.01613209533691406, 0.01637171173095703, 0.01623142433166504, 0.033683456420898435, 0.01619660758972168, 0.016097280502319337, 0.015988767623901366, 0.01602457618713379, 0.015972319602966308, 0.0162478084564209, 0.016290815353393554, 0.01619353675842285, 0.016153600692749022, 0.016243711471557617, 0.016117759704589844, 0.01617407989501953, 0.016130048751831053, 0.01618636894226074, 0.016126976013183594, 0.01623040008544922, 0.016299007415771484, 0.01618227195739746, 0.016124927520751953, 0.01623756790161133, 0.016141311645507812, 0.016236543655395508, 0.016192512512207033, 0.016059392929077147, 0.016097280502319337, 0.016154624938964843, 0.01619046401977539, 0.01619558334350586, 0.016199680328369142, 0.016146432876586913, 0.016102399826049805, 0.01601433563232422, 0.01601945686340332, 0.01602355194091797, 0.016053247451782226, 0.015981568336486816, 0.016020479202270507, 0.01598259162902832, 0.01601638412475586, 0.01600921630859375, 0.0162293758392334, 0.016097280502319337, 0.01595084762573242, 0.015972352027893065, 0.01617407989501953, 0.016175104141235352, 0.01616486358642578, 0.01617919921875, 0.016134143829345703, 0.01616383934020996, 0.016096256256103517, 0.016740352630615234, 0.016089088439941408, 0.01599283218383789, 0.016094207763671875, 0.01616998481750488, 0.01620070457458496, 0.016137216567993166, 0.016137216567993166, 0.01623961639404297, 0.016120832443237306, 0.016189472198486328, 0.03327791976928711, 0.01597542381286621, 0.015980544090270995, 0.016103424072265626, 0.01615564727783203, 0.015962112426757814, 0.01595084762573242, 0.015969280242919923, 0.016104448318481446, 0.016142335891723633, 0.015981568336486816, 0.01601638412475586, 0.015953920364379884, 0.016144384384155275, 0.016160768508911134, 0.01617817687988281, 0.016304128646850585, 0.016135168075561524, 0.016153600692749022, 0.016122880935668944, 0.01621708869934082, 0.016087039947509766, 0.016300031661987305, 0.01620479965209961, 0.016251903533935547, 0.016359424591064452, 0.01614028739929199, 0.016117759704589844, 0.016021503448486327, 0.016013311386108398, 0.016082944869995116, 0.01602457618713379, 0.015954943656921388, 0.01602355194091797, 0.016012287139892577, 0.016038911819458008, 0.01594777584075928, 0.015985664367675782, 0.015935487747192383, 0.01597644805908203, 0.015980544090270995, 0.01594777584075928, 0.016036863327026366, 0.01616383934020996, 0.016175104141235352, 0.01621504020690918, 0.01641472053527832, 0.01621401596069336, 0.016152576446533205, 0.016167936325073243, 0.01622118377685547, 0.01621401596069336, 0.016103424072265626, 0.01620684814453125, 0.01613209533691406, 0.016257024765014647, 0.016166912078857423, 0.01615667152404785, 0.0162293758392334, 0.01622118377685547, 0.01601126480102539, 0.015986687660217287, 0.015972352027893065, 0.0335728645324707, 0.016160768508911134, 0.01619049644470215, 0.016162784576416015, 0.01618227195739746, 0.016151552200317384, 0.016152576446533205, 0.016112640380859376, 0.01617203140258789, 0.01614847946166992, 0.01618124771118164, 0.01615564727783203, 0.016150527954101563, 0.016134143829345703, 0.016022527694702148, 0.01600614356994629, 0.016081951141357423, 0.016113632202148436, 0.01615564727783203, 0.01603276824951172, 0.015979519844055178, 0.016130048751831053, 0.01618227195739746, 0.01613107109069824, 0.01618841552734375, 0.016161792755126952, 0.01618022346496582, 0.016089088439941408, 0.016158720016479493, 0.016108543395996093, 0.016102399826049805, 0.01598464012145996, 0.015952896118164063, 0.015961088180541993, 0.01596723175048828, 0.016013311386108398, 0.015960063934326172, 0.016010240554809572, 0.015953920364379884, 0.015965184211730956, 0.015956992149353026, 0.015961088180541993, 0.01594777584075928, 0.016070655822753906, 0.015953920364379884, 0.01600819206237793, 0.016092159271240233, 0.01618227195739746, 0.01613107109069824, 0.016194559097290038, 0.01616486358642578, 0.01618124771118164, 0.016124927520751953, 0.016517120361328123, 0.015945728302001954, 0.01601740837097168, 0.016160768508911134, 0.016138240814208983, 0.015935487747192383, 0.015965184211730956, 0.015946751594543458, 0.015997952461242675, 0.015981568336486816, 0.033538047790527346, 0.016126976013183594, 0.016175104141235352, 0.016110591888427735, 0.01619353675842285, 0.01608399963378906, 0.01604707145690918, 0.015921152114868165, 0.015965184211730956, 0.016121856689453123, 0.016136192321777345, 0.01599180793762207, 0.016067583084106444, 0.016134143829345703, 0.016165887832641602, 0.01618022346496582, 0.01603379249572754, 0.016064512252807618, 0.01597542381286621, 0.015932415962219237, 0.016075775146484374, 0.01617919921875, 0.01606553649902344, 0.01602457618713379, 0.01601126480102539, 0.016060415267944335, 0.016, 0.01603276824951172, 0.016030719757080078, 0.01602764892578125, 0.015955967903137205, 0.016191488265991212, 0.015993856430053712, 0.01620582389831543, 0.016141311645507812, 0.016219135284423827, 0.016104448318481446, 0.016043008804321288, 0.01601638412475586, 0.015997952461242675, 0.01607372856140137, 0.016022527694702148, 0.016004095077514647, 0.015953920364379884, 0.01599180793762207, 0.016092159271240233, 0.016067583084106444, 0.016114688873291014, 0.01620684814453125, 0.01620070457458496, 0.016227327346801757, 0.016280576705932616, 0.01682636833190918, 0.016002048492431642, 0.016133119583129883, 0.016471040725708007, 0.016415775299072264, 0.016180192947387696, 0.01601433563232422, 0.015978495597839357, 0.01600819206237793, 0.015927295684814453, 0.016035839080810545, 0.03345817565917969, 0.016160768508911134, 0.01614028739929199, 0.01617203140258789, 0.01617718315124512, 0.01610032081604004, 0.016136192321777345, 0.016101375579833984, 0.016126976013183594, 0.016005119323730468, 0.01601433563232422, 0.016119808197021485, 0.016118783950805664, 0.01621504020690918, 0.016160768508911134, 0.01616281509399414, 0.01619353675842285, 0.016128000259399415, 0.01618943977355957, 0.016130048751831053, 0.016295936584472655, 0.016111616134643555, 0.016101375579833984, 0.016129024505615236, 0.016107519149780272, 0.01638604736328125, 0.01613926315307617, 0.016110591888427735, 0.016166912078857423, 0.016111616134643555, 0.016154624938964843, 0.016161792755126952, 0.016194559097290038, 0.01621299171447754, 0.016311296463012694, 0.017582080841064454, 0.01613107109069824, 0.01601945686340332, 0.01616896057128906, 0.01616896057128906, 0.016695295333862305, 0.01620684814453125, 0.016210943222045898, 0.01616998481750488, 0.01616592025756836, 0.016148448944091797, 0.016020479202270507, 0.016154624938964843, 0.016116735458374023, 0.016029695510864257, 0.01601945686340332, 0.01617305564880371, 0.015930368423461915, 0.01603993606567383, 0.01600716781616211, 0.016004095077514647, 0.016062463760375977, 0.016119808197021485, 0.016070655822753906, 0.016220159530639648, 0.01617407989501953, 0.01600819206237793, 0.0160501766204834]",tokens/s,60.87126844613523,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,.,.,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 373, in cached_file raise EnvironmentError( OSError: . does not appear to have a file named config.json. Checkout 'https://huggingface.co/./tree/None' for available files. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,a,a,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/a/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66947492-789c207f7cc768ab1fc98ebb;7d259cac-2af5-4fbc-b861-9b7e93c14ae3) Repository Not Found for url: https://huggingface.co/a/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: a is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-7B,Qwen/Qwen-7B,cuda,0,42,,,True,,,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",qwen,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-66b,facebook/opt-66b,cuda,0,42,,,True,,,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",opt,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,stabilityai/stablelm-base-alpha-7b,stabilityai/stablelm-base-alpha-7b,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.42.3,,0.31.0,,,,1.20.0,,,,0.11.1,,,,MB,1263.24736,17176.199168,0.0,16529.752064,16322.781696,s,10,0.6743778381347658,0.06743778381347658,0.0001527504409071489,0.06738692474365235,0.06746442642211914,0.06767945976257324,0.06785148643493652,"[0.06789449310302734, 0.06739657592773438, 0.06738883209228516, 0.0673861083984375, 0.06738428497314453, 0.06738230133056641, 0.06736777496337891, 0.06741664123535156, 0.06737308502197266, 0.06738774108886719]",tokens/s,3796.0915309444326,kWh,7.965522491158523e-07,4.3630755597663713e-07,4.945197841315525e-06,6.178057646408014e-06,tokens/kWh,41436971.72344143,MB,1263.57504,17176.199168,0.0,16529.752064,16442.131456,s,10,22.018045654296877,2.2018045654296876,0.0016303193812182248,2.201161376953125,2.2028490234375,2.2046382568359375,2.2060696435546876,"[2.206427490234375, 2.200530517578125, 2.201146240234375, 2.201730224609375, 2.201761474609375, 2.201018798828125, 2.201176513671875, 2.202451416015625, 2.2006982421875, 2.201104736328125]",tokens/s,28.61289370962195,kWh,2.599383024872461e-05,1.424559139638557e-05,0.00012815054752408417,0.00016838996916919432,tokens/kWh,374131.54899208434,,s,629,22.589461528778088,0.0359132933684866,0.008014884036052092,0.034941951751708986,0.03503964004516602,0.035121151733398435,0.10240909301757813,"[0.03525632095336914, 0.035248126983642575, 0.03524198532104492, 0.03528396987915039, 0.035253246307373046, 0.03520204925537109, 0.035280895233154294, 0.03522867202758789, 0.03525836944580078, 0.03525529479980469, 0.03521331024169922, 0.03521638488769531, 0.03501875305175781, 0.03488870239257812, 0.03489996719360351, 0.03486105728149414, 0.03488460922241211, 0.03490816116333008, 0.03488358306884766, 0.034874366760253905, 0.03487641525268555, 0.03488358306884766, 0.035179519653320314, 0.03512319946289062, 0.03491635131835937, 0.03489382553100586, 0.034923519134521484, 0.034900993347167966, 0.03489689636230469, 0.034933761596679686, 0.03489996719360351, 0.0349224967956543, 0.034941951751708986, 0.034947071075439456, 0.03492147064208984, 0.03496448135375976, 0.034947071075439456, 0.03494604873657227, 0.03496448135375976, 0.034936832427978515, 0.03495935821533203, 0.03496038436889649, 0.034966529846191405, 0.034956287384033204, 0.03496038436889649, 0.03496550369262695, 0.03495731353759766, 0.035007488250732424, 0.034977790832519534, 0.034985984802246094, 0.03500236892700195, 0.03500236892700195, 0.0350115852355957, 0.03505766296386719, 0.035019775390625, 0.03501875305175781, 0.03501363372802734, 0.03501772689819336, 0.03502899169921875, 0.03504537582397461, 0.03503206253051758, 0.03503513717651367, 0.10240204620361328, 0.034813953399658204, 0.03482112121582031, 0.03482828903198242, 0.03484467315673828, 0.03484262466430664, 0.03483852767944336, 0.03486003112792969, 0.03482316970825195, 0.03486105728149414, 0.03483443069458008, 0.03486515045166016, 0.03486105728149414, 0.03486412811279297, 0.03487846374511719, 0.034874366760253905, 0.03486310577392578, 0.03486207962036133, 0.034852863311767575, 0.03487539291381836, 0.03488051223754883, 0.03488972854614258, 0.034909183502197266, 0.035019775390625, 0.034874366760253905, 0.03490611267089844, 0.03501363372802734, 0.034900993347167966, 0.034885631561279294, 0.03488972854614258, 0.03489689636230469, 0.03490611267089844, 0.0349224967956543, 0.034928638458251955, 0.0349194221496582, 0.0349378547668457, 0.034928638458251955, 0.034936832427978515, 0.034955265045166016, 0.034966529846191405, 0.03495935821533203, 0.034933761596679686, 0.034941951751708986, 0.034955265045166016, 0.03498188781738281, 0.03498291015625, 0.034988033294677735, 0.034966529846191405, 0.03496755218505859, 0.0350013427734375, 0.034988033294677735, 0.03500543975830078, 0.03499520111083984, 0.03501772689819336, 0.03501875305175781, 0.03503411102294922, 0.03501363372802734, 0.03504127883911133, 0.03501260757446289, 0.03501260757446289, 0.0350382080078125, 0.03503308868408203, 0.03502899169921875, 0.1024194564819336, 0.03482112121582031, 0.03486822509765625, 0.03483135986328125, 0.03483852767944336, 0.03483647918701172, 0.03482624053955078, 0.03488051223754883, 0.03484262466430664, 0.03484774398803711, 0.034855934143066404, 0.03484364700317383, 0.03486515045166016, 0.03486822509765625, 0.034912254333496096, 0.034857982635498046, 0.034874366760253905, 0.03487027359008789, 0.03486617660522461, 0.034872318267822264, 0.034900993347167966, 0.03489791870117188, 0.03490611267089844, 0.03487641525268555, 0.03492761611938477, 0.03489894485473633, 0.034931713104248044, 0.034904064178466795, 0.03491020965576172, 0.03491123199462891, 0.0349306869506836, 0.03493478393554687, 0.03492966461181641, 0.034941951751708986, 0.03494297790527344, 0.034947071075439456, 0.03497881698608398, 0.03494297790527344, 0.03497062301635742, 0.034971649169921876, 0.03498188781738281, 0.03495116806030273, 0.034948097229003904, 0.034956287384033204, 0.03498086547851562, 0.03496448135375976, 0.034993152618408206, 0.03496857452392578, 0.03497267150878906, 0.03498393630981445, 0.03500236892700195, 0.03499622344970703, 0.03499622344970703, 0.03499622344970703, 0.03501363372802734, 0.03502284622192383, 0.0353259506225586, 0.03503308868408203, 0.03501260757446289, 0.035037185668945314, 0.035056640625, 0.035059711456298825, 0.035059711456298825, 0.10242559814453125, 0.034953216552734374, 0.0348221435546875, 0.03481497573852539, 0.034835456848144535, 0.03484262466430664, 0.034840576171875, 0.03491430282592774, 0.03484672164916992, 0.03486515045166016, 0.03483852767944336, 0.034974720001220705, 0.034923519134521484, 0.03489894485473633, 0.03486822509765625, 0.034920448303222655, 0.03487027359008789, 0.03487744140625, 0.0348590087890625, 0.03489279937744141, 0.034895870208740236, 0.034950145721435545, 0.034900993347167966, 0.034895870208740236, 0.03489177703857422, 0.03489484786987305, 0.03491020965576172, 0.034912254333496096, 0.03491123199462891, 0.034909183502197266, 0.034931713104248044, 0.03505049514770508, 0.034936832427978515, 0.03494400024414063, 0.03494911956787109, 0.03495731353759766, 0.0349306869506836, 0.03498393630981445, 0.035125247955322264, 0.03499724960327148, 0.03498086547851562, 0.034971649169921876, 0.03496857452392578, 0.03499622344970703, 0.03496857452392578, 0.035004417419433595, 0.034977790832519534, 0.03497369766235352, 0.03495731353759766, 0.03501260757446289, 0.03498086547851562, 0.03500236892700195, 0.035004417419433595, 0.03501875305175781, 0.03500339126586914, 0.03502489471435547, 0.035023872375488284, 0.03506073760986328, 0.03500543975830078, 0.03501567840576172, 0.035108863830566404, 0.03505561447143555, 0.035056640625, 0.10241024017333984, 0.03482624053955078, 0.03481804656982422, 0.03483852767944336, 0.03483852767944336, 0.03484262466430664, 0.034841598510742186, 0.03487744140625, 0.03483443069458008, 0.034855934143066404, 0.03484262466430664, 0.0348590087890625, 0.03486412811279297, 0.03487948989868164, 0.03491328048706055, 0.03492659378051758, 0.03486310577392578, 0.03487027359008789, 0.03486207962036133, 0.03487846374511719, 0.03487027359008789, 0.03487744140625, 0.03486207962036133, 0.034950145721435545, 0.034890750885009765, 0.034895870208740236, 0.034890750885009765, 0.03492147064208984, 0.034907135009765625, 0.03492454528808594, 0.03496755218505859, 0.03493273544311523, 0.034947071075439456, 0.035053569793701174, 0.03509145736694336, 0.03494604873657227, 0.034931713104248044, 0.034948097229003904, 0.03498700714111328, 0.03513241577148438, 0.034948097229003904, 0.03502899169921875, 0.03498086547851562, 0.03498188781738281, 0.03497881698608398, 0.03498495864868164, 0.03500851058959961, 0.03495935821533203, 0.03498495864868164, 0.03502489471435547, 0.03500543975830078, 0.03526758575439453, 0.035026943206787106, 0.03500339126586914, 0.034991104125976565, 0.03503206253051758, 0.03502592086791992, 0.035037185668945314, 0.03503411102294922, 0.03505868911743164, 0.03504230499267578, 0.03507199859619141, 0.03503923034667969, 0.10242355346679688, 0.03482726287841797, 0.03487744140625, 0.03483750534057617, 0.03483750534057617, 0.03486617660522461, 0.03483955383300781, 0.03483647918701172, 0.03483955383300781, 0.034857982635498046, 0.03483750534057617, 0.03485696029663086, 0.034855934143066404, 0.03488153457641602, 0.03486105728149414, 0.034895870208740236, 0.03491839981079101, 0.03487846374511719, 0.03487846374511719, 0.03489894485473633, 0.0348671989440918, 0.03494604873657227, 0.03487641525268555, 0.03490611267089844, 0.034917377471923826, 0.03490304183959961, 0.03491430282592774, 0.03492761611938477, 0.03488972854614258, 0.03490611267089844, 0.03490611267089844, 0.03494400024414063, 0.034939903259277344, 0.034941951751708986, 0.034941951751708986, 0.03497062301635742, 0.03502182388305664, 0.03497574234008789, 0.03493478393554687, 0.03500236892700195, 0.034966529846191405, 0.034961406707763674, 0.034938880920410156, 0.034993152618408206, 0.034976768493652347, 0.03498188781738281, 0.03496857452392578, 0.03496448135375976, 0.03498188781738281, 0.03500032043457031, 0.03499929428100586, 0.03499827194213867, 0.03504844665527344, 0.03501567840576172, 0.03500236892700195, 0.03504537582397461, 0.035007488250732424, 0.035043327331542966, 0.03501772689819336, 0.03503104019165039, 0.03502899169921875, 0.035007488250732424, 0.03501772689819336, 0.10242457580566407, 0.03482931137084961, 0.03482316970825195, 0.034852863311767575, 0.03483443069458008, 0.03483852767944336, 0.034830337524414064, 0.03486617660522461, 0.03488051223754883, 0.03486515045166016, 0.034855934143066404, 0.03488460922241211, 0.03487334442138672, 0.0348590087890625, 0.03488265609741211, 0.03495516967773438, 0.03488358306884766, 0.03488051223754883, 0.03487744140625, 0.03487539291381836, 0.03487948989868164, 0.03489177703857422, 0.03488256072998047, 0.03489689636230469, 0.03489177703857422, 0.03491328048706055, 0.034900993347167966, 0.03492966461181641, 0.034939903259277344, 0.03494911956787109, 0.034933761596679686, 0.03492454528808594, 0.034909183502197266, 0.03493273544311523, 0.03491328048706055, 0.034925567626953126, 0.03493273544311523, 0.034947071075439456, 0.03494400024414063, 0.034994174957275394, 0.035146751403808595, 0.03495935821533203, 0.03496345520019531, 0.0350115852355957, 0.03498700714111328, 0.03498086547851562, 0.03495219039916992, 0.03498495864868164, 0.03498393630981445, 0.03497983932495117, 0.03497881698608398, 0.03501055908203125, 0.035064830780029296, 0.03502284622192383, 0.034991104125976565, 0.03502796936035156, 0.03500543975830078, 0.03502592086791992, 0.03503513717651367, 0.03503411102294922, 0.03503308868408203, 0.03505152130126953, 0.03502796936035156, 0.10252082824707032, 0.03483340835571289, 0.034830337524414064, 0.034841598510742186, 0.03483955383300781, 0.034825214385986326, 0.034825214385986326, 0.03497881698608398, 0.03485081481933594, 0.03487539291381836, 0.03484467315673828, 0.034852863311767575, 0.03488358306884766, 0.03488358306884766, 0.03488870239257812, 0.034917377471923826, 0.03488153457641602, 0.034993152618408206, 0.03526348876953125, 0.034907135009765625, 0.03488460922241211, 0.03494911956787109, 0.03492454528808594, 0.034887680053710936, 0.03488665771484375, 0.034915328979492184, 0.03491020965576172, 0.034928638458251955, 0.034909183502197266, 0.03490816116333008, 0.0349224967956543, 0.0349194221496582, 0.0349224967956543, 0.03522969436645508, 0.03502796936035156, 0.03494297790527344, 0.034950145721435545, 0.03505868911743164, 0.035135486602783206, 0.03499212646484375, 0.034966529846191405, 0.034953216552734374, 0.03497062301635742, 0.034971649169921876, 0.03496550369262695, 0.034994174957275394, 0.03497267150878906, 0.03497369766235352, 0.03507097625732422, 0.03499827194213867, 0.03501567840576172, 0.03501260757446289, 0.03500032043457031, 0.03501875305175781, 0.035020801544189455, 0.03502284622192383, 0.03500032043457031, 0.035043327331542966, 0.035046398162841795, 0.03503411102294922, 0.03503513717651367, 0.035043327331542966, 0.0350382080078125, 0.10246041870117187, 0.03483443069458008, 0.0348221435546875, 0.03482112121582031, 0.03481804656982422, 0.03483443069458008, 0.03483647918701172, 0.034835456848144535, 0.03486515045166016, 0.03487027359008789, 0.03485081481933594, 0.034852863311767575, 0.03486515045166016, 0.03486003112792969, 0.03487027359008789, 0.034869247436523435, 0.03485081481933594, 0.03486105728149414, 0.03485081481933594, 0.034872318267822264, 0.034890750885009765, 0.03488870239257812, 0.03487027359008789, 0.034887680053710936, 0.03487641525268555, 0.034895870208740236, 0.03488972854614258, 0.03492659378051758, 0.034890750885009765, 0.03491328048706055, 0.034915328979492184, 0.03491635131835937, 0.0349194221496582, 0.03493478393554687, 0.03492966461181641, 0.03492966461181641, 0.03492659378051758, 0.034966529846191405, 0.0349378547668457, 0.03496755218505859, 0.034956287384033204, 0.03495219039916992, 0.034971649169921876, 0.034974720001220705, 0.03496038436889649, 0.034985984802246094, 0.03497574234008789, 0.03497062301635742, 0.034971649169921876, 0.03499827194213867, 0.0350382080078125, 0.03503923034667969, 0.03499929428100586, 0.03501055908203125, 0.03502592086791992, 0.03502284622192383, 0.0351907844543457, 0.03505459213256836, 0.03500339126586914, 0.035019775390625, 0.0350382080078125, 0.03504947280883789, 0.03502796936035156, 0.10240614318847656, 0.03482419204711914, 0.03484467315673828, 0.034813953399658204, 0.034840576171875, 0.03485184097290039, 0.034841598510742186, 0.03484569549560547, 0.0348590087890625, 0.03486515045166016, 0.03483955383300781, 0.0348487663269043, 0.0348590087890625, 0.03511808013916016, 0.034887680053710936, 0.034904064178466795, 0.034872318267822264, 0.034871295928955076, 0.03486412811279297, 0.03488153457641602, 0.03487334442138672, 0.0349194221496582, 0.03488460922241211, 0.03486617660522461, 0.03488665771484375, 0.034902015686035154, 0.034925567626953126, 0.03493273544311523, 0.03489382553100586, 0.03492761611938477, 0.03490611267089844, 0.034933761596679686, 0.034915328979492184, 0.03494297790527344, 0.03492761611938477, 0.03495219039916992, 0.03493478393554687, 0.034947071075439456, 0.034976768493652347, 0.034961406707763674, 0.03493580627441406, 0.03496550369262695, 0.03506073760986328, 0.03498700714111328, 0.034976768493652347, 0.03499929428100586, 0.03498700714111328, 0.034971649169921876, 0.03496550369262695, 0.03503308868408203, 0.03502182388305664, 0.03499622344970703, 0.03500339126586914, 0.03501260757446289, 0.03503206253051758, 0.03502592086791992, 0.0350115852355957, 0.03503308868408203, 0.03503923034667969, 0.035019775390625, 0.03502489471435547, 0.03505152130126953, 0.03502489471435547]",tokens/s,27.844842569561877,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,1,1,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/1/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6693260f-1035c1c560a5526070a611b7;d77af3df-7531-49f8-8411-84c053c5a6c4) Repository Not Found for url: https://huggingface.co/1/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: 1 is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-7b,tiiuae/falcon-7b,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ self.load_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 250, in load_model_with_no_weights self.pretrained_model = self.automodel_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 559, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3704, in from_pretrained config = cls._autoset_attn_implementation( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1490, in _autoset_attn_implementation config = cls._check_and_enable_sdpa( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1656, in _check_and_enable_sdpa raise ValueError( ValueError: FalconForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,meta-llama/Llama-2-7b-hf,meta-llama/Llama-2-7b-hf,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.1,,,,MB,1227.173888,14563.147776,0.0,13916.700672,13738.740736,s,10,0.9368582153320313,0.09368582153320312,0.00029103446743837454,0.09357998657226563,0.09383599014282226,0.0941846996307373,0.09446366722106934,"[0.09453340911865235, 0.09357814025878906, 0.09357881927490234, 0.09349008178710938, 0.09358115386962891, 0.09375849914550781, 0.09352550506591797, 0.09354483032226563, 0.0936296615600586, 0.09363811492919921]",tokens/s,2732.537280566742,kWh,1.1091232609278441e-06,6.077465462512239e-07,5.8868764644298375e-06,7.603746271608905e-06,tokens/kWh,33667614.732998185,MB,1227.173888,14563.147776,0.0,13916.700672,13889.779712,s,10,22.537959716796877,2.2537959716796876,0.0008759771139935921,2.2536420898437504,2.254865625,2.254890454101562,2.2549103173828122,"[2.254915283203125, 2.254842041015625, 2.254397216796875, 2.253551513671875, 2.253302734375, 2.253732666015625, 2.252381103515625, 2.25331982421875, 2.2526572265625, 2.254860107421875]",tokens/s,27.952840803530215,kWh,2.6543364576678166e-05,1.4546689112517989e-05,0.0001127380239909712,0.00015382807768016734,tokens/kWh,409548.1198886647,,s,629,23.343781822204623,0.037112530718926215,0.011189030628809745,0.03576115036010742,0.03585863723754883,0.035902259826660156,0.12990176574707032,"[0.03586764907836914, 0.035659774780273434, 0.03563724899291992, 0.03562905502319336, 0.03574476623535156, 0.03574784088134766, 0.03612057495117187, 0.03586969757080078, 0.035714046478271484, 0.03576627349853516, 0.03570380783081055, 0.035688449859619144, 0.035714046478271484, 0.03577753448486328, 0.0360263671875, 0.03580825424194336, 0.035716094970703126, 0.03568742370605469, 0.03570687866210937, 0.0356864013671875, 0.035716094970703126, 0.03570892715454101, 0.035722240447998044, 0.03571916961669922, 0.035829761505126956, 0.035724288940429685, 0.035757057189941405, 0.035714046478271484, 0.03571814346313477, 0.03568947219848633, 0.03570073699951172, 0.03574169540405273, 0.03574476623535156, 0.035794944763183595, 0.03602534484863281, 0.03585945510864258, 0.03575807952880859, 0.03574169540405273, 0.03579289627075195, 0.035783679962158206, 0.03579084777832031, 0.03576422500610352, 0.03578675079345703, 0.035740673065185545, 0.035796993255615236, 0.03577753448486328, 0.03580723190307617, 0.03573657608032227, 0.03580928039550781, 0.03578060913085938, 0.035813377380371096, 0.0357918701171875, 0.03584511947631836, 0.03582668685913086, 0.035811328887939455, 0.035806209564208984, 0.03582361602783203, 0.03579904174804688, 0.03581644821166992, 0.035805248260498045, 0.03581228637695313, 0.0358287353515625, 0.1300193328857422, 0.0356495361328125, 0.03566387176513672, 0.03567411041259766, 0.03566796875, 0.03569049453735352, 0.035722240447998044, 0.03572531127929687, 0.03571507263183594, 0.03569356918334961, 0.03572531127929687, 0.03570278549194336, 0.03572531127929687, 0.035681278228759765, 0.035699710845947266, 0.03571814346313477, 0.03576115036010742, 0.03570175933837891, 0.03570687866210937, 0.035745792388916016, 0.03579084777832031, 0.03584307098388672, 0.035790878295898436, 0.03576726531982422, 0.035757057189941405, 0.0358021125793457, 0.035768318176269534, 0.03575193786621094, 0.035716094970703126, 0.035781631469726564, 0.035850238800048825, 0.035789825439453124, 0.03580928039550781, 0.035811328887939455, 0.03583488082885742, 0.03575807952880859, 0.03575500869750976, 0.03578675079345703, 0.035860481262207033, 0.035775489807128906, 0.03580825424194336, 0.03580313491821289, 0.035827713012695314, 0.03583180618286133, 0.035800064086914066, 0.0358205451965332, 0.03590348815917969, 0.035827713012695314, 0.03582259368896484, 0.035796993255615236, 0.0358021125793457, 0.035844097137451174, 0.035947521209716796, 0.03580313491821289, 0.035833854675292966, 0.035850238800048825, 0.035850238800048825, 0.03583180618286133, 0.035833854675292966, 0.035850238800048825, 0.03592396926879883, 0.03582876968383789, 0.03585327911376953, 0.12993740844726562, 0.03567411041259766, 0.03566796875, 0.03569049453735352, 0.03570691299438476, 0.03572323226928711, 0.03569561767578125, 0.0356577262878418, 0.03569260787963867, 0.03573241424560547, 0.03575296020507813, 0.03571507263183594, 0.03571507263183594, 0.03571200180053711, 0.03570892715454101, 0.035707904815673826, 0.03570892715454101, 0.03570483016967774, 0.03572019195556641, 0.03577958297729492, 0.03571814346313477, 0.03570892715454101, 0.03570175933837891, 0.03577241516113281, 0.03573555374145508, 0.03571916961669922, 0.03579391860961914, 0.035833854675292966, 0.035846145629882815, 0.035745792388916016, 0.03575193786621094, 0.03575603103637695, 0.035833854675292966, 0.03601408004760742, 0.0358666877746582, 0.03574470520019531, 0.03575091171264649, 0.03576422500610352, 0.03574272155761719, 0.03577036666870117, 0.03575398254394531, 0.03576627349853516, 0.03577036666870117, 0.03578265762329102, 0.035767295837402346, 0.03578572845458984, 0.03579391860961914, 0.03589324951171875, 0.0360079345703125, 0.03583180618286133, 0.035794944763183595, 0.0358809585571289, 0.03581644821166992, 0.035862529754638675, 0.035827713012695314, 0.03580825424194336, 0.0357918701171875, 0.035800064086914066, 0.03584000015258789, 0.03583078384399414, 0.0358809585571289, 0.035831871032714846, 0.03590444946289063, 0.12986883544921876, 0.035616737365722656, 0.03567411041259766, 0.0356577262878418, 0.035714046478271484, 0.035672065734863284, 0.0357918701171875, 0.03592297744750977, 0.035671009063720706, 0.03566592025756836, 0.0356864013671875, 0.035681278228759765, 0.03568537521362305, 0.035694591522216795, 0.03563315200805664, 0.03566899108886719, 0.03564652633666992, 0.03566483306884766, 0.035688449859619144, 0.035724288940429685, 0.03568947219848633, 0.0357212142944336, 0.035748863220214845, 0.03569561767578125, 0.03575807952880859, 0.0357314567565918, 0.03579084777832031, 0.035757057189941405, 0.03576934432983398, 0.03578572845458984, 0.03570483016967774, 0.035767295837402346, 0.035746814727783204, 0.035783679962158206, 0.03574476623535156, 0.03579289627075195, 0.03572633743286133, 0.035827713012695314, 0.03578265762329102, 0.03576422500610352, 0.03580313491821289, 0.035784702301025394, 0.035773441314697264, 0.03577446365356445, 0.03581951904296875, 0.03578572845458984, 0.03582156753540039, 0.0357918701171875, 0.03588198471069336, 0.035813377380371096, 0.03579084777832031, 0.03582361602783203, 0.03583795166015625, 0.03589120101928711, 0.03585126495361328, 0.035827713012695314, 0.03582566452026367, 0.03584000015258789, 0.03585740661621094, 0.035945472717285154, 0.0358021125793457, 0.03582668685913086, 0.03579084777832031, 0.12991897583007814, 0.03563014221191406, 0.035665855407714844, 0.03561471939086914, 0.03564137649536133, 0.0356607666015625, 0.03569868850708008, 0.0356495361328125, 0.035661823272705076, 0.03565670394897461, 0.03562496185302735, 0.03564851379394531, 0.03564851379394531, 0.03566080093383789, 0.03564031982421875, 0.03567001724243164, 0.03566080093383789, 0.03580928039550781, 0.03584000015258789, 0.03576934432983398, 0.035678207397460936, 0.03575296020507813, 0.03567103958129883, 0.03574784088134766, 0.03590041732788086, 0.03589017486572266, 0.035745792388916016, 0.035705856323242184, 0.03571916961669922, 0.035727359771728515, 0.03579391860961914, 0.03577036666870117, 0.035745792388916016, 0.0357386245727539, 0.03577756881713867, 0.03579388809204102, 0.03576319885253906, 0.03579289627075195, 0.035729408264160156, 0.03576422500610352, 0.035722240447998044, 0.03579391860961914, 0.03577446365356445, 0.03589017486572266, 0.035775489807128906, 0.03577958297729492, 0.03581644821166992, 0.035765247344970705, 0.03575091171264649, 0.035784702301025394, 0.03582566452026367, 0.03583795166015625, 0.03603046417236328, 0.035885055541992186, 0.03578572845458984, 0.035800064086914066, 0.0358021125793457, 0.03586764907836914, 0.03580928039550781, 0.035885055541992186, 0.035817470550537106, 0.03586150360107422, 0.03585638427734375, 0.13014016723632812, 0.03572326278686523, 0.035757057189941405, 0.03573964691162109, 0.03574169540405273, 0.03570380783081055, 0.03574476623535156, 0.03572326278686523, 0.035783679962158206, 0.03570380783081055, 0.035727359771728515, 0.035705856323242184, 0.0357509765625, 0.035719104766845707, 0.035765247344970705, 0.035729408264160156, 0.03580108642578125, 0.035705856323242184, 0.035762176513671876, 0.035732479095458985, 0.03574070358276367, 0.035756000518798826, 0.03571814346313477, 0.03570892715454101, 0.035705856323242184, 0.035773441314697264, 0.03571712112426758, 0.0357283821105957, 0.03573452758789063, 0.0357561264038086, 0.03573443222045898, 0.03576115036010742, 0.035730430603027344, 0.035774494171142576, 0.035808223724365235, 0.035746814727783204, 0.035757057189941405, 0.03579391860961914, 0.03576934432983398, 0.035765247344970705, 0.03576115036010742, 0.0358021125793457, 0.035762176513671876, 0.03578265762329102, 0.03574991989135742, 0.035791839599609375, 0.03585433578491211, 0.03579596710205078, 0.035776512145996094, 0.03581647872924805, 0.0357611198425293, 0.03580825424194336, 0.035798015594482424, 0.035813377380371096, 0.03578879928588867, 0.03580928039550781, 0.0358021125793457, 0.03584310531616211, 0.035839969635009766, 0.035811328887939455, 0.035817470550537106, 0.03582361602783203, 0.03579289627075195, 0.12990463256835938, 0.0356506233215332, 0.035687358856201175, 0.03567923355102539, 0.03564031982421875, 0.035659774780273434, 0.03563315200805664, 0.03566387176513672, 0.03566796875, 0.035675136566162106, 0.03567001724243164, 0.03566694259643555, 0.03574272155761719, 0.035707904815673826, 0.035768318176269534, 0.03571200180053711, 0.0356864013671875, 0.03568025588989258, 0.03568435287475586, 0.03568025588989258, 0.03568230438232422, 0.035678207397460936, 0.03569356918334961, 0.03570380783081055, 0.035705856323242184, 0.03570278549194336, 0.035730430603027344, 0.035726398468017576, 0.03575084686279297, 0.03573555374145508, 0.035743743896484374, 0.03572326278686523, 0.0357314567565918, 0.035732479095458985, 0.03575091171264649, 0.03575296020507813, 0.035748863220214845, 0.035748863220214845, 0.03574272155761719, 0.03575500869750976, 0.03572531127929687, 0.035765247344970705, 0.03576118469238281, 0.03576623916625977, 0.035776512145996094, 0.035824703216552733, 0.035798976898193356, 0.035760128021240234, 0.03577753448486328, 0.035767295837402346, 0.03585228729248047, 0.03611545562744141, 0.03584204864501953, 0.035789825439453124, 0.03579084777832031, 0.03580825424194336, 0.035789825439453124, 0.03578572845458984, 0.035794944763183595, 0.03581644821166992, 0.03580416107177734, 0.03579596710205078, 0.03581951904296875, 0.12989439392089844, 0.03562188720703125, 0.03564748764038086, 0.03563212966918945, 0.03562393569946289, 0.03563827133178711, 0.035656734466552736, 0.035691486358642575, 0.035783679962158206, 0.03569049453735352, 0.03566080093383789, 0.03566188812255859, 0.03566995239257813, 0.035683326721191407, 0.035724288940429685, 0.035707904815673826, 0.0357386245727539, 0.03568742370605469, 0.03565670394897461, 0.035688449859619144, 0.03569356918334961, 0.035714046478271484, 0.035692543029785154, 0.03579084777832031, 0.03620249557495117, 0.03587481689453125, 0.03575398254394531, 0.03571814346313477, 0.035746814727783204, 0.035743743896484374, 0.0357283821105957, 0.0357386245727539, 0.035796993255615236, 0.035737598419189456, 0.03574272155761719, 0.035737598419189456, 0.03576422500610352, 0.03577958297729492, 0.03575398254394531, 0.03578572845458984, 0.03574784088134766, 0.035773441314697264, 0.035727359771728515, 0.03574784088134766, 0.035768318176269534, 0.0358021125793457, 0.03579084777832031, 0.035784702301025394, 0.03580928039550781, 0.03581644821166992, 0.03584511947631836, 0.03580825424194336, 0.035885055541992186, 0.03588915252685547, 0.03582463836669922, 0.035813377380371096, 0.0357918701171875, 0.03584000015258789, 0.03590659332275391, 0.03582358551025391, 0.03580928039550781, 0.035827713012695314, 0.035827713012695314, 0.12999679565429687, 0.035642368316650394, 0.03561983871459961, 0.03564134216308594, 0.035688449859619144, 0.03566899108886719, 0.03565465545654297, 0.03566592025756836, 0.03563622283935547, 0.03564134216308594, 0.035681278228759765, 0.03565465545654297, 0.03562393569946289, 0.035694591522216795, 0.03562905502319336, 0.03568947219848633, 0.03563417434692383, 0.03572326278686523, 0.035675136566162106, 0.03567001724243164, 0.03563008117675781, 0.03570483016967774, 0.03573452758789063, 0.03569356918334961, 0.03568947219848633, 0.035737598419189456, 0.03567001724243164, 0.03570175933837891, 0.035737598419189456, 0.03578265762329102, 0.03580416107177734, 0.03574784088134766, 0.03581849670410156, 0.035730430603027344, 0.03571712112426758, 0.035740673065185545, 0.03577753448486328, 0.035716094970703126, 0.03574995040893555, 0.03573241424560547, 0.03574687957763672, 0.035853248596191406, 0.035999744415283204, 0.035773441314697264, 0.035724288940429685, 0.03578060913085938, 0.03577139282226562, 0.03581235122680664, 0.035783679962158206, 0.03579596710205078, 0.035781631469726564, 0.03585843276977539, 0.03577241516113281, 0.03578060913085938, 0.035781631469726564, 0.035768318176269534, 0.035970046997070314, 0.036127742767333985, 0.035866622924804685, 0.035844097137451174, 0.035888126373291016, 0.035888126373291016, 0.035798015594482424, 0.13012582397460937, 0.035737598419189456, 0.035760128021240234, 0.03591372680664062, 0.035920894622802735, 0.03570380783081055, 0.035705856323242184, 0.035727359771728515, 0.035714046478271484, 0.03569664001464844, 0.035681278228759765, 0.035730430603027344, 0.035714046478271484, 0.03570073699951172, 0.03569664001464844, 0.03573555374145508, 0.03568435287475586, 0.035724288940429685, 0.03569356918334961, 0.0357386245727539, 0.03566080093383789, 0.03571507263183594, 0.03567411041259766, 0.03570892715454101, 0.03569664001464844, 0.035740673065185545, 0.0358287353515625, 0.03573964691162109, 0.035716094970703126, 0.035765247344970705, 0.03575091171264649, 0.03582668685913086, 0.0357386245727539, 0.035748863220214845, 0.03573350524902344, 0.03575296020507813, 0.035806209564208984, 0.035807296752929686, 0.035908576965332034, 0.03578774261474609, 0.03573350524902344, 0.035768318176269534, 0.035748863220214845, 0.03582668685913086, 0.03581033706665039, 0.03578569412231445, 0.03577036666870117, 0.035844097137451174, 0.03585228729248047, 0.03586457443237305, 0.03585228729248047, 0.035862529754638675, 0.03586355209350586, 0.035871742248535156, 0.03582156753540039, 0.03584102249145508, 0.03602329635620117, 0.035904510498046875, 0.03587481689453125, 0.03587276840209961, 0.035814399719238284, 0.03588608169555664, 0.03582361602783203]",tokens/s,26.94507705695294,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-rw-1b,tiiuae/falcon-rw-1b,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ self.load_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 250, in load_model_with_no_weights self.pretrained_model = self.automodel_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 559, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3704, in from_pretrained config = cls._autoset_attn_implementation( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1490, in _autoset_attn_implementation config = cls._check_and_enable_sdpa( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1656, in _check_and_enable_sdpa raise ValueError( ValueError: FalconForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,8,8,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/8/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66932440-2945d1b75858294e40744279;08a06f95-02b2-4b4a-9c56-206dcc7fe394) Repository Not Found for url: https://huggingface.co/8/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: 8 is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-4B,,cuda,0,42,,,,,True,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,217063f5c507ed7cc255df7e1f64c4333a0b4dfe,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,,MB,1290.756096,10129.768448,0.0,9483.321344,8972.087296,s,10,0.4185474510192871,0.04185474510192871,0.00026355926287794073,0.04174934387207031,0.04193587455749512,0.042284225273132324,0.04256290584564209,"[0.04263257598876953, 0.04171980667114258, 0.04173174285888672, 0.041715999603271485, 0.041858463287353515, 0.04182236862182617, 0.04174137496948242, 0.04175606536865235, 0.04174262237548828, 0.041826431274414065]",tokens/s,6116.3913285474355,kWh,4.952807576919034e-07,2.7139040796108354e-07,2.817547581790843e-06,3.58421874744383e-06,tokens/kWh,71424212.09156735,MB,1290.756096,10129.768448,0.0,9483.321344,9040.66816,s,10,20.668836547851566,2.0668836547851566,0.029378578381633064,2.079330078125,2.08822685546875,2.089942041015625,2.091314189453125,"[2.0848232421875, 2.067904296875, 2.081076416015625, 2.086474365234375, 2.070119140625, 2.077583740234375, 2.087845703125, 2.0916572265625, 2.0213123779296875, 2.0000400390625]",tokens/s,30.480670672558283,kWh,2.3955011945486294e-05,1.3127950983070979e-05,7.515879257700655e-05,0.00011224175550556382,tokens/kWh,561288.4413312397,,s,629,21.010035703659028,0.033402282517740946,0.005035072653597517,0.03315097427368164,0.03340987014770508,0.03374016647338867,0.07439762329101564,"[0.03436646270751953, 0.03327078247070313, 0.03454873657226563, 0.03412275314331055, 0.034154495239257815, 0.03465318298339844, 0.03433779144287109, 0.03325235366821289, 0.0340398063659668, 0.03427840042114258, 0.034767871856689454, 0.033584129333496096, 0.032748542785644534, 0.03382374572753906, 0.03416476821899414, 0.033783775329589844, 0.033255424499511715, 0.0333496322631836, 0.03317452621459961, 0.03333017730712891, 0.03315097427368164, 0.03203788757324219, 0.031921152114868165, 0.032102401733398435, 0.0319866886138916, 0.0330618896484375, 0.03199385643005371, 0.033235969543457033, 0.032851966857910156, 0.03199897575378418, 0.031900672912597655, 0.03203788757324219, 0.032290817260742184, 0.033124351501464845, 0.0331673583984375, 0.03333222579956055, 0.03309260940551758, 0.03211264038085938, 0.03197952079772949, 0.031953920364379884, 0.0319836483001709, 0.03284681701660156, 0.033154048919677735, 0.03329232025146484, 0.03308553695678711, 0.03326348876953125, 0.033122303009033204, 0.033285118103027346, 0.03312332916259766, 0.03337011337280273, 0.033132545471191405, 0.03337420654296875, 0.03311820983886719, 0.03310899353027344, 0.03311001586914063, 0.03355340957641602, 0.033296382904052735, 0.032075775146484374, 0.03204095840454101, 0.03204403305053711, 0.032119838714599606, 0.03200815963745117, 0.07428300476074219, 0.03204915237426758, 0.03276287841796875, 0.03362815856933594, 0.03222118377685547, 0.032, 0.03194879913330078, 0.03199590492248535, 0.03199897575378418, 0.03198259162902832, 0.032189441680908204, 0.03191193580627441, 0.032487422943115234, 0.03319193649291992, 0.03322470474243164, 0.033189888000488284, 0.03314688110351562, 0.03324006271362305, 0.03316633605957031, 0.03308748626708984, 0.033235969543457033, 0.03325132751464844, 0.033219585418701174, 0.0333834228515625, 0.0333383674621582, 0.033081344604492184, 0.03285504150390625, 0.033154048919677735, 0.032863231658935545, 0.032949249267578126, 0.03259187316894531, 0.032626686096191404, 0.03304447937011719, 0.033261566162109374, 0.033291263580322264, 0.03318067169189453, 0.03317657470703125, 0.033205249786376956, 0.03212595367431641, 0.03206553649902344, 0.032, 0.03197337532043457, 0.03213824081420898, 0.03300249481201172, 0.032849918365478514, 0.03328409576416016, 0.0328908805847168, 0.032833534240722655, 0.032775169372558595, 0.0326297607421875, 0.033107967376708985, 0.033111038208007815, 0.033157119750976564, 0.03263488006591797, 0.03330047988891602, 0.03333324813842774, 0.03289907073974609, 0.03249356842041016, 0.03288063812255859, 0.033188865661621096, 0.03322367858886719, 0.033209342956542966, 0.03313356781005859, 0.0757422103881836, 0.033228832244873045, 0.03325027084350586, 0.03324313735961914, 0.03297792053222656, 0.033219585418701174, 0.033116161346435545, 0.03313868713378906, 0.033132545471191405, 0.03323292922973633, 0.03325641632080078, 0.03317452621459961, 0.03322470474243164, 0.03322060775756836, 0.033170433044433595, 0.03319193649291992, 0.03322371292114258, 0.03207779312133789, 0.03199897575378418, 0.03176652717590332, 0.03352268981933594, 0.03369676971435547, 0.03322060775756836, 0.03328204727172852, 0.03332710266113281, 0.033170433044433595, 0.03317964935302734, 0.032075775146484374, 0.03204608154296875, 0.032715774536132815, 0.03324620819091797, 0.0332861442565918, 0.032979969024658204, 0.03318783950805664, 0.03312844848632813, 0.033231937408447265, 0.033207233428955076, 0.033203201293945314, 0.03307827377319336, 0.03339980697631836, 0.03396198272705078, 0.03347251129150391, 0.03326566314697266, 0.03314995193481445, 0.032048126220703126, 0.03250995254516602, 0.03348582458496094, 0.03334143829345703, 0.03323904037475586, 0.03314585494995117, 0.033413120269775394, 0.033347583770751955, 0.032043006896972655, 0.032045055389404296, 0.032121856689453124, 0.03208703994750976, 0.03203788757324219, 0.033339393615722655, 0.03300044631958008, 0.03324518585205078, 0.03322367858886719, 0.03329539108276367, 0.03322876739501953, 0.07429737854003907, 0.03316630554199219, 0.033212448120117186, 0.032016353607177736, 0.03205017471313477, 0.032026622772216795, 0.032086017608642575, 0.03179110336303711, 0.03310182571411133, 0.033209342956542966, 0.03326464080810547, 0.03341516876220703, 0.03340902328491211, 0.03337625503540039, 0.03327897644042969, 0.03294412612915039, 0.033157119750976564, 0.03340595245361328, 0.0332042236328125, 0.03326873779296875, 0.03316633605957031, 0.03326464080810547, 0.03314790344238281, 0.0332564468383789, 0.0334090576171875, 0.033300449371337894, 0.03326566314697266, 0.033172481536865236, 0.033188865661621096, 0.033263614654541016, 0.033160190582275394, 0.033347583770751955, 0.033127422332763674, 0.03317145538330078, 0.0332492790222168, 0.03335782241821289, 0.03334348678588867, 0.03325439834594727, 0.03291033554077148, 0.03325747299194336, 0.0334919662475586, 0.033225727081298825, 0.03326668930053711, 0.03328409576416016, 0.03324415969848633, 0.03332505416870117, 0.032331775665283204, 0.03325337600708008, 0.0333496322631836, 0.03322163009643555, 0.033909759521484374, 0.033544193267822264, 0.03201740646362305, 0.03291852951049805, 0.03319705581665039, 0.033432575225830076, 0.033274879455566404, 0.03340390396118164, 0.033259521484375, 0.03321855926513672, 0.03321651077270508, 0.0333271369934082, 0.0332072639465332, 0.07446630096435547, 0.03224883270263672, 0.032099327087402346, 0.03196313667297363, 0.03209011077880859, 0.032086017608642575, 0.03194367980957031, 0.031898624420166014, 0.033124351501464845, 0.033255424499511715, 0.033377281188964845, 0.03313459014892578, 0.03216998291015625, 0.03340390396118164, 0.03347251129150391, 0.0334919662475586, 0.033355777740478515, 0.033205249786376956, 0.0332677116394043, 0.033331199645996096, 0.03319193649291992, 0.03325747299194336, 0.033307647705078124, 0.033173503875732424, 0.03314790344238281, 0.03198873519897461, 0.03201638412475586, 0.03196928024291992, 0.03211980819702148, 0.03206758499145508, 0.032102401733398435, 0.03215769577026367, 0.03209625625610352, 0.032080894470214845, 0.03218022537231445, 0.03209529495239258, 0.03237984085083008, 0.03227340698242188, 0.032919551849365236, 0.03329945755004883, 0.033500160217285156, 0.03320115280151367, 0.03326259231567383, 0.03316326522827148, 0.03323392105102539, 0.03325439834594727, 0.03291852951049805, 0.03331686401367188, 0.03341823959350586, 0.03274444961547852, 0.03343360137939453, 0.03319910430908203, 0.03335372924804687, 0.03331891250610351, 0.03329433441162109, 0.033205249786376956, 0.03310182571411133, 0.03326566314697266, 0.03323187255859375, 0.03319705581665039, 0.03327283096313476, 0.03322163009643555, 0.03366912078857422, 0.07443660736083985, 0.03198054313659668, 0.032008190155029294, 0.03196931266784668, 0.03191190338134765, 0.03202867126464844, 0.03271680068969727, 0.0332564468383789, 0.03338751983642578, 0.03325747299194336, 0.03331590270996094, 0.03324409484863281, 0.0328950080871582, 0.033342430114746094, 0.033181697845458984, 0.0332861442565918, 0.033546241760253906, 0.03204403305053711, 0.032438270568847655, 0.03320115280151367, 0.033165313720703124, 0.033143806457519534, 0.03312025451660156, 0.03313459014892578, 0.0332677116394043, 0.033175552368164066, 0.03258572769165039, 0.03326873779296875, 0.033298431396484376, 0.03211980819702148, 0.0321003532409668, 0.03203481674194336, 0.03213516616821289, 0.03231846237182617, 0.032126976013183595, 0.03201228713989258, 0.032010238647460935, 0.03321548843383789, 0.03329945755004883, 0.03334656143188477, 0.03334143829345703, 0.033260543823242186, 0.033363967895507815, 0.03335168075561523, 0.033259521484375, 0.03318272018432617, 0.03332403182983398, 0.03323289489746094, 0.0331673583984375, 0.03371212768554688, 0.03372851181030274, 0.033363967895507815, 0.0332677116394043, 0.0328837776184082, 0.03335782241821289, 0.033311710357666016, 0.033747936248779295, 0.03338137435913086, 0.033377281188964845, 0.03335270309448242, 0.03328921508789062, 0.033442817687988284, 0.03343462371826172, 0.0755445785522461, 0.03338444900512695, 0.03315718460083008, 0.03272083282470703, 0.03316428756713867, 0.03339571380615235, 0.033242111206054685, 0.03336191940307617, 0.03320115280151367, 0.032546817779541014, 0.03316223907470703, 0.032246784210205076, 0.033113086700439456, 0.03319398498535156, 0.03322982406616211, 0.03322163009643555, 0.03319193649291992, 0.03322777557373047, 0.03322367858886719, 0.03315302276611328, 0.03317657470703125, 0.03316124725341797, 0.03321238327026367, 0.032625663757324216, 0.03324518585205078, 0.0331929931640625, 0.0331437759399414, 0.03355750274658203, 0.033188865661621096, 0.033285118103027346, 0.03277926254272461, 0.03316844940185547, 0.03324415969848633, 0.033103809356689454, 0.03220377731323242, 0.032284671783447266, 0.03313663864135742, 0.03325132751464844, 0.03317452621459961, 0.03313875198364258, 0.03311404800415039, 0.033105918884277344, 0.033377281188964845, 0.03314176177978516, 0.03314176177978516, 0.0332042236328125, 0.03335075378417969, 0.0332143669128418, 0.03320217514038086, 0.03312332916259766, 0.03328307342529297, 0.03311820983886719, 0.03315814590454102, 0.03324006271362305, 0.03326566314697266, 0.03322982406616211, 0.033102848052978515, 0.03321139144897461, 0.0331673583984375, 0.03312332916259766, 0.03326668930053711, 0.03319500732421875, 0.033312767028808594, 0.07551487731933594, 0.03368960189819336, 0.033721343994140625, 0.03336294555664063, 0.033235969543457033, 0.03315302276611328, 0.03338649749755859, 0.03318783950805664, 0.033446910858154294, 0.03319091033935547, 0.033309696197509765, 0.03312025451660156, 0.03314179229736328, 0.033258529663085935, 0.03340591812133789, 0.03320624160766601, 0.03329536056518555, 0.033172481536865236, 0.03329740905761719, 0.0330967025756836, 0.03337011337280273, 0.03489996719360351, 0.03385958480834961, 0.03330252838134766, 0.0335022087097168, 0.033430526733398434, 0.03335987091064453, 0.033963008880615236, 0.03410943984985351, 0.032249855041503905, 0.032261119842529294, 0.032817150115966795, 0.03314176177978516, 0.03206655883789063, 0.03203788757324219, 0.03201331329345703, 0.03194367980957031, 0.03304550552368164, 0.0331776008605957, 0.033148929595947264, 0.03319091033935547, 0.03304959869384766, 0.033157119750976564, 0.03298611068725586, 0.03308544158935547, 0.032626686096191404, 0.03200307083129883, 0.03321139144897461, 0.03322470474243164, 0.03313459014892578, 0.03320115280151367, 0.03347558212280274, 0.03328102493286133, 0.0332872314453125, 0.03326252746582031, 0.03331174468994141, 0.03413094329833984, 0.03375001525878906, 0.03327078247070313, 0.033288192749023435, 0.033185791015625, 0.03320012664794922, 0.03325235366821289, 0.07448780822753906, 0.032089088439941404, 0.032043041229248045, 0.03203683090209961, 0.034157569885253904, 0.0336732177734375, 0.03428768157958984, 0.03340588760375977, 0.032369663238525394, 0.03200921630859375, 0.032094207763671875, 0.03193548774719238, 0.03218431854248047, 0.03201740646362305, 0.03202969741821289, 0.03194777679443359, 0.03194572830200195, 0.031788063049316403, 0.032191455841064455, 0.03201228713989258, 0.032045055389404296, 0.03203379058837891, 0.03211161422729492, 0.03199283218383789, 0.03198566436767578, 0.03219558334350586, 0.032421886444091795, 0.03290828704833984, 0.03200102233886719, 0.03205017471313477, 0.03209830474853516, 0.03204095840454101, 0.03172352027893066, 0.03197747230529785, 0.032064510345458985, 0.03163750457763672, 0.031916032791137694, 0.03201945495605469, 0.03200307083129883, 0.031631359100341795, 0.031677440643310545, 0.0316364803314209, 0.03211264038085938, 0.03177369689941406, 0.03172659111022949, 0.0318023681640625, 0.03213312149047851, 0.03198259162902832, 0.03172147178649903, 0.031697919845581055, 0.031704063415527346, 0.0320552978515625, 0.03160268783569336, 0.03163955116271973, 0.0316682243347168, 0.03165388870239258, 0.03165286445617676, 0.03202252960205078, 0.031747072219848634, 0.03199283218383789, 0.03194879913330078, 0.03181875228881836, 0.03170201683044434, 0.0744632339477539, 0.03174604797363281, 0.0316866569519043, 0.031632383346557616, 0.031579135894775394, 0.031661056518554685, 0.031752191543579104, 0.031676416397094724, 0.0316231689453125, 0.03168870353698731, 0.031689760208129886, 0.031696863174438476, 0.03158732795715332, 0.0315996150970459, 0.03169177627563476, 0.031604736328125, 0.03157196807861328, 0.03178803253173828, 0.03202969741821289, 0.03196416091918945, 0.03194470405578613, 0.03197238349914551, 0.03190985679626465, 0.031749120712280275, 0.03162521553039551, 0.032216064453125, 0.032008190155029294, 0.03175635147094726, 0.03226105499267578, 0.03185868835449219, 0.03172454452514648, 0.03180441665649414, 0.03156172752380371, 0.03164672088623047, 0.03160883140563965, 0.031562751770019534, 0.03160883140563965, 0.03164159965515137, 0.031835136413574217, 0.03164262390136719, 0.03163443183898926, 0.031899648666381834, 0.031801343917846676, 0.031647743225097655, 0.03156787109375, 0.03167948722839355, 0.03161292839050293, 0.031643648147583005, 0.03162623977661133, 0.031697919845581055, 0.03167231941223145, 0.03163955116271973, 0.03165081596374512, 0.03171225547790527, 0.03177984046936035, 0.03164262390136719, 0.03167436790466309, 0.03164672088623047, 0.03177574348449707, 0.03207987213134766, 0.03177369689941406, 0.03167436790466309, 0.031719423294067385]",tokens/s,29.9380738268072,,,main,False,False,,,,,,,,,,,,,,,,,,,,,,, bfloat16-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,internlm/internlm2-20b,internlm/internlm2-20b,cuda,0,42,,,True,,,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",internlm2,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,i,i,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/i/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669474c7-11c53e2b0a32eabf7141f0c3;61e2f5e0-389a-4406-8dc9-85d494e86ef4) Repository Not Found for url: https://huggingface.co/i/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: i is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,stabilityai/stablelm-3b-4e1t,stabilityai/stablelm-3b-4e1t,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1248.722944,6715.604992,0.0,6069.157888,5763.863552,s,10,0.3015369281768798,0.03015369281768799,0.00029932930486819073,0.030029088020324706,0.030538534164428713,0.03073526668548584,0.030892652702331543,"[0.03093199920654297, 0.030049375534057617, 0.030092960357666017, 0.029983840942382812, 0.02999942398071289, 0.029926080703735352, 0.030000255584716796, 0.0300183048248291, 0.030039871215820312, 0.030494815826416017]",tokens/s,8489.839090283225,kWh,3.568219259186214e-07,1.9552162898641277e-07,2.169461285752255e-06,2.7218048406572892e-06,tokens/kWh,94055237.23669273,MB,1248.722944,6736.576512,0.0,6090.129408,5863.102976,s,10,14.880395874023437,1.4880395874023438,0.012550265247513353,1.4854010620117188,1.507277880859375,1.507602197265625,1.507861650390625,"[1.507205810546875, 1.4989151611328124, 1.487388427734375, 1.49186767578125, 1.4719144287109376, 1.4717625732421875, 1.507926513671875, 1.4812845458984376, 1.478717041015625, 1.4834136962890625]",tokens/s,42.33758330984896,kWh,1.8004725348803372e-05,9.86657931815692e-06,5.9096726616047514e-05,8.696803128300781e-05,tokens/kWh,724404.1180487112,,s,629,15.125405706405617,0.02404674993069259,0.003655537751274818,0.02347315216064453,0.024345395278930662,0.024510873794555663,0.05393481719970703,"[0.02405887985229492, 0.023488512039184572, 0.023574527740478517, 0.023568416595458986, 0.02432611274719238, 0.02532966423034668, 0.02715648078918457, 0.0249036808013916, 0.024391679763793944, 0.02426470375061035, 0.024376319885253905, 0.02433433532714844, 0.024373247146606446, 0.024376319885253905, 0.024285184860229493, 0.023608320236206053, 0.02370047950744629, 0.023575584411621095, 0.023516128540039063, 0.023569408416748046, 0.023540735244750977, 0.023529504776000975, 0.024470560073852538, 0.02481350326538086, 0.02452787208557129, 0.024437759399414064, 0.024422399520874022, 0.02447871971130371, 0.024201215744018553, 0.024164352416992187, 0.023334911346435547, 0.023409664154052736, 0.023574527740478517, 0.023645183563232423, 0.02338921546936035, 0.02358268737792969, 0.02330624008178711, 0.023407615661621094, 0.02332467269897461, 0.02328166389465332, 0.02330009651184082, 0.023775232315063476, 0.023933952331542968, 0.02391142463684082, 0.023969791412353517, 0.023642112731933593, 0.02350284767150879, 0.023556095123291015, 0.02449612808227539, 0.02407731246948242, 0.023613439559936524, 0.023372800827026367, 0.02347110366821289, 0.023564287185668945, 0.02344550323486328, 0.023302143096923827, 0.023254016876220703, 0.023396415710449217, 0.023318464279174805, 0.023591936111450194, 0.023558143615722657, 0.023635967254638672, 0.05416447830200195, 0.023436288833618164, 0.023521280288696288, 0.02353049659729004, 0.023374879837036133, 0.023309280395507812, 0.023406591415405274, 0.02332467269897461, 0.02332159996032715, 0.023822336196899413, 0.025536512374877928, 0.02478489685058594, 0.024374336242675782, 0.024349632263183593, 0.02428313636779785, 0.024457216262817383, 0.024086528778076172, 0.02452070426940918, 0.023537727355957033, 0.023306175231933592, 0.023609344482421874, 0.023345151901245118, 0.023262208938598632, 0.02332876777648926, 0.02324684715270996, 0.02331340789794922, 0.023282688140869142, 0.02333695983886719, 0.023249919891357423, 0.023860223770141603, 0.023302143096923827, 0.023397375106811523, 0.023331840515136718, 0.02328883171081543, 0.02333695983886719, 0.02325606346130371, 0.023254016876220703, 0.02368409538269043, 0.024555519104003908, 0.02434048080444336, 0.02477568054199219, 0.024549375534057616, 0.024353792190551758, 0.02437222480773926, 0.024452096939086915, 0.024352767944335937, 0.02413670349121094, 0.023988224029541014, 0.024000511169433594, 0.02392166328430176, 0.024070144653320313, 0.023979007720947267, 0.02429644775390625, 0.024370176315307617, 0.024357887268066408, 0.023441440582275392, 0.02338915252685547, 0.023348224639892577, 0.023347200393676756, 0.02332159996032715, 0.02332467269897461, 0.02327859115600586, 0.02354380798339844, 0.053994495391845705, 0.023410688400268553, 0.023232511520385742, 0.02326425552368164, 0.023249919891357423, 0.023358463287353515, 0.023248895645141602, 0.02329497528076172, 0.023446527481079102, 0.023357440948486328, 0.023419904708862304, 0.023508991241455078, 0.024246271133422852, 0.02407935905456543, 0.023417919158935548, 0.02404243278503418, 0.024442880630493165, 0.024069120407104492, 0.024037376403808593, 0.02405068778991699, 0.024060928344726562, 0.023391231536865235, 0.025050111770629883, 0.0250644474029541, 0.024574975967407226, 0.024395776748657227, 0.024415231704711913, 0.023508991241455078, 0.02395955276489258, 0.02384998321533203, 0.023625728607177734, 0.023571456909179687, 0.023545856475830077, 0.02332979202270508, 0.02330009651184082, 0.02327039909362793, 0.02327347183227539, 0.023443456649780273, 0.02330931282043457, 0.02327654457092285, 0.023388160705566406, 0.023570432662963867, 0.023330816268920897, 0.023340032577514647, 0.023376895904541017, 0.02326016044616699, 0.023248960494995117, 0.023310304641723633, 0.02332259178161621, 0.023371776580810546, 0.023323648452758788, 0.023384063720703126, 0.023301151275634764, 0.023423967361450197, 0.023412736892700195, 0.02333798408508301, 0.023400447845458985, 0.02335443115234375, 0.023664575576782226, 0.023591936111450194, 0.023573503494262696, 0.023375871658325196, 0.023406591415405274, 0.0538787841796875, 0.023500799179077148, 0.023626752853393555, 0.0232857608795166, 0.023388160705566406, 0.023395328521728515, 0.023365663528442382, 0.02331439971923828, 0.023302143096923827, 0.02332262420654297, 0.023319551467895508, 0.02367487907409668, 0.02347417640686035, 0.023394304275512694, 0.02325606346130371, 0.02355200004577637, 0.02347929573059082, 0.023596031188964844, 0.023449632644653322, 0.023644128799438478, 0.02352332878112793, 0.023817216873168946, 0.02352230453491211, 0.023580671310424805, 0.023237632751464843, 0.023428096771240234, 0.02332979202270508, 0.02335436820983887, 0.02332979202270508, 0.023585792541503905, 0.023622655868530275, 0.0233123836517334, 0.023319551467895508, 0.023241727828979493, 0.023430143356323242, 0.0235284481048584, 0.023541759490966797, 0.023319551467895508, 0.023438335418701172, 0.023584768295288085, 0.02347315216064453, 0.02328166389465332, 0.023373823165893554, 0.024606719970703125, 0.025400320053100587, 0.024451072692871095, 0.02425753593444824, 0.024394752502441407, 0.02424831962585449, 0.024389631271362306, 0.024344575881958007, 0.02430156707763672, 0.024012800216674804, 0.023966720581054687, 0.024081407546997072, 0.024374271392822267, 0.02438860893249512, 0.02434867286682129, 0.02424015998840332, 0.02334511947631836, 0.023198720932006835, 0.023568384170532225, 0.023609344482421874, 0.054042686462402345, 0.023474111557006835, 0.023588863372802735, 0.023438335418701172, 0.023442432403564452, 0.023414783477783203, 0.02329599952697754, 0.023128063201904296, 0.023218175888061524, 0.023214080810546874, 0.023257087707519532, 0.02325196838378906, 0.023198720932006835, 0.02327654457092285, 0.0233175048828125, 0.023391231536865235, 0.023422975540161133, 0.02327142333984375, 0.023147520065307618, 0.023226367950439454, 0.023243776321411135, 0.023209983825683594, 0.02349260711669922, 0.023622655868530275, 0.023383039474487305, 0.02344960021972656, 0.02350387191772461, 0.02345471954345703, 0.02350182342529297, 0.02351103973388672, 0.023360511779785157, 0.023472127914428712, 0.02346598434448242, 0.02345471954345703, 0.02348953628540039, 0.023564287185668945, 0.023428096771240234, 0.02345881652832031, 0.02349465560913086, 0.023377920150756838, 0.02328371238708496, 0.02329190444946289, 0.02324787139892578, 0.023196672439575194, 0.02326016044616699, 0.02326323127746582, 0.023206911087036132, 0.02328678321838379, 0.02304921531677246, 0.023795711517333985, 0.02386227226257324, 0.023135263442993163, 0.023203807830810545, 0.023230464935302734, 0.02313523292541504, 0.023315456390380858, 0.02329190444946289, 0.02332057571411133, 0.023342079162597656, 0.02334726333618164, 0.02318636894226074, 0.02347315216064453, 0.023245824813842773, 0.05386240005493164, 0.02328473663330078, 0.02332876777648926, 0.023245824813842773, 0.023373823165893554, 0.023197696685791015, 0.023205888748168944, 0.023217151641845703, 0.023174144744873046, 0.023222272872924804, 0.023348224639892577, 0.023244800567626952, 0.02390630340576172, 0.02372096061706543, 0.02350387191772461, 0.023554048538208007, 0.023597055435180665, 0.023584768295288085, 0.023442432403564452, 0.023340032577514647, 0.02327347183227539, 0.02323967933654785, 0.02326118469238281, 0.02331648063659668, 0.023266304016113282, 0.023237632751464843, 0.023249919891357423, 0.023147520065307618, 0.023612415313720703, 0.02371993637084961, 0.02327654457092285, 0.02327859115600586, 0.02323967933654785, 0.023198720932006835, 0.023229440689086913, 0.02332876777648926, 0.023397439956665038, 0.023549888610839845, 0.02349260711669922, 0.02348646354675293, 0.02369945526123047, 0.023477279663085937, 0.02350588798522949, 0.02329804801940918, 0.02320899200439453, 0.023221216201782226, 0.023351295471191406, 0.02327039909362793, 0.023198720932006835, 0.02330931282043457, 0.02326118469238281, 0.023194623947143556, 0.023262208938598632, 0.02334617614746094, 0.023282688140869142, 0.02347007942199707, 0.02327039909362793, 0.023292959213256834, 0.02334102439880371, 0.023329856872558594, 0.023330751419067382, 0.023334911346435547, 0.023334911346435547, 0.054063102722167966, 0.02329599952697754, 0.027489280700683592, 0.023642112731933593, 0.02323967933654785, 0.023605247497558594, 0.02347007942199707, 0.023794687271118165, 0.023441408157348635, 0.0235100154876709, 0.02346291160583496, 0.023542783737182618, 0.023931903839111326, 0.024778751373291014, 0.02372812843322754, 0.02390630340576172, 0.024247295379638673, 0.02453196716308594, 0.02411008071899414, 0.024204288482666016, 0.024252416610717774, 0.024220672607421875, 0.02408550453186035, 0.024116256713867187, 0.02392470359802246, 0.024040447235107423, 0.02411622428894043, 0.02416640090942383, 0.024030208587646484, 0.024122367858886717, 0.024040447235107423, 0.024396799087524415, 0.024262655258178712, 0.024193023681640623, 0.02409676742553711, 0.024022016525268555, 0.023996416091918944, 0.024031232833862305, 0.024155136108398437, 0.024204288482666016, 0.024242176055908202, 0.024185855865478514, 0.02384588813781738, 0.024396799087524415, 0.02502348709106445, 0.024414207458496092, 0.023544832229614256, 0.02348953628540039, 0.023443456649780273, 0.023587839126586914, 0.023662591934204103, 0.02323148727416992, 0.023383039474487305, 0.02344550323486328, 0.023151615142822265, 0.023538688659667968, 0.023516159057617187, 0.02351411247253418, 0.023610368728637695, 0.023657472610473632, 0.023517183303833008, 0.023595008850097656, 0.02348134422302246, 0.054089729309082034, 0.02324684715270996, 0.023299072265625, 0.023143423080444335, 0.023211008071899415, 0.023327743530273438, 0.023154687881469727, 0.023166976928710937, 0.023370752334594725, 0.02326425552368164, 0.02326937675476074, 0.02327654457092285, 0.023266304016113282, 0.023174144744873046, 0.02325196838378906, 0.02318547248840332, 0.023305152893066405, 0.02352025604248047, 0.02344960021972656, 0.023446527481079102, 0.02322329521179199, 0.023250944137573244, 0.02346598434448242, 0.023645183563232423, 0.023222272872924804, 0.023244800567626952, 0.023230464935302734, 0.023389184951782226, 0.023641088485717773, 0.02353152084350586, 0.023143423080444335, 0.023421951293945312, 0.023439359664916993, 0.023348224639892577, 0.023554048538208007, 0.023541759490966797, 0.02332057571411133, 0.023576576232910155, 0.023299072265625, 0.02330112075805664, 0.02348543930053711, 0.023617536544799804, 0.02349158477783203, 0.023600128173828124, 0.02322329521179199, 0.02327859115600586, 0.023430143356323242, 0.023555072784423828, 0.025754623413085938, 0.02454015922546387, 0.023628799438476563, 0.023800832748413086, 0.023954431533813478, 0.023345151901245118, 0.02349567985534668, 0.023776256561279296, 0.023582719802856447, 0.0236943359375, 0.023876672744750978, 0.02378848075866699, 0.02447667121887207, 0.02409062385559082, 0.023642112731933593, 0.053956607818603515, 0.02349465560913086, 0.02352742385864258, 0.023399423599243165, 0.023340032577514647, 0.023244800567626952, 0.02325299263000488, 0.023229440689086913, 0.023195648193359376, 0.023220224380493162, 0.023203840255737306, 0.02323148727416992, 0.023923776626586915, 0.02353555107116699, 0.02353152084350586, 0.02393907165527344, 0.023735296249389647, 0.0236759033203125, 0.02353971290588379, 0.023271455764770507, 0.023575519561767577, 0.02329599952697754, 0.02326323127746582, 0.02327244758605957, 0.02349158477783203, 0.02329292869567871, 0.023435264587402343, 0.023362560272216795, 0.02328473663330078, 0.023315456390380858, 0.023630847930908205, 0.023582719802856447, 0.023395328521728515, 0.023216127395629883, 0.023259136199951173, 0.023347200393676756, 0.023339008331298827, 0.02357254409790039, 0.023504831314086913, 0.023665664672851562, 0.023319551467895508, 0.02348134422302246, 0.02330419158935547, 0.023545856475830077, 0.02348646354675293, 0.023399456024169922, 0.023434207916259765, 0.02347315216064453, 0.02322329521179199, 0.023243776321411135, 0.023284767150878908, 0.023274463653564455, 0.02330931282043457, 0.023368703842163087, 0.023541759490966797, 0.023606271743774415, 0.023504896163940428, 0.02347929573059082, 0.023593984603881835, 0.024039424896240235, 0.023569408416748046, 0.023673856735229492, 0.02495078468322754, 0.055223297119140625, 0.0233123836517334, 0.023559167861938478, 0.023833599090576172, 0.02348953628540039, 0.023440383911132814, 0.02351513671875, 0.023631872177124022, 0.023561248779296873, 0.023585760116577148, 0.02350492858886719, 0.023351264953613282, 0.023340032577514647, 0.023603200912475586, 0.023581695556640626, 0.02346905517578125, 0.023310367584228515, 0.023294944763183594, 0.02324787139892578, 0.023362560272216795, 0.02347110366821289, 0.02329497528076172, 0.023525375366210938, 0.02352230453491211, 0.02330112075805664, 0.023375871658325196, 0.023573503494262696, 0.023245824813842773, 0.023340032577514647, 0.023517215728759765, 0.023287776947021485, 0.02330828857421875, 0.02330624008178711, 0.023837696075439452, 0.02346700859069824, 0.023440383911132814, 0.02369843292236328, 0.02351820755004883, 0.02355200004577637, 0.0233175048828125, 0.0233175048828125, 0.02366464042663574, 0.02345881652832031, 0.02364419174194336, 0.023471071243286134, 0.02390425682067871, 0.0239554557800293, 0.023352319717407227, 0.02324787139892578, 0.023352319717407227, 0.023342079162597656, 0.02345267105102539, 0.023432191848754884, 0.02371993637084961, 0.023976959228515626, 0.023941120147705077, 0.02353971290588379, 0.023971839904785155, 0.023639039993286134, 0.02390323257446289, 0.0239554557800293, 0.023649280548095702, 0.02406399917602539]",tokens/s,41.58566138385414,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,huggyllama/llama-13b,huggyllama/llama-13b,cuda,0,42,,,True,,,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",llama,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-2.7b,facebook/opt-2.7b,cuda,0,42,,,True,,,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",opt,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,microsoft/phi-1_5,microsoft/phi-1_5,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1281.302528,3855.089664,0.0,3208.64256,2982.448128,s,10,0.2068157157897949,0.020681571578979492,0.00025223375901512455,0.020605776786804197,0.020716234207153322,0.02107498073577881,0.0213619779586792,"[0.021433727264404297, 0.020558847427368163, 0.02063580894470215, 0.020620288848876952, 0.02056262397766113, 0.020636512756347657, 0.020603328704833983, 0.020608224868774415, 0.020594112396240233, 0.020562240600585938]",tokens/s,12378.169571029865,kWh,2.4537378093403793e-07,1.344271180800455e-07,1.1843843038921175e-06,1.5641852029062008e-06,tokens/kWh,163663484.04547048,MB,1281.630208,3855.089664,0.0,3208.64256,3022.612992,s,10,11.67772607421875,1.167772607421875,0.009795152085250813,1.1701090087890624,1.1784019287109375,1.1804039672851563,1.1820055981445312,"[1.17795703125, 1.161013916015625, 1.14857666015625, 1.1743184814453125, 1.1568768310546875, 1.182406005859375, 1.16305908203125, 1.173300048828125, 1.1711513671875, 1.169066650390625]",tokens/s,53.94885922104895,kWh,1.3632492338212307e-05,7.46837398083023e-06,3.701328514310734e-05,5.811415146214988e-05,tokens/kWh,1084073.300821269,,s,629,11.843653614044197,0.018829338019148156,0.002516662092082329,0.018708480834960937,0.01890058174133301,0.019062169647216798,0.03916615753173828,"[0.019811328887939454, 0.018520063400268554, 0.018324480056762696, 0.0200130558013916, 0.019280895233154297, 0.01903308868408203, 0.01906380844116211, 0.018881536483764647, 0.01883955192565918, 0.018844671249389648, 0.018787328720092773, 0.01883955192565918, 0.018761728286743166, 0.018748416900634765, 0.01880575942993164, 0.01878118324279785, 0.01886412811279297, 0.018185216903686522, 0.018102272033691406, 0.01807155227661133, 0.018160640716552736, 0.018928640365600585, 0.018852863311767578, 0.01869824028015137, 0.018708480834960937, 0.018721792221069337, 0.018719743728637696, 0.01876479911804199, 0.01876479911804199, 0.018898944854736328, 0.01900339126586914, 0.0192675838470459, 0.019198976516723632, 0.018765823364257812, 0.018882560729980468, 0.018860031127929687, 0.01880166435241699, 0.018687999725341797, 0.01880166435241699, 0.018736127853393555, 0.018740224838256835, 0.01881497573852539, 0.018742271423339844, 0.018808832168579103, 0.018775039672851563, 0.018743295669555664, 0.018770944595336913, 0.018782207489013672, 0.018899967193603515, 0.018817024230957033, 0.01877299118041992, 0.0184586238861084, 0.018214912414550782, 0.01801625633239746, 0.017946624755859376, 0.017928192138671875, 0.017881088256835938, 0.017983488082885742, 0.018107391357421874, 0.018298879623413086, 0.018021440505981444, 0.01790457534790039, 0.039158782958984374, 0.01807974433898926, 0.018158592224121094, 0.018129919052124025, 0.018092031478881835, 0.018141183853149414, 0.018143232345581056, 0.018184192657470705, 0.018163711547851562, 0.01823232078552246, 0.018020351409912108, 0.01802342414855957, 0.018184192657470705, 0.01815449523925781, 0.018076671600341796, 0.018119680404663087, 0.018033664703369142, 0.018174976348876954, 0.01819647979736328, 0.018129919052124025, 0.018095104217529297, 0.018076671600341796, 0.018130943298339842, 0.01808076858520508, 0.01802137565612793, 0.0184586238861084, 0.01886617660522461, 0.019013631820678712, 0.01882111930847168, 0.018780223846435545, 0.018735040664672853, 0.018720767974853517, 0.018695167541503906, 0.018824192047119142, 0.018774015426635742, 0.01879347229003906, 0.018732032775878905, 0.01884364891052246, 0.018255872726440428, 0.017918975830078124, 0.018164735794067383, 0.01822105598449707, 0.018082815170288084, 0.018428928375244142, 0.018062335968017578, 0.018172927856445312, 0.01840640068054199, 0.01820057678222656, 0.01825279998779297, 0.018880512237548826, 0.01882624053955078, 0.01876479911804199, 0.018779136657714843, 0.018787328720092773, 0.01879452705383301, 0.018790367126464844, 0.01883647918701172, 0.018786304473876952, 0.01879449653625488, 0.018761728286743166, 0.018771968841552734, 0.018935808181762694, 0.01883443260192871, 0.03990630340576172, 0.01927987289428711, 0.019092512130737305, 0.01883235168457031, 0.0187740478515625, 0.018793439865112303, 0.01879654312133789, 0.01839411163330078, 0.018700288772583007, 0.018759679794311524, 0.01877299118041992, 0.018027519226074217, 0.017958911895751953, 0.01814630317687988, 0.018099199295043944, 0.018127872467041017, 0.018164735794067383, 0.018173952102661133, 0.018148351669311523, 0.01798860740661621, 0.018168832778930662, 0.01808793640136719, 0.0181790714263916, 0.018093055725097656, 0.018062335968017578, 0.01797225570678711, 0.01810736083984375, 0.018118656158447266, 0.01798860740661621, 0.018095104217529297, 0.01847091293334961, 0.018741247177124023, 0.018341888427734376, 0.017971200942993162, 0.018062335968017578, 0.01820262336730957, 0.018052095413208007, 0.017952768325805665, 0.018000896453857423, 0.01796505546569824, 0.018142208099365235, 0.017992704391479493, 0.018036735534667968, 0.017933311462402343, 0.01817804718017578, 0.018119680404663087, 0.018151424407958985, 0.018121728897094725, 0.017979391098022462, 0.017910783767700195, 0.018081792831420897, 0.018114559173583983, 0.018126848220825196, 0.018017280578613282, 0.01821286392211914, 0.017970176696777345, 0.018078720092773438, 0.018059263229370116, 0.018168832778930662, 0.018144256591796876, 0.018129919052124025, 0.018089984893798827, 0.017954816818237306, 0.03987148666381836, 0.01863270378112793, 0.018745344161987306, 0.018784255981445314, 0.018768896102905275, 0.018724863052368163, 0.019107839584350587, 0.01884671974182129, 0.018743295669555664, 0.018701311111450195, 0.018728960037231446, 0.018537471771240235, 0.01843507194519043, 0.018705408096313478, 0.018718719482421875, 0.01855897521972656, 0.018019327163696287, 0.0178606071472168, 0.018081792831420897, 0.018094079971313477, 0.018273279190063475, 0.018775039672851563, 0.01882009506225586, 0.018662399291992187, 0.0184453125, 0.018505727767944336, 0.018746368408203123, 0.01878937530517578, 0.01862553596496582, 0.018795520782470702, 0.018957311630249024, 0.018925567626953126, 0.01880166435241699, 0.018663423538208008, 0.018711551666259766, 0.018807807922363282, 0.018768896102905275, 0.01889587211608887, 0.01879449653625488, 0.018755615234375, 0.018808799743652342, 0.01878118324279785, 0.018938880920410156, 0.018721792221069337, 0.018774015426635742, 0.018774015426635742, 0.018880512237548826, 0.01887539291381836, 0.018873344421386717, 0.01880473518371582, 0.018447359085083007, 0.0186060791015625, 0.018737152099609376, 0.018851839065551757, 0.018755584716796874, 0.018754560470581053, 0.01862758445739746, 0.01817190361022949, 0.018434080123901367, 0.018396127700805665, 0.017946624755859376, 0.017912832260131836, 0.017884159088134767, 0.03920281600952148, 0.018101247787475586, 0.018019327163696287, 0.017945600509643556, 0.017887231826782226, 0.01817190361022949, 0.018108415603637695, 0.018116607666015624, 0.018225151062011717, 0.018096128463745118, 0.01821286392211914, 0.018107391357421874, 0.018114591598510744, 0.01810633659362793, 0.017942527770996093, 0.01802137565612793, 0.018114559173583983, 0.018119680404663087, 0.018108415603637695, 0.018075647354125975, 0.018111488342285157, 0.018139135360717772, 0.018120704650878908, 0.01807257652282715, 0.018116640090942382, 0.018140127182006836, 0.017898496627807618, 0.017938432693481447, 0.017966079711914062, 0.017976320266723633, 0.018124799728393554, 0.01807360076904297, 0.018099199295043944, 0.018191360473632814, 0.018123775482177733, 0.017909759521484374, 0.017924127578735353, 0.01807663917541504, 0.018093055725097656, 0.018101247787475586, 0.018267135620117187, 0.018225151062011717, 0.0184965763092041, 0.019235776901245116, 0.01905971145629883, 0.01877299118041992, 0.01876479911804199, 0.018733055114746093, 0.018760704040527345, 0.018732032775878905, 0.019904512405395508, 0.01905356788635254, 0.018762752532958983, 0.018679807662963867, 0.01881395149230957, 0.019001344680786132, 0.019094528198242186, 0.01900748825073242, 0.01903923225402832, 0.01881395149230957, 0.018752511978149415, 0.018744319915771485, 0.018768896102905275, 0.03986636734008789, 0.01883647918701172, 0.018803712844848632, 0.01875660705566406, 0.018731008529663085, 0.01879654312133789, 0.01903001594543457, 0.018766847610473633, 0.018717695236206054, 0.018767871856689454, 0.01880575942993164, 0.018885631561279297, 0.018757631301879883, 0.018861055374145508, 0.018729984283447267, 0.018742271423339844, 0.018777088165283205, 0.018785280227661134, 0.019211263656616212, 0.01919385528564453, 0.018997247695922852, 0.018874368667602538, 0.018778112411499022, 0.018743295669555664, 0.018885631561279297, 0.018787328720092773, 0.018803712844848632, 0.018865152359008788, 0.01885593605041504, 0.018766847610473633, 0.018769920349121092, 0.018761728286743166, 0.018917375564575196, 0.018763776779174804, 0.018811904907226562, 0.018697280883789063, 0.0187524471282959, 0.018740224838256835, 0.018800640106201173, 0.018702335357666015, 0.018808832168579103, 0.018675712585449217, 0.018720767974853517, 0.01876479911804199, 0.018732032775878905, 0.018747392654418944, 0.018777088165283205, 0.018791423797607423, 0.01879654312133789, 0.018514944076538087, 0.018373632431030275, 0.017957887649536132, 0.01824665641784668, 0.018174016952514648, 0.01859884834289551, 0.018907136917114258, 0.018757631301879883, 0.01877299118041992, 0.018791423797607423, 0.018975744247436522, 0.018745344161987306, 0.018761728286743166, 0.018783231735229493, 0.03928985595703125, 0.01825484848022461, 0.018139135360717772, 0.018082815170288084, 0.018225151062011717, 0.01842995262145996, 0.01820262336730957, 0.018132991790771484, 0.018206720352172853, 0.018141183853149414, 0.018197504043579102, 0.018094079971313477, 0.01823027229309082, 0.018167808532714845, 0.01801420783996582, 0.019503103256225587, 0.019194879531860352, 0.01882316780090332, 0.018817024230957033, 0.018758655548095703, 0.018777088165283205, 0.018787328720092773, 0.018930688858032226, 0.0188538875579834, 0.01879654312133789, 0.01824870491027832, 0.017950719833374023, 0.018108415603637695, 0.01822208023071289, 0.018579456329345705, 0.01821183967590332, 0.018135040283203126, 0.018172927856445312, 0.018176000595092775, 0.018152448654174806, 0.018163711547851562, 0.018181119918823242, 0.018190336227416993, 0.0182108154296875, 0.018268159866333008, 0.01809715270996094, 0.017961984634399415, 0.017941503524780272, 0.01805516815185547, 0.0188272647857666, 0.018779136657714843, 0.01880678367614746, 0.018647039413452148, 0.018488319396972656, 0.01861222457885742, 0.01887539291381836, 0.018758655548095703, 0.018832384109497072, 0.018455551147460936, 0.018284543991088868, 0.018724863052368163, 0.01877299118041992, 0.01884569549560547, 0.0188538875579834, 0.018675712585449217, 0.018808832168579103, 0.01879859161376953, 0.018766847610473633, 0.03983257675170898, 0.018760704040527345, 0.01882111930847168, 0.018751487731933594, 0.018694143295288086, 0.01848422431945801, 0.018942975997924806, 0.018903039932250978, 0.018890752792358398, 0.018724863052368163, 0.018746368408203123, 0.018905120849609373, 0.018867168426513672, 0.01878835105895996, 0.018759679794311524, 0.018284543991088868, 0.01823232078552246, 0.018134048461914062, 0.018978784561157227, 0.018748416900634765, 0.01876479911804199, 0.018776063919067384, 0.01886412811279297, 0.01885699272155762, 0.018868192672729493, 0.018745344161987306, 0.018738176345825194, 0.018680896759033203, 0.018373567581176756, 0.018557952880859374, 0.01880268859863281, 0.018437120437622072, 0.01866035270690918, 0.01879449653625488, 0.018786304473876952, 0.0187064323425293, 0.018728960037231446, 0.018701311111450195, 0.01864192008972168, 0.018856960296630858, 0.01887846374511719, 0.018702335357666015, 0.01877299118041992, 0.01881292724609375, 0.01884979248046875, 0.018740224838256835, 0.018774015426635742, 0.018683904647827147, 0.019269632339477538, 0.018865152359008788, 0.0186746883392334, 0.01840742492675781, 0.01841663932800293, 0.01788313674926758, 0.017934335708618163, 0.018140159606933593, 0.018233343124389647, 0.017922048568725587, 0.01798041534423828, 0.01803980827331543, 0.018112512588500978, 0.018155519485473632, 0.018116607666015624, 0.039169025421142575, 0.018086912155151368, 0.018190336227416993, 0.018098175048828127, 0.01821696090698242, 0.018068479537963866, 0.018184192657470705, 0.018124799728393554, 0.018381856918334962, 0.018158559799194337, 0.018077695846557617, 0.018109439849853515, 0.018177024841308592, 0.01824358367919922, 0.018456575393676757, 0.018713600158691408, 0.01879452705383301, 0.018790367126464844, 0.018744319915771485, 0.01886720085144043, 0.018817024230957033, 0.018703359603881836, 0.01879654312133789, 0.019148799896240236, 0.01902387237548828, 0.01881395149230957, 0.018774015426635742, 0.018743295669555664, 0.018722816467285155, 0.018937856674194335, 0.019924991607666014, 0.020033536911010744, 0.019301376342773437, 0.018844671249389648, 0.018931711196899414, 0.01887948799133301, 0.01899728012084961, 0.018385887145996094, 0.01819647979736328, 0.01820467185974121, 0.01883852767944336, 0.01877299118041992, 0.01886617660522461, 0.018763776779174804, 0.018807807922363282, 0.01878118324279785, 0.01880575942993164, 0.018885631561279297, 0.01887027168273926, 0.018961408615112304, 0.018891775131225585, 0.01880575942993164, 0.018233343124389647, 0.018134016036987305, 0.01814630317687988, 0.018123775482177733, 0.018207744598388673, 0.018159616470336915, 0.018291711807250977, 0.018136064529418947, 0.018159616470336915, 0.018168832778930662, 0.018132991790771484, 0.03897753524780274, 0.01790771293640137, 0.017945600509643556, 0.01817087936401367, 0.018183168411254884, 0.01819545555114746, 0.018176000595092775, 0.018155519485473632, 0.01813811111450195, 0.018184192657470705, 0.018118656158447266, 0.018229248046875, 0.01809715270996094, 0.018325504302978517, 0.018780160903930664, 0.018732032775878905, 0.018790399551391602, 0.01871571159362793, 0.018745279312133788, 0.018913280487060546, 0.018744319915771485, 0.018752511978149415, 0.018498559951782227, 0.018455615997314455, 0.018637760162353516, 0.018530303955078126, 0.01886412811279297, 0.01880473518371582, 0.01880985641479492, 0.01883443260192871, 0.018505727767944336, 0.018326528549194337, 0.01814630317687988, 0.01818623924255371, 0.018143232345581056, 0.018149375915527344, 0.017979391098022462, 0.01822208023071289, 0.018117631912231445, 0.01818726348876953, 0.01900441551208496, 0.018865152359008788, 0.018771968841552734, 0.01887539291381836, 0.018844671249389648, 0.018803712844848632, 0.01883135986328125, 0.018760704040527345, 0.01880268859863281, 0.018799615859985352, 0.018494464874267577, 0.018504703521728515, 0.018561023712158203, 0.01884671974182129, 0.0200263671875, 0.019103744506835937, 0.018423807144165038, 0.018963455200195312, 0.01885491180419922, 0.01869824028015137, 0.01881907272338867, 0.018778112411499022, 0.01883443260192871]",tokens/s,53.108611624214724,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,google/recurrentgemma-2b,google/recurrentgemma-2b,cuda,0,42,,,True,,,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",recurrent_gemma,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,stabilityai/stablelm-2-1_6b,stabilityai/stablelm-2-1_6b,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1263.665152,4276.617216,0.0,3632.267264,3408.132096,s,10,0.19949964714050294,0.019949964714050295,0.0002634110509256969,0.019849407196044923,0.020051040267944336,0.020388576316833495,0.020658605155944823,"[0.020726112365722655, 0.01983452796936035, 0.019976032257080077, 0.019860031127929688, 0.019847679138183593, 0.019810176849365233, 0.019833568572998048, 0.01993631935119629, 0.01985113525390625, 0.019824064254760743]",tokens/s,12832.102896889093,kWh,2.367375350109959e-07,1.297196071719301e-07,1.1859149990541087e-06,1.552372141237035e-06,tokens/kWh,164908911.465006,MB,1263.665152,4278.714368,0.0,3632.267264,3474.14784,s,10,11.340074707031249,1.134007470703125,0.017892579925606965,1.1269158325195312,1.1675471557617187,1.1677104675292969,1.1678411169433593,"[1.1675108642578125, 1.1208251953125, 1.12455859375, 1.130638671875, 1.1247318115234375, 1.1166605224609376, 1.167873779296875, 1.1390423583984375, 1.119133056640625, 1.129099853515625]",tokens/s,55.55518956232075,kWh,1.3274553809949536e-05,7.274062950731538e-06,3.679097649354576e-05,5.733959325422684e-05,tokens/kWh,1098717.246225947,,s,629,11.499658243179313,0.01828244553764598,0.0024469567160428553,0.017869823455810546,0.018637005233764648,0.01880657958984375,0.038081085357666015,"[0.01884774398803711, 0.01860812759399414, 0.0185743350982666, 0.0184453125, 0.01886617660522461, 0.01882931137084961, 0.01926348876953125, 0.01861427116394043, 0.01864192008972168, 0.018528255462646484, 0.018469888687133788, 0.018545663833618165, 0.01879859161376953, 0.018585599899291993, 0.018709503173828124, 0.018785280227661134, 0.01863580894470215, 0.018593759536743165, 0.018550783157348632, 0.018528255462646484, 0.01887846374511719, 0.018746368408203123, 0.018538496017456055, 0.018334720611572267, 0.01846067237854004, 0.018226175308227538, 0.01903411293029785, 0.022605823516845702, 0.01947648048400879, 0.01963827133178711, 0.018687999725341797, 0.018511871337890624, 0.018373632431030275, 0.01879449653625488, 0.018298879623413086, 0.018463743209838866, 0.018551807403564453, 0.018501632690429686, 0.018333696365356447, 0.018592767715454102, 0.018457599639892578, 0.018525184631347655, 0.018374656677246092, 0.018017280578613282, 0.017897472381591797, 0.01785651206970215, 0.018173952102661133, 0.018081792831420897, 0.017719295501708983, 0.017928192138671875, 0.01800294494628906, 0.018084863662719726, 0.018745344161987306, 0.018118656158447266, 0.01796505546569824, 0.017923072814941408, 0.01820262336730957, 0.0178606071472168, 0.017819648742675782, 0.017984512329101563, 0.017900543212890627, 0.01789030456542969, 0.0381952018737793, 0.017870847702026366, 0.01784524726867676, 0.017769472122192383, 0.018292736053466797, 0.018486335754394533, 0.017918912887573243, 0.01781760025024414, 0.017871904373168945, 0.017918943405151367, 0.017780736923217775, 0.017930240631103517, 0.01782374382019043, 0.01784524726867676, 0.017900543212890627, 0.01780940818786621, 0.017913856506347657, 0.01780735969543457, 0.01763020706176758, 0.017731584548950196, 0.017713151931762695, 0.01789132881164551, 0.017888256072998047, 0.017942527770996093, 0.01791391944885254, 0.01770182418823242, 0.017869823455810546, 0.017869823455810546, 0.017781759262084963, 0.017686527252197267, 0.01782476806640625, 0.017871871948242187, 0.017888256072998047, 0.017958911895751953, 0.01779199981689453, 0.01788729667663574, 0.017980352401733398, 0.017812480926513673, 0.017787904739379884, 0.017778688430786133, 0.017715200424194336, 0.017729536056518554, 0.017673248291015624, 0.017790016174316407, 0.017834911346435545, 0.017758304595947266, 0.01781443214416504, 0.017750015258789064, 0.01765888023376465, 0.017469440460205078, 0.017543167114257813, 0.017501184463500977, 0.017442815780639647, 0.017552383422851564, 0.017527807235717775, 0.017427455902099608, 0.017501184463500977, 0.01761075210571289, 0.01739571189880371, 0.01741619110107422, 0.017682464599609374, 0.017755104064941407, 0.018130943298339842, 0.038694911956787106, 0.017531904220581054, 0.01760166358947754, 0.018078592300415038, 0.018190336227416993, 0.0181146240234375, 0.018285503387451173, 0.017655807495117186, 0.017790048599243165, 0.017686431884765624, 0.017709056854248048, 0.017706144332885743, 0.017629024505615234, 0.01817308807373047, 0.01808880043029785, 0.01805721664428711, 0.017702911376953127, 0.01760870361328125, 0.017690624237060547, 0.018164735794067383, 0.018199615478515625, 0.01773049545288086, 0.017687551498413084, 0.017640512466430665, 0.017819583892822264, 0.017896448135375977, 0.017967103958129883, 0.017886207580566405, 0.01800601577758789, 0.01783705520629883, 0.01794867134094238, 0.017704959869384765, 0.0178657283782959, 0.018155519485473632, 0.01804697608947754, 0.017895423889160156, 0.017922048568725587, 0.01785958480834961, 0.017952768325805665, 0.01794047927856445, 0.017904640197753906, 0.017993728637695314, 0.017728511810302734, 0.017718271255493166, 0.017839103698730468, 0.017970176696777345, 0.017928192138671875, 0.017917951583862304, 0.01785753631591797, 0.017836032867431642, 0.01789132881164551, 0.017528831481933595, 0.017690624237060547, 0.017417375564575194, 0.017654624938964844, 0.017476608276367187, 0.017512447357177736, 0.017732608795166017, 0.017730560302734375, 0.017711103439331053, 0.017747968673706056, 0.01785036849975586, 0.017827840805053712, 0.038109184265136715, 0.017896448135375977, 0.017785856246948242, 0.017747968673706056, 0.017863679885864257, 0.01794047927856445, 0.01781657600402832, 0.0178288631439209, 0.018043903350830077, 0.01860403251647949, 0.01850992012023926, 0.018368480682373046, 0.017993663787841795, 0.017885183334350584, 0.01863987159729004, 0.01838083267211914, 0.018540512084960936, 0.018571264266967775, 0.01838595199584961, 0.01786467170715332, 0.017897472381591797, 0.017870847702026366, 0.017821855545043945, 0.01789423942565918, 0.01785753631591797, 0.01785241508483887, 0.017904640197753906, 0.017864704132080078, 0.017886207580566405, 0.017912832260131836, 0.017862655639648437, 0.01785241508483887, 0.01790771293640137, 0.0176629753112793, 0.017690624237060547, 0.01764352035522461, 0.017605632781982423, 0.017690624237060547, 0.017778688430786133, 0.017887231826782226, 0.01764352035522461, 0.017689599990844726, 0.017674240112304687, 0.017656831741333007, 0.017714176177978515, 0.01765990447998047, 0.01776639938354492, 0.017946624755859376, 0.018001920700073244, 0.01787392044067383, 0.017918975830078124, 0.0178606071472168, 0.017961984634399415, 0.017880064010620117, 0.017879039764404296, 0.01788313674926758, 0.017908735275268553, 0.017911808013916015, 0.017887231826782226, 0.017933311462402343, 0.017961984634399415, 0.01840947151184082, 0.018391040802001952, 0.03800883102416992, 0.017836032867431642, 0.017935359954833984, 0.0178657283782959, 0.0178606071472168, 0.01785958480834961, 0.017929216384887696, 0.01760972785949707, 0.018041856765747072, 0.017744895935058593, 0.017680383682250975, 0.017710079193115236, 0.017723392486572266, 0.017844224929809572, 0.017727487564086913, 0.017938432693481447, 0.017853439331054686, 0.017928192138671875, 0.017920000076293945, 0.017926143646240233, 0.01784115219116211, 0.01781350326538086, 0.017854463577270507, 0.0178288631439209, 0.01781350326538086, 0.017777664184570312, 0.017692703247070313, 0.01784419250488281, 0.017942623138427736, 0.017874847412109374, 0.017878015518188475, 0.01783705520629883, 0.017672191619873046, 0.01756979179382324, 0.017896448135375977, 0.01780940818786621, 0.017870847702026366, 0.017880064010620117, 0.017829887390136717, 0.017933311462402343, 0.017691648483276368, 0.017722400665283203, 0.01789743995666504, 0.017904640197753906, 0.017884159088134767, 0.01786675262451172, 0.017715200424194336, 0.01763942337036133, 0.017699840545654297, 0.017613824844360353, 0.0176363525390625, 0.01824051284790039, 0.019198976516723632, 0.018696191787719727, 0.018225151062011717, 0.01825798416137695, 0.017827903747558594, 0.017515392303466798, 0.01743155288696289, 0.01765376091003418, 0.017583168029785157, 0.01768953514099121, 0.017624063491821287, 0.037850112915039064, 0.017649663925170898, 0.017490943908691405, 0.01738444709777832, 0.017460224151611328, 0.017447935104370118, 0.017443840026855468, 0.01741414451599121, 0.01745305633544922, 0.01765990447998047, 0.017498111724853514, 0.017584127426147463, 0.01738764762878418, 0.017346431732177734, 0.01743052864074707, 0.017526784896850587, 0.01743769645690918, 0.017398847579956054, 0.017348575592041015, 0.017396703720092773, 0.017445087432861328, 0.017409824371337892, 0.01741414451599121, 0.01765171241760254, 0.017551359176635743, 0.01760358428955078, 0.017538047790527343, 0.017663999557495116, 0.01763532829284668, 0.01761587142944336, 0.017689599990844726, 0.01764352035522461, 0.017552383422851564, 0.017697792053222656, 0.017903615951538086, 0.018372671127319336, 0.018416576385498047, 0.018362367630004883, 0.018371583938598633, 0.018062335968017578, 0.01831532859802246, 0.01775814437866211, 0.01764249610900879, 0.017722368240356445, 0.017723392486572266, 0.017605632781982423, 0.01780121612548828, 0.017976383209228515, 0.017961919784545897, 0.017884159088134767, 0.01782476806640625, 0.017724416732788087, 0.017570816040039062, 0.017950719833374023, 0.018543615341186523, 0.01839308738708496, 0.018249727249145507, 0.01800294494628906, 0.01761177635192871, 0.017582080841064454, 0.01784012794494629, 0.01781760025024414, 0.01779622459411621, 0.0381162223815918, 0.017648767471313477, 0.017774463653564453, 0.01767945671081543, 0.017769376754760743, 0.01789132881164551, 0.01785139274597168, 0.017900543212890627, 0.017864704132080078, 0.017830911636352538, 0.017892351150512697, 0.017859615325927735, 0.01850569534301758, 0.019381248474121093, 0.01983795166015625, 0.018976768493652343, 0.01903001594543457, 0.019076095581054688, 0.01887948799133301, 0.0186296329498291, 0.018323455810546875, 0.01839206314086914, 0.018590879440307618, 0.018406240463256836, 0.019336191177368164, 0.02059878349304199, 0.018817184448242187, 0.018888544082641602, 0.018720767974853517, 0.0186296329498291, 0.01865216064453125, 0.01863372802734375, 0.01859280014038086, 0.01830089569091797, 0.01840127944946289, 0.018692096710205077, 0.018547712326049806, 0.018573312759399413, 0.018811904907226562, 0.018626560211181642, 0.0186296329498291, 0.018333696365356447, 0.018538496017456055, 0.018677759170532226, 0.018694143295288086, 0.018552831649780274, 0.018734079360961914, 0.01859174346923828, 0.01863065528869629, 0.018649120330810547, 0.01866851234436035, 0.01824665641784668, 0.018742271423339844, 0.018626560211181642, 0.018760704040527345, 0.01863680076599121, 0.018888704299926756, 0.018742271423339844, 0.01761996841430664, 0.017937408447265626, 0.018754560470581053, 0.018646015167236327, 0.018474048614501953, 0.038491073608398436, 0.01862451171875, 0.018637823104858398, 0.018675712585449217, 0.0182476806640625, 0.018325504302978517, 0.018504703521728515, 0.018465791702270508, 0.018594816207885743, 0.018499584197998048, 0.0184770565032959, 0.01846272087097168, 0.018290687561035156, 0.018323455810546875, 0.01768448066711426, 0.017693695068359376, 0.017713151931762695, 0.01822412872314453, 0.017855487823486327, 0.0178288631439209, 0.01780735969543457, 0.017899551391601563, 0.01788822364807129, 0.01756985664367676, 0.017651647567749024, 0.017674240112304687, 0.017722368240356445, 0.017649663925170898, 0.01773465538024902, 0.017724416732788087, 0.01764352035522461, 0.017772544860839845, 0.01773465538024902, 0.017633279800415038, 0.017681407928466796, 0.017698816299438477, 0.017729536056518554, 0.01789132881164551, 0.01787392044067383, 0.01785753631591797, 0.017871871948242187, 0.017898496627807618, 0.01784217643737793, 0.0178606071472168, 0.01786777687072754, 0.017723392486572266, 0.01780531120300293, 0.018555904388427736, 0.01862144088745117, 0.018695167541503906, 0.018594816207885743, 0.018564096450805666, 0.018564096450805666, 0.018296831130981444, 0.017720319747924804, 0.017872896194458008, 0.01793846321105957, 0.018650079727172853, 0.01860710334777832, 0.01869004821777344, 0.018527231216430663, 0.017903615951538086, 0.017898496627807618, 0.038161407470703124, 0.017778688430786133, 0.017705984115600586, 0.017822719573974608, 0.017829887390136717, 0.017661951065063478, 0.017723392486572266, 0.01765068817138672, 0.01819340705871582, 0.01789030456542969, 0.01787596893310547, 0.017937408447265626, 0.017811456680297853, 0.017887231826782226, 0.0178155517578125, 0.01800704002380371, 0.01803059196472168, 0.01794047927856445, 0.01775923156738281, 0.01770086479187012, 0.017928192138671875, 0.01822208023071289, 0.01841663932800293, 0.018326528549194337, 0.018387968063354493, 0.01815449523925781, 0.01760870361328125, 0.017763328552246094, 0.017912832260131836, 0.017625087738037108, 0.017707008361816406, 0.017468416213989257, 0.01760665512084961, 0.01775923156738281, 0.017729536056518554, 0.017589248657226563, 0.01760051155090332, 0.017683456420898438, 0.017667072296142578, 0.017540096282958984, 0.017803264617919923, 0.017765375137329103, 0.017762304306030274, 0.017482751846313475, 0.01758720016479492, 0.01759129524230957, 0.017686527252197267, 0.017584127426147463, 0.01745305633544922, 0.01744076728820801, 0.017519775390625, 0.01746211242675781, 0.01759040069580078, 0.017678207397460937, 0.01759436798095703, 0.01762713623046875, 0.017665023803710937, 0.017604608535766602, 0.017708032608032227, 0.01757900810241699, 0.0176629753112793, 0.01758515167236328, 0.017698816299438477, 0.03851980972290039, 0.017605632781982423, 0.017757247924804688, 0.017596351623535157, 0.017496063232421876, 0.017955839157104494, 0.018095104217529297, 0.017902591705322265, 0.017825920104980467, 0.01748467254638672, 0.0174716796875, 0.017650495529174803, 0.017705087661743165, 0.017675167083740236, 0.017752031326293945, 0.01775721549987793, 0.0176977596282959, 0.017581056594848633, 0.01759231948852539, 0.01760883140563965, 0.01782156753540039, 0.01788313674926758, 0.017892351150512697, 0.017768447875976562, 0.01786675262451172, 0.017938528060913086, 0.01778268814086914, 0.017868799209594728, 0.017764352798461915, 0.017687551498413084, 0.017722368240356445, 0.017803264617919923, 0.018184192657470705, 0.018524160385131837, 0.018824192047119142, 0.01858252716064453, 0.01865318489074707, 0.018654207229614257, 0.018593791961669923, 0.018655263900756835, 0.01783600044250488, 0.01798246383666992, 0.018033664703369142, 0.018075647354125975, 0.01785241508483887, 0.017914880752563478, 0.017885183334350584, 0.017909759521484374, 0.017753087997436523, 0.017929216384887696, 0.017915903091430666, 0.017873952865600586, 0.017821664810180663, 0.017863679885864257, 0.01786675262451172, 0.017855487823486327, 0.01789132881164551, 0.0182609920501709, 0.01783705520629883, 0.017803264617919923, 0.017864704132080078, 0.017904640197753906, 0.017830911636352538]",tokens/s,54.697277666757834,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,togethercomputer/RedPajama-INCITE-Base-3B-v1,togethercomputer/RedPajama-INCITE-Base-3B-v1,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.42.3,,0.31.0,,,,1.20.0,,,,0.11.1,,,,MB,1264.766976,6791.102464,0.0,6144.65536,6020.360704,s,10,0.3191828784942627,0.03191828784942626,0.00016736062221300383,0.03186444759368896,0.03199633312225342,0.03220202207565308,0.0323665732383728,"[0.03240771102905273, 0.03186579132080078, 0.03183760070800781, 0.03180099105834961, 0.03186310386657715, 0.031950624465942386, 0.031843807220458986, 0.03189072036743164, 0.03184345626831055, 0.03187907218933105]",tokens/s,8020.48033427337,kWh,3.7628885118655933e-07,2.0618902313305993e-07,1.971235653420372e-06,2.553713527739991e-06,tokens/kWh,100246169.8303949,MB,1264.766976,6791.102464,0.0,6144.65536,6124.220416,s,10,13.277724731445312,1.3277724731445313,0.008849238502958255,1.3250625,1.3352716918945313,1.3424027526855469,1.3481076013183593,"[1.3495338134765624, 1.3274376220703126, 1.3205426025390625, 1.3187171630859376, 1.3319588623046874, 1.3209241943359376, 1.3226873779296875, 1.3308304443359376, 1.33368701171875, 1.3214056396484375]",tokens/s,47.447888304837825,kWh,1.5747257891544112e-05,8.62936996738627e-06,5.626900538317971e-05,8.06456332421101e-05,tokens/kWh,781195.4282864231,,s,629,13.541893121719358,0.02152924184693062,0.003825742965098097,0.020925439834594727,0.021629338455200195,0.021909708404541015,0.05297971313476563,"[0.022074367523193358, 0.0218470401763916, 0.0218920955657959, 0.021716991424560548, 0.02205183982849121, 0.022158336639404298, 0.022076416015625, 0.021808128356933593, 0.02207027244567871, 0.021828607559204103, 0.022004735946655272, 0.021743616104125976, 0.021956607818603514, 0.021753856658935547, 0.0218787841796875, 0.02189926338195801, 0.02230784034729004, 0.021600255966186522, 0.021813247680664064, 0.02168012809753418, 0.022402048110961914, 0.02267955207824707, 0.022037504196166992, 0.02172211265563965, 0.021841920852661133, 0.02147942352294922, 0.02131865692138672, 0.020923391342163086, 0.02091007995605469, 0.020788223266601562, 0.020900863647460938, 0.021388288497924804, 0.021513216018676756, 0.02084659194946289, 0.020929536819458007, 0.02081177520751953, 0.020934656143188478, 0.02081279945373535, 0.021165056228637694, 0.02089369583129883, 0.021172224044799806, 0.021635072708129883, 0.0210513916015625, 0.020883455276489257, 0.02085171127319336, 0.020746240615844725, 0.020876287460327148, 0.020799488067626954, 0.02123776054382324, 0.021089279174804687, 0.02089779281616211, 0.020960256576538085, 0.02087424087524414, 0.020876287460327148, 0.020926464080810548, 0.020896768569946288, 0.02086092758178711, 0.02086809539794922, 0.021539840698242187, 0.02106572723388672, 0.021200895309448242, 0.02084249687194824, 0.05304422378540039, 0.020823040008544923, 0.02083328056335449, 0.02083737564086914, 0.021402624130249022, 0.021428224563598632, 0.02092748832702637, 0.021517311096191406, 0.020867071151733398, 0.020783103942871094, 0.0208721923828125, 0.021001216888427734, 0.021164031982421876, 0.02083430480957031, 0.021163007736206055, 0.02149273681640625, 0.02145382308959961, 0.020884479522705078, 0.020855808258056642, 0.022568960189819336, 0.022962175369262695, 0.021841920852661133, 0.021736448287963867, 0.021531648635864258, 0.021008384704589843, 0.020965375900268556, 0.020925439834594727, 0.020918272018432618, 0.02085068893432617, 0.02084556770324707, 0.02086092758178711, 0.02081996726989746, 0.02099404716491699, 0.021073919296264648, 0.02089779281616211, 0.02090291213989258, 0.020888576507568358, 0.02092748832702637, 0.020980735778808594, 0.020936704635620116, 0.020817920684814452, 0.02087731170654297, 0.0208353271484375, 0.02085990333557129, 0.02085683250427246, 0.02082918357849121, 0.020785152435302736, 0.020858879089355468, 0.02086092758178711, 0.020805631637573242, 0.02085683250427246, 0.020930559158325195, 0.020805631637573242, 0.02081279945373535, 0.02082918357849121, 0.02082713508605957, 0.02084556770324707, 0.020939775466918945, 0.021342208862304687, 0.021606399536132814, 0.021497856140136717, 0.02084864044189453, 0.021114879608154297, 0.05298688125610351, 0.020787200927734374, 0.020989952087402345, 0.02085990333557129, 0.020785152435302736, 0.021110784530639647, 0.020989952087402345, 0.02088140869140625, 0.02082508850097656, 0.020944896697998046, 0.02087936019897461, 0.020900863647460938, 0.020808704376220705, 0.02083020782470703, 0.021024768829345702, 0.02082713508605957, 0.020926464080810548, 0.020958208084106447, 0.021016576766967773, 0.020925439834594727, 0.02091110420227051, 0.020853759765625, 0.020840448379516603, 0.020924415588378906, 0.02085171127319336, 0.02086297607421875, 0.020982784271240236, 0.021317632675170898, 0.0208353271484375, 0.02081177520751953, 0.020876287460327148, 0.02083635139465332, 0.020931583404541015, 0.02086297607421875, 0.02086809539794922, 0.020888576507568358, 0.020840448379516603, 0.02084249687194824, 0.020936704635620116, 0.020899839401245117, 0.020799488067626954, 0.020918272018432618, 0.02082611274719238, 0.020898815155029296, 0.020993024826049804, 0.02165760040283203, 0.021401599884033205, 0.02086604881286621, 0.020932607650756836, 0.02101759910583496, 0.020939775466918945, 0.020904960632324218, 0.020924415588378906, 0.021179391860961915, 0.020975616455078124, 0.021000192642211913, 0.021415935516357423, 0.02147225570678711, 0.021388288497924804, 0.02089369583129883, 0.020912128448486327, 0.02102272033691406, 0.020912128448486327, 0.053177345275878904, 0.021015552520751952, 0.02106777572631836, 0.020914176940917968, 0.02080460739135742, 0.02088243293762207, 0.021112831115722656, 0.02128998374938965, 0.020912128448486327, 0.020973567962646485, 0.020867071151733398, 0.020876287460327148, 0.02084966468811035, 0.020938751220703124, 0.021193727493286133, 0.02084659194946289, 0.020808704376220705, 0.020898815155029296, 0.02104934310913086, 0.02085990333557129, 0.020920320510864256, 0.020957183837890626, 0.020774911880493165, 0.02105446434020996, 0.020798463821411133, 0.021013504028320314, 0.02104217529296875, 0.021006336212158205, 0.02084147262573242, 0.02119987106323242, 0.02107084846496582, 0.020912128448486327, 0.02083430480957031, 0.020963327407836914, 0.02085478401184082, 0.02088652801513672, 0.021390335083007812, 0.021097471237182617, 0.02088243293762207, 0.020951040267944337, 0.02082713508605957, 0.02086092758178711, 0.02079641532897949, 0.02088243293762207, 0.02087116813659668, 0.02088140869140625, 0.02079641532897949, 0.02082611274719238, 0.020808704376220705, 0.02086809539794922, 0.02086092758178711, 0.020891647338867187, 0.02080460739135742, 0.020923391342163086, 0.02083328056335449, 0.02088140869140625, 0.020899839401245117, 0.0209039363861084, 0.020781055450439453, 0.02083737564086914, 0.0209039363861084, 0.02084249687194824, 0.020982784271240236, 0.052961280822753906, 0.020779008865356444, 0.02104319953918457, 0.020884479522705078, 0.020896768569946288, 0.02086297607421875, 0.020942848205566408, 0.020917247772216797, 0.02089472007751465, 0.020888576507568358, 0.022194175720214843, 0.021534719467163087, 0.021571584701538086, 0.02087424087524414, 0.020913152694702147, 0.02081996726989746, 0.020925439834594727, 0.020777984619140624, 0.020883455276489257, 0.021139455795288087, 0.021307392120361326, 0.020990976333618162, 0.020914176940917968, 0.020880384445190428, 0.0208721923828125, 0.020904960632324218, 0.02090291213989258, 0.020936704635620116, 0.021406719207763672, 0.022392831802368163, 0.021988351821899413, 0.021724159240722657, 0.021749759674072267, 0.02110873603820801, 0.02087936019897461, 0.021013504028320314, 0.02083430480957031, 0.020890623092651366, 0.02089574432373047, 0.020899839401245117, 0.02085990333557129, 0.021005311965942384, 0.020952064514160155, 0.02086297607421875, 0.0208855037689209, 0.02082099151611328, 0.020772863388061523, 0.020805631637573242, 0.020809728622436522, 0.020932607650756836, 0.021006336212158205, 0.02082918357849121, 0.020818944931030273, 0.020993024826049804, 0.020981760025024415, 0.021703680038452147, 0.02210201644897461, 0.02166886329650879, 0.021527551651000978, 0.021558271408081055, 0.021775360107421874, 0.02167398452758789, 0.02210201644897461, 0.053921791076660154, 0.02130227279663086, 0.020979711532592774, 0.02082918357849121, 0.020891647338867187, 0.021428224563598632, 0.02168627166748047, 0.02132275199890137, 0.020928512573242186, 0.02087833595275879, 0.02083020782470703, 0.020899839401245117, 0.020773887634277344, 0.021111808776855468, 0.021365760803222656, 0.02108006477355957, 0.021352447509765626, 0.021038080215454103, 0.020831232070922853, 0.020814847946166993, 0.02088755226135254, 0.020858879089355468, 0.021391359329223633, 0.020864000320434572, 0.02084351921081543, 0.02082815933227539, 0.020955135345458984, 0.02084966468811035, 0.02081996726989746, 0.02087321662902832, 0.02091110420227051, 0.020864000320434572, 0.020975616455078124, 0.021006336212158205, 0.02085478401184082, 0.020785152435302736, 0.021148767471313477, 0.020857759475708008, 0.020928512573242186, 0.020914176940917968, 0.020899839401245117, 0.020921344757080077, 0.0210565128326416, 0.0208855037689209, 0.020950016021728517, 0.020876287460327148, 0.020831232070922853, 0.020875263214111327, 0.020790271759033203, 0.02087116813659668, 0.020934656143188478, 0.02089472007751465, 0.02085171127319336, 0.02090188789367676, 0.020920320510864256, 0.020931583404541015, 0.020940799713134766, 0.020876287460327148, 0.020806655883789063, 0.020876287460327148, 0.0208035831451416, 0.020867071151733398, 0.02086092758178711, 0.05291929626464844, 0.02103910446166992, 0.02109951972961426, 0.020853759765625, 0.02083635139465332, 0.020925439834594727, 0.020942848205566408, 0.02087731170654297, 0.021301248550415038, 0.02275328063964844, 0.02209587287902832, 0.021178367614746094, 0.02108723258972168, 0.021184511184692383, 0.020948991775512696, 0.02085171127319336, 0.020734975814819336, 0.02086809539794922, 0.02079641532897949, 0.02086092758178711, 0.020922367095947265, 0.020875263214111327, 0.020771839141845702, 0.020814847946166993, 0.020763647079467772, 0.02084249687194824, 0.0214517765045166, 0.02123980712890625, 0.020805631637573242, 0.02080460739135742, 0.020880384445190428, 0.020883455276489257, 0.020928512573242186, 0.020980735778808594, 0.020888576507568358, 0.020989952087402345, 0.02086809539794922, 0.020966400146484376, 0.020972543716430665, 0.020867071151733398, 0.020884479522705078, 0.020900863647460938, 0.020924415588378906, 0.02109951972961426, 0.020973567962646485, 0.021110784530639647, 0.02107084846496582, 0.020913152694702147, 0.020793344497680662, 0.02090188789367676, 0.02101862335205078, 0.020959232330322267, 0.020958208084106447, 0.02088140869140625, 0.020800512313842775, 0.02088652801513672, 0.020913152694702147, 0.02124492835998535, 0.0208721923828125, 0.020891647338867187, 0.020929536819458007, 0.02102579116821289, 0.020943872451782225, 0.0530145263671875, 0.020928512573242186, 0.02085683250427246, 0.02086604881286621, 0.020777984619140624, 0.020974592208862306, 0.02144972801208496, 0.021369855880737306, 0.021155839920043946, 0.021142528533935546, 0.021315584182739256, 0.020962303161621093, 0.021009408950805664, 0.021148672103881837, 0.021006336212158205, 0.02104422378540039, 0.020899839401245117, 0.021227519989013673, 0.020953088760375976, 0.021402624130249022, 0.020922367095947265, 0.02130534362792969, 0.021916671752929686, 0.02188390350341797, 0.02149273681640625, 0.02085068893432617, 0.021357568740844726, 0.02085478401184082, 0.0212541446685791, 0.02087424087524414, 0.020934656143188478, 0.021146623611450196, 0.02090188789367676, 0.02143846321105957, 0.020934656143188478, 0.02108723258972168, 0.020799488067626954, 0.021352447509765626, 0.020920320510864256, 0.02088140869140625, 0.021404672622680664, 0.020931583404541015, 0.021153791427612305, 0.021279743194580078, 0.021565439224243164, 0.02088960075378418, 0.02107596778869629, 0.020976640701293944, 0.020853759765625, 0.021405696868896484, 0.021105663299560547, 0.02123161506652832, 0.02145280075073242, 0.02126540756225586, 0.021014528274536134, 0.020967424392700194, 0.021627904891967774, 0.020929536819458007, 0.02099404716491699, 0.020937728881835937, 0.021081087112426757, 0.02091007995605469, 0.02107904052734375, 0.05324492645263672, 0.0214835205078125, 0.021004287719726563, 0.021217279434204102, 0.020823040008544923, 0.021420032501220702, 0.020930559158325195, 0.021184511184692383, 0.020961280822753905, 0.02143436813354492, 0.02103091239929199, 0.02109542465209961, 0.020964351654052735, 0.02083328056335449, 0.0212807674407959, 0.021171199798583985, 0.02150399971008301, 0.020907007217407226, 0.02103193664550781, 0.020987903594970703, 0.021121023178100586, 0.02109951972961426, 0.020832256317138673, 0.02092748832702637, 0.020864000320434572, 0.021012479782104493, 0.02110771179199219, 0.020898815155029296, 0.021626880645751953, 0.021501951217651367, 0.021342208862304687, 0.021361663818359376, 0.021554176330566405, 0.020998144149780275, 0.020984832763671874, 0.02165043258666992, 0.021566463470458985, 0.021020671844482423, 0.021194751739501954, 0.02171801567077637, 0.021308416366577147, 0.02171801567077637, 0.0214968318939209, 0.02126438331604004, 0.021778432846069336, 0.02286079978942871, 0.021828607559204103, 0.020990976333618162, 0.020896768569946288, 0.02087731170654297, 0.020968448638916014, 0.02085273551940918, 0.02086092758178711, 0.0208855037689209, 0.020810752868652343, 0.0208855037689209, 0.02090291213989258, 0.02084454345703125, 0.020967424392700194, 0.0208721923828125, 0.02089369583129883, 0.021005311965942384, 0.020913152694702147, 0.053082111358642575, 0.020774911880493165, 0.021001216888427734, 0.020770816802978515, 0.020822015762329102, 0.021238784790039062, 0.02247270393371582, 0.021526527404785157, 0.02104115104675293, 0.020951040267944337, 0.0208855037689209, 0.020891647338867187, 0.02091007995605469, 0.020892671585083008, 0.02085273551940918, 0.020883455276489257, 0.020823040008544923, 0.02085068893432617, 0.02089779281616211, 0.02082611274719238, 0.020923391342163086, 0.020933631896972657, 0.0211015682220459, 0.02088652801513672, 0.020959232330322267, 0.02188902473449707, 0.02131455993652344, 0.020921344757080077, 0.02085683250427246, 0.020912128448486327, 0.02088960075378418, 0.02087116813659668, 0.021037055969238282, 0.02089472007751465, 0.02085683250427246, 0.020785152435302736, 0.02088140869140625, 0.020970495223999023, 0.020973567962646485, 0.02087424087524414, 0.020884479522705078, 0.02090188789367676, 0.020840448379516603, 0.0209039363861084, 0.020899839401245117, 0.02088140869140625, 0.02084454345703125, 0.02126643180847168, 0.02090598487854004, 0.020916223526000977, 0.02089574432373047, 0.020876287460327148, 0.0208855037689209, 0.02086195182800293, 0.020950016021728517, 0.02088652801513672, 0.02087833595275879, 0.021171199798583985, 0.020847616195678712, 0.021045248031616212, 0.020810752868652343, 0.020891647338867187, 0.020938751220703124]",tokens/s,46.44845401941397,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,meta-llama/Meta-Llama-3-8B,meta-llama/Meta-Llama-3-8B,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.1,,,,MB,1263.304704,17201.364992,0.0,16554.917888,16436.726784,s,10,0.8410996170043946,0.08410996170043945,0.00028077510301055336,0.08406849670410158,0.08432462005615234,0.08459322204589843,0.08480810363769531,"[0.08486182403564453, 0.08403580474853516, 0.08412857818603516, 0.08405661010742188, 0.08426493072509765, 0.08408377838134766, 0.08391302490234374, 0.08408038330078126, 0.08386335754394532, 0.08381132507324218]",tokens/s,3043.634723218076,kWh,9.952621090979804e-07,5.453488007104078e-07,6.541121012537723e-06,8.081731922346112e-06,tokens/kWh,31676378.585653912,MB,1263.632384,17201.364992,0.0,16554.917888,16436.729344,s,10,24.693271240234374,2.4693271240234376,0.0028284303330679,2.4685008544921874,2.4710369140625,2.4741880126953126,2.4767088916015627,"[2.477339111328125, 2.4673857421875, 2.468317626953125, 2.46868408203125, 2.467552734375, 2.467927978515625, 2.470336669921875, 2.469690673828125, 2.4673388671875, 2.46869775390625]",tokens/s,25.513023117548695,kWh,2.9084454136235377e-05,1.5939373986752988e-05,0.0001271281248116623,0.00017215195293465068,tokens/kWh,365955.7671350668,,s,629,25.409200088500953,0.04039618456041491,0.010046905879209297,0.039163902282714845,0.03925729293823242,0.039377509307861325,0.12367265563964844,"[0.03994015884399414, 0.03992364883422852, 0.03995647811889649, 0.03991142272949219, 0.03999846267700195, 0.039769088745117184, 0.03983564758300781, 0.03988991928100586, 0.039689216613769535, 0.03948031997680664, 0.03979776000976563, 0.03974655914306641, 0.03930112075805664, 0.03923865509033203, 0.039139328002929685, 0.03913113784790039, 0.0392355842590332, 0.03912704086303711, 0.03916185760498047, 0.039142398834228515, 0.03916799926757813, 0.039160831451416016, 0.03912908935546875, 0.0391383056640625, 0.03913011169433594, 0.039119873046875, 0.03914956665039063, 0.039136257171630856, 0.039209983825683595, 0.039142398834228515, 0.039169025421142575, 0.0391536636352539, 0.039160831451416016, 0.039155712127685545, 0.03923865509033203, 0.039175167083740234, 0.039136257171630856, 0.039160831451416016, 0.0391833610534668, 0.03911679840087891, 0.03924582290649414, 0.0391464958190918, 0.039180286407470705, 0.03914035034179687, 0.03916287994384766, 0.03916697692871094, 0.03926015853881836, 0.03916697692871094, 0.03917311859130859, 0.03915676879882812, 0.03922224044799805, 0.039174144744873046, 0.03922227096557617, 0.039206912994384766, 0.03926323318481445, 0.03915059280395508, 0.039191551208496093, 0.03918438339233398, 0.03925299072265625, 0.03918438339233398, 0.03918131256103516, 0.039163902282714845, 0.12374527740478515, 0.0390984001159668, 0.03906044769287109, 0.03909939193725586, 0.03905331039428711, 0.03916697692871094, 0.039076862335205076, 0.03911167907714844, 0.03909427261352539, 0.03914342498779297, 0.03913113784790039, 0.03920383834838867, 0.03914035034179687, 0.039122943878173826, 0.03913420867919922, 0.03911171340942383, 0.039118816375732425, 0.03914342498779297, 0.03909120178222656, 0.03916185760498047, 0.03908812713623047, 0.0391383056640625, 0.03913523101806641, 0.03920793533325195, 0.03914342498779297, 0.03911065673828125, 0.03913420867919922, 0.03911372756958008, 0.03915059280395508, 0.03915980911254883, 0.03915161514282227, 0.03924889755249023, 0.039144447326660156, 0.039175167083740234, 0.03912704086303711, 0.03919769668579102, 0.03913523101806641, 0.03915779113769531, 0.03916182327270508, 0.03913420867919922, 0.039215103149414066, 0.03922332763671875, 0.039185375213623044, 0.039193599700927735, 0.03915059280395508, 0.03917107009887695, 0.0391536636352539, 0.03917107009887695, 0.03917311859130859, 0.03919974517822265, 0.03919257736206055, 0.03918643188476562, 0.03918745422363281, 0.039185409545898435, 0.0391536636352539, 0.03924787139892578, 0.039177215576171875, 0.03918854522705078, 0.03916281509399414, 0.03922739028930664, 0.03921612930297851, 0.03918131256103516, 0.039258113861083986, 0.12390809631347656, 0.03910451126098633, 0.03909529495239258, 0.039109630584716795, 0.039093246459960936, 0.0391536636352539, 0.03926732635498047, 0.03929702377319336, 0.03910041427612305, 0.03915161514282227, 0.03912908935546875, 0.03910860824584961, 0.03913011169433594, 0.03911270523071289, 0.03912089538574219, 0.03912908935546875, 0.03913216018676758, 0.03912908935546875, 0.03913318252563477, 0.039155712127685545, 0.03909222412109375, 0.039223297119140625, 0.039548992156982425, 0.03942598342895508, 0.03920588684082031, 0.0391464958190918, 0.03917311859130859, 0.039191551208496093, 0.03910758590698242, 0.039174144744873046, 0.039160831451416016, 0.039155712127685545, 0.03912704086303711, 0.03915776062011719, 0.03916697692871094, 0.03914547348022461, 0.039139328002929685, 0.03917004776000976, 0.0391383056640625, 0.039128063201904296, 0.039103488922119144, 0.03917311859130859, 0.03917619323730469, 0.03918643188476562, 0.03911577606201172, 0.03916185760498047, 0.039155712127685545, 0.03920281600952148, 0.03922431945800781, 0.039293952941894535, 0.03921100616455078, 0.039198719024658206, 0.03914342498779297, 0.03920793533325195, 0.039163902282714845, 0.03918950271606445, 0.039147518157958985, 0.03918643188476562, 0.039174144744873046, 0.03922022247314453, 0.039180286407470705, 0.03917824172973633, 0.03917004776000976, 0.12378623962402344, 0.03908201599121094, 0.039092193603515624, 0.03911270523071289, 0.039057407379150394, 0.03912499237060547, 0.03911270523071289, 0.03914035034179687, 0.039126014709472655, 0.03922739028930664, 0.039139328002929685, 0.039139328002929685, 0.03917619323730469, 0.03913113784790039, 0.03914547348022461, 0.039126014709472655, 0.03917622375488281, 0.039481311798095706, 0.03928575897216797, 0.039213054656982424, 0.039221248626708984, 0.039136257171630856, 0.039163902282714845, 0.039180286407470705, 0.039137279510498044, 0.03912704086303711, 0.03911679840087891, 0.03915776062011719, 0.03907993698120117, 0.039206912994384766, 0.03909939193725586, 0.03916697692871094, 0.03913318252563477, 0.0391536636352539, 0.03914547348022461, 0.03917004776000976, 0.03913420867919922, 0.039158782958984374, 0.03914035034179687, 0.039169025421142575, 0.03908816146850586, 0.039206878662109375, 0.03934003067016602, 0.03942502212524414, 0.03931545639038086, 0.03927961730957031, 0.039147518157958985, 0.03921408081054688, 0.03916799926757813, 0.039223297119140625, 0.03913216018676758, 0.039201793670654295, 0.03916697692871094, 0.039201793670654295, 0.0391833610534668, 0.039242752075195314, 0.039182334899902346, 0.03919257736206055, 0.0391833610534668, 0.0392540168762207, 0.03917619323730469, 0.03918438339233398, 0.03917619323730469, 0.12409446716308593, 0.03913318252563477, 0.0391383056640625, 0.03911475372314453, 0.03907993698120117, 0.039136257171630856, 0.03911065673828125, 0.03911167907714844, 0.039065601348876954, 0.0391014404296875, 0.03911475372314453, 0.0391464958190918, 0.03916799926757813, 0.03910351943969727, 0.0391126708984375, 0.03915059280395508, 0.03915980911254883, 0.03917004776000976, 0.03912704086303711, 0.03913011169433594, 0.039155712127685545, 0.03922227096557617, 0.039229438781738284, 0.03910860824584961, 0.03919462585449219, 0.039213054656982424, 0.03926528167724609, 0.039226367950439454, 0.039128063201904296, 0.03914342498779297, 0.03914652633666992, 0.039132129669189455, 0.039160831451416016, 0.03914956665039063, 0.03916595077514649, 0.039219230651855466, 0.039118816375732425, 0.03914547348022461, 0.03911270523071289, 0.039142398834228515, 0.03912704086303711, 0.039163902282714845, 0.03920800018310547, 0.0391607666015625, 0.03915980911254883, 0.03916287994384766, 0.039228416442871096, 0.03922227096557617, 0.03915161514282227, 0.03918643188476562, 0.0391649284362793, 0.03920793533325195, 0.039144447326660156, 0.039191551208496093, 0.03917004776000976, 0.03919257736206055, 0.03917011260986328, 0.03916896057128906, 0.03914956665039063, 0.03922534561157227, 0.039136257171630856, 0.039169025421142575, 0.03914547348022461, 0.12379647827148438, 0.03911577606201172, 0.039103488922119144, 0.0391014404296875, 0.03907583999633789, 0.03909836959838867, 0.03920383834838867, 0.03918745422363281, 0.03910451126098633, 0.0391464958190918, 0.0391649284362793, 0.03919257736206055, 0.03918131256103516, 0.03912707138061523, 0.039118816375732425, 0.03914854431152344, 0.03911577606201172, 0.039160831451416016, 0.03913420867919922, 0.039185409545898435, 0.039160831451416016, 0.0391383056640625, 0.03912908935546875, 0.039126014709472655, 0.03915980911254883, 0.03923353576660156, 0.03917004776000976, 0.039137279510498044, 0.039093246459960936, 0.039177215576171875, 0.03914956665039063, 0.03917311859130859, 0.03916799926757813, 0.039158782958984374, 0.03913318252563477, 0.03917619323730469, 0.03915673446655273, 0.039196670532226564, 0.03917004776000976, 0.039169025421142575, 0.03913011169433594, 0.039158782958984374, 0.03914547348022461, 0.03922227096557617, 0.039259136199951174, 0.03916185760498047, 0.03916595077514649, 0.039209983825683595, 0.039152641296386716, 0.039193599700927735, 0.039155712127685545, 0.039272449493408204, 0.03914854431152344, 0.03920588684082031, 0.039155712127685545, 0.03921001434326172, 0.03915670394897461, 0.03917926406860352, 0.03925100708007812, 0.039238590240478516, 0.039191551208496093, 0.03920793533325195, 0.03922227096557617, 0.12371250915527343, 0.039376895904541014, 0.03933491134643555, 0.039190528869628906, 0.039322654724121095, 0.03914441680908203, 0.039177215576171875, 0.03915980911254883, 0.03909939193725586, 0.03913523101806641, 0.03908607864379883, 0.03915059280395508, 0.039128063201904296, 0.0391014404296875, 0.03911577606201172, 0.03913216018676758, 0.039158782958984374, 0.039259136199951174, 0.039375873565673826, 0.03920793533325195, 0.03910758590698242, 0.03923353576660156, 0.03944345474243164, 0.039198719024658206, 0.03917004776000976, 0.03912089538574219, 0.03913318252563477, 0.03913113784790039, 0.039198719024658206, 0.039137279510498044, 0.039174144744873046, 0.03918131256103516, 0.039174144744873046, 0.039147518157958985, 0.039215103149414066, 0.039231487274169925, 0.03923353576660156, 0.039228416442871096, 0.039288833618164064, 0.03917926406860352, 0.039226367950439454, 0.039191551208496093, 0.03924889755249023, 0.03932876968383789, 0.039191551208496093, 0.039163902282714845, 0.039218177795410154, 0.039213054656982424, 0.039288833618164064, 0.039370750427246096, 0.039198719024658206, 0.03923763275146484, 0.03922534561157227, 0.03925708770751953, 0.03916697692871094, 0.03918950271606445, 0.039177215576171875, 0.039175201416015625, 0.039152606964111325, 0.039218177795410154, 0.03915673446655273, 0.039218177795410154, 0.039190528869628906, 0.12386815643310548, 0.039094337463378905, 0.03917407989501953, 0.03915059280395508, 0.039258113861083986, 0.03911884689331055, 0.03912089538574219, 0.039137279510498044, 0.039163902282714845, 0.039142398834228515, 0.03914547348022461, 0.03942607879638672, 0.039961566925048826, 0.03926835250854492, 0.03912396621704101, 0.03916185760498047, 0.039185409545898435, 0.03914137649536133, 0.03912396621704101, 0.03916287994384766, 0.03914547348022461, 0.03911372756958008, 0.03912089538574219, 0.03910451126098633, 0.039139328002929685, 0.03911884689331055, 0.03911475372314453, 0.03911679840087891, 0.039103488922119144, 0.0391536636352539, 0.03911679840087891, 0.039163902282714845, 0.03912192153930664, 0.03915673446655273, 0.039163902282714845, 0.039215103149414066, 0.03921920013427734, 0.03940966415405273, 0.039332862854003905, 0.039191551208496093, 0.03911679840087891, 0.039136257171630856, 0.039198719024658206, 0.0393779182434082, 0.03918131256103516, 0.03917107009887695, 0.039137279510498044, 0.03919974517822265, 0.039161888122558594, 0.0392130241394043, 0.03911577606201172, 0.03921100616455078, 0.03923353576660156, 0.039272449493408204, 0.03918745422363281, 0.039213054656982424, 0.03918643188476562, 0.03921408081054688, 0.039196670532226564, 0.03926015853881836, 0.03918438339233398, 0.03918643188476562, 0.039185409545898435, 0.12357017517089844, 0.039090175628662106, 0.03908095932006836, 0.039093246459960936, 0.03905843353271484, 0.03910246276855469, 0.03909529495239258, 0.03912192153930664, 0.03912704086303711, 0.03913420867919922, 0.0391383056640625, 0.03913523101806641, 0.03909632110595703, 0.039122943878173826, 0.03914342498779297, 0.03917926406860352, 0.03911577606201172, 0.039175167083740234, 0.03916595077514649, 0.03930214309692383, 0.03920076751708984, 0.03917926406860352, 0.039144447326660156, 0.03917311859130859, 0.03914854431152344, 0.03912704086303711, 0.039139328002929685, 0.03912908935546875, 0.0390830078125, 0.03913523101806641, 0.03916595077514649, 0.0391475830078125, 0.039129024505615236, 0.03917311859130859, 0.03915161514282227, 0.039185409545898435, 0.039147518157958985, 0.03920896148681641, 0.03917107009887695, 0.039160831451416016, 0.039185409545898435, 0.039163902282714845, 0.0391464958190918, 0.03917619323730469, 0.039190528869628906, 0.03915468978881836, 0.039163902282714845, 0.039221248626708984, 0.039147518157958985, 0.039155712127685545, 0.03916287994384766, 0.03917107009887695, 0.03917824172973633, 0.039193599700927735, 0.03919257736206055, 0.039190528869628906, 0.03919462585449219, 0.03918438339233398, 0.03918131256103516, 0.03920588684082031, 0.039174144744873046, 0.03919564819335938, 0.0391833610534668, 0.12353126525878906, 0.03917107009887695, 0.03911782455444336, 0.0391464958190918, 0.039087104797363284, 0.03915161514282227, 0.039288833618164064, 0.03912192153930664, 0.039090175628662106, 0.039126014709472655, 0.03915776062011719, 0.03911372756958008, 0.03912396621704101, 0.03911065673828125, 0.03911679840087891, 0.03909529495239258, 0.03910451126098633, 0.03914451217651367, 0.039094207763671875, 0.03919769668579102, 0.03924991989135742, 0.03915161514282227, 0.039144447326660156, 0.03916185760498047, 0.03915468978881836, 0.03910758590698242, 0.039163902282714845, 0.039144447326660156, 0.039122943878173826, 0.03916185760498047, 0.039122943878173826, 0.03919564819335938, 0.03917004776000976, 0.039174144744873046, 0.03916697692871094, 0.03916697692871094, 0.03916287994384766, 0.03918950271606445, 0.03917824172973633, 0.039206912994384766, 0.03910246276855469, 0.03918438339233398, 0.03920383834838867, 0.03918438339233398, 0.03912704086303711, 0.03917107009887695, 0.03912704086303711, 0.039204864501953124, 0.039172096252441405, 0.039182334899902346, 0.039158782958984374, 0.03922227096557617, 0.0392806396484375, 0.039206912994384766, 0.03917619323730469, 0.03922431945800781, 0.039160831451416016, 0.039204864501953124, 0.039180286407470705, 0.039231487274169925, 0.03948441696166992, 0.0397127685546875, 0.03921408081054688]",tokens/s,24.75481313103816,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen2-beta-14B,Qwen/Qwen2-beta-14B,cuda,0,42,,,True,,,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",qwen2,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-32B,Qwen/Qwen1.5-32B,cuda,0,42,,,True,,,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",qwen2,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,0,0,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/0/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669325a8-212a78595d04b9e02ce5edcc;d5fe5847-b092-469e-b3fc-ba7b1252c9a9) Repository Not Found for url: https://huggingface.co/0/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: 0 is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,2,2,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/2/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669324db-3fb2e96c1cf8ea1524ed1076;20f2f0ff-8680-4a55-8c2a-a17ba66b7097) Repository Not Found for url: https://huggingface.co/2/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: 2 is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/xglm-564M,facebook/xglm-564M,cuda,0,42,,,True,,,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",xglm,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-14B,Qwen/Qwen-14B,cuda,0,42,,,True,,,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",qwen,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,huggyllama/llama-7b,huggyllama/llama-7b,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.1,,,,MB,1235.922944,14529.593344,0.0,13883.14624,13705.186304,s,10,0.9381064758300781,0.0938106475830078,0.0002782777167609782,0.0937690544128418,0.09394051055908202,0.09426363220214844,0.09452212951660156,"[0.09458675384521484, 0.093621826171875, 0.09378880310058593, 0.09377235412597656, 0.09377174377441407, 0.09347682952880859, 0.09386870574951171, 0.09369475555419922, 0.09376636505126953, 0.09375833892822266]",tokens/s,2728.901319794002,kWh,1.1091462254276654e-06,6.077594067647053e-07,5.774441275831742e-06,7.4913469080241126e-06,tokens/kWh,34172760.004718766,MB,1235.922944,14529.593344,0.0,13883.14624,13856.22528,s,10,22.542584716796878,2.254258471679688,0.003272167817286955,2.2532373046875,2.2557274169921877,2.259641296386719,2.262772399902344,"[2.26355517578125, 2.25456201171875, 2.253303466796875, 2.254278564453125, 2.252525146484375, 2.253171142578125, 2.25120556640625, 2.254857666015625, 2.25266845703125, 2.25245751953125]",tokens/s,27.947105796194517,kWh,2.6553851566736822e-05,1.455239421363509e-05,0.0001132887095301677,0.00015439495531053963,tokens/kWh,408044.42006078525,,s,629,23.349358585357674,0.037121396797071014,0.011210628935323555,0.03575497436523437,0.03586355209350586,0.035984383392333986,0.13010292602539061,"[0.03647084808349609, 0.03641644668579101, 0.03629056167602539, 0.03638272094726563, 0.03651686477661133, 0.036311038970947264, 0.036367359161376955, 0.03647795104980469, 0.03638272094726563, 0.036400127410888675, 0.03618201446533203, 0.03621068954467774, 0.03609907150268555, 0.035798015594482424, 0.03578572845458984, 0.03580723190307617, 0.03576934432983398, 0.03578777694702148, 0.03570483016967774, 0.03568742370605469, 0.035789825439453124, 0.03580108642578125, 0.0357130241394043, 0.035705856323242184, 0.03575603103637695, 0.03582156753540039, 0.03575807952880859, 0.03580825424194336, 0.035705856323242184, 0.035724288940429685, 0.03573555374145508, 0.035716094970703126, 0.035724288940429685, 0.03574169540405273, 0.03575296020507813, 0.03576934432983398, 0.0357314567565918, 0.03577753448486328, 0.03576934432983398, 0.035813377380371096, 0.03587071990966797, 0.0359024658203125, 0.03580108642578125, 0.03586355209350586, 0.03577036666870117, 0.03579391860961914, 0.035866622924804685, 0.035862529754638675, 0.03592704010009766, 0.03588608169555664, 0.03587276840209961, 0.035783679962158206, 0.035794944763183595, 0.03580723190307617, 0.035860481262207033, 0.03587891387939453, 0.035773441314697264, 0.035846145629882815, 0.03582668685913086, 0.035781631469726564, 0.035813377380371096, 0.03589836883544922, 0.13003878784179687, 0.03566592025756836, 0.0356577262878418, 0.03570995330810547, 0.035710975646972655, 0.0358389778137207, 0.03573964691162109, 0.03566899108886719, 0.03567308807373047, 0.03573964691162109, 0.03577241516113281, 0.03572332763671875, 0.03569452667236328, 0.03570175933837891, 0.035947521209716796, 0.035970046997070314, 0.035806209564208984, 0.03571916961669922, 0.03567718505859375, 0.035716094970703126, 0.03569356918334961, 0.035722240447998044, 0.03570278549194336, 0.035736640930175784, 0.03576518249511719, 0.03578060913085938, 0.035757057189941405, 0.03576115036010742, 0.03575603103637695, 0.0357283821105957, 0.03575398254394531, 0.035781631469726564, 0.035757057189941405, 0.03574272155761719, 0.03570995330810547, 0.03574272155761719, 0.03575193786621094, 0.035768318176269534, 0.03573350524902344, 0.03586764907836914, 0.03602739334106445, 0.03579596710205078, 0.03579289627075195, 0.0358021125793457, 0.0357918701171875, 0.035794944763183595, 0.03586355209350586, 0.03580416107177734, 0.03577753448486328, 0.035781631469726564, 0.035775489807128906, 0.03579391860961914, 0.03581542587280274, 0.035784702301025394, 0.03577753448486328, 0.03578879928588867, 0.03580723190307617, 0.03581951904296875, 0.035811328887939455, 0.036013057708740234, 0.03587481689453125, 0.035896320343017575, 0.035860481262207033, 0.13015245056152344, 0.03561471939086914, 0.03559526443481445, 0.03567001724243164, 0.03563417434692383, 0.035643390655517575, 0.035642368316650394, 0.0356495361328125, 0.035707904815673826, 0.035655681610107424, 0.03566796875, 0.035642368316650394, 0.03563622283935547, 0.03566592025756836, 0.03565055847167969, 0.035798015594482424, 0.036050945281982424, 0.03570073699951172, 0.03568230438232422, 0.035655681610107424, 0.035746814727783204, 0.0358389778137207, 0.035789825439453124, 0.035724288940429685, 0.03571814346313477, 0.0357283821105957, 0.035776512145996094, 0.03575807952880859, 0.035789825439453124, 0.035796993255615236, 0.03573452758789063, 0.035783679962158206, 0.035953662872314454, 0.03597318267822266, 0.03580307388305664, 0.035732479095458985, 0.0357314567565918, 0.03570175933837891, 0.03574272155761719, 0.03571507263183594, 0.03573555374145508, 0.03575398254394531, 0.03574169540405273, 0.03575296020507813, 0.03571712112426758, 0.03571916961669922, 0.035748863220214845, 0.035748863220214845, 0.03582156753540039, 0.03580928039550781, 0.03580313491821289, 0.035860481262207033, 0.035781631469726564, 0.035759105682373046, 0.035746814727783204, 0.03577753448486328, 0.035773441314697264, 0.03576627349853516, 0.035858463287353516, 0.035872737884521486, 0.036124671936035156, 0.035975166320800785, 0.03581951904296875, 0.13026611328125, 0.03571507263183594, 0.035800064086914066, 0.035724288940429685, 0.03587583923339844, 0.03587481689453125, 0.03567923355102539, 0.035694591522216795, 0.035746814727783204, 0.035707904815673826, 0.03571712112426758, 0.035745792388916016, 0.03571814346313477, 0.035683326721191407, 0.035705856323242184, 0.03571507263183594, 0.03572531127929687, 0.035683326721191407, 0.03578265762329102, 0.035705856323242184, 0.03569356918334961, 0.0357386245727539, 0.035745792388916016, 0.03572531127929687, 0.035705856323242184, 0.03573350524902344, 0.035724288940429685, 0.03572326278686523, 0.03576319885253906, 0.035737598419189456, 0.035762176513671876, 0.03577036666870117, 0.035757057189941405, 0.03575296020507813, 0.035743743896484374, 0.03577036666870117, 0.03576319885253906, 0.03577753448486328, 0.03580928039550781, 0.03579084777832031, 0.0358021125793457, 0.03583180618286133, 0.03582156753540039, 0.035787841796875, 0.035798976898193356, 0.035789825439453124, 0.03582361602783203, 0.03581235122680664, 0.035829761505126956, 0.03577753448486328, 0.035814399719238284, 0.035811328887939455, 0.03581545639038086, 0.03579388809204102, 0.03580723190307617, 0.03586867141723633, 0.03581951904296875, 0.0358287353515625, 0.035871742248535156, 0.03583078384399414, 0.03584102249145508, 0.03585331344604492, 0.03583283233642578, 0.13040640258789063, 0.03571712112426758, 0.03564646530151367, 0.03563315200805664, 0.0356577262878418, 0.03562905502319336, 0.035694591522216795, 0.03573964691162109, 0.03569356918334961, 0.035683326721191407, 0.03568435287475586, 0.03564851379394531, 0.03565875244140625, 0.03563731384277344, 0.03567814254760742, 0.0356864013671875, 0.03572326278686523, 0.035778560638427735, 0.03568025588989258, 0.03568742370605469, 0.03575091171264649, 0.03570278549194336, 0.03570483016967774, 0.035705856323242184, 0.03568537521362305, 0.035707904815673826, 0.03571712112426758, 0.03571916961669922, 0.035729408264160156, 0.0357386245727539, 0.03572326278686523, 0.0356864013671875, 0.03577753448486328, 0.03573452758789063, 0.03576115036010742, 0.035781631469726564, 0.035746849060058594, 0.03575497436523437, 0.03571916961669922, 0.03576115036010742, 0.035746814727783204, 0.035746814727783204, 0.035760128021240234, 0.03577958297729492, 0.035760128021240234, 0.03577859115600586, 0.035772384643554686, 0.03577958297729492, 0.035776512145996094, 0.035789825439453124, 0.03580416107177734, 0.035798015594482424, 0.03583692932128906, 0.03582156753540039, 0.0358205451965332, 0.03581849670410156, 0.03580416107177734, 0.0358389778137207, 0.0357918701171875, 0.035800064086914066, 0.035811328887939455, 0.03583795166015625, 0.035871742248535156, 0.12986778259277343, 0.03565158462524414, 0.03564748764038086, 0.03563212966918945, 0.03562700653076172, 0.03564748764038086, 0.035659774780273434, 0.035661823272705076, 0.035681278228759765, 0.03568435287475586, 0.035675136566162106, 0.03571507263183594, 0.035691520690917966, 0.035699710845947266, 0.03566284942626953, 0.035659774780273434, 0.0356577262878418, 0.03569868850708008, 0.035683326721191407, 0.03567718505859375, 0.03572022247314453, 0.03574166488647461, 0.03575296020507813, 0.035692543029785154, 0.03572633743286133, 0.03575193786621094, 0.03573964691162109, 0.0357314567565918, 0.0357386245727539, 0.035730430603027344, 0.03572531127929687, 0.03599052810668945, 0.03640422439575195, 0.03587071990966797, 0.03575091171264649, 0.03575296020507813, 0.035727359771728515, 0.035746814727783204, 0.03580108642578125, 0.03573964691162109, 0.03577139282226562, 0.035760128021240234, 0.03574784088134766, 0.035745792388916016, 0.03574784088134766, 0.03583795166015625, 0.035778560638427735, 0.035768318176269534, 0.03581951904296875, 0.03581849670410156, 0.035765247344970705, 0.03578675079345703, 0.03579289627075195, 0.035871742248535156, 0.035794944763183595, 0.03578060913085938, 0.03580108642578125, 0.035790878295898436, 0.03583894348144531, 0.035776512145996094, 0.035807262420654296, 0.0358502082824707, 0.035814399719238284, 0.13022720336914062, 0.03561471939086914, 0.03560038375854492, 0.03563724899291992, 0.035620864868164064, 0.03561471939086914, 0.03560345458984375, 0.03562188720703125, 0.03568537521362305, 0.03563212966918945, 0.035729408264160156, 0.035694591522216795, 0.03567001724243164, 0.03566284942626953, 0.03563008117675781, 0.03565158462524414, 0.035659774780273434, 0.0356495361328125, 0.03568742370605469, 0.035655681610107424, 0.03567001724243164, 0.03567923355102539, 0.03569664001464844, 0.0356864013671875, 0.03563417434692383, 0.035681278228759765, 0.03568435287475586, 0.03567718505859375, 0.0357314567565918, 0.03568025588989258, 0.03571814346313477, 0.0357130241394043, 0.035724288940429685, 0.03574476623535156, 0.035781631469726564, 0.03584204864501953, 0.0357498893737793, 0.0357386245727539, 0.035767295837402346, 0.035710975646972655, 0.03573350524902344, 0.035737598419189456, 0.03573350524902344, 0.03572326278686523, 0.035716094970703126, 0.03574169540405273, 0.03575193786621094, 0.035737598419189456, 0.035915775299072264, 0.03585539245605469, 0.03604169464111328, 0.03591372680664062, 0.03579391860961914, 0.03576934432983398, 0.03583078384399414, 0.03580108642578125, 0.03580416107177734, 0.035767295837402346, 0.03578265762329102, 0.035748863220214845, 0.035784702301025394, 0.03578675079345703, 0.03578060913085938, 0.13018418884277344, 0.035768318176269534, 0.035722240447998044, 0.03574476623535156, 0.0357130241394043, 0.03566796875, 0.03569868850708008, 0.0357283821105957, 0.03570278549194336, 0.0356864013671875, 0.03573350524902344, 0.03586150360107422, 0.03599871826171875, 0.035716094970703126, 0.03570995330810547, 0.03573555374145508, 0.03574476623535156, 0.03578988647460937, 0.035861438751220706, 0.035762176513671876, 0.03576319885253906, 0.035716094970703126, 0.0357130241394043, 0.0357283821105957, 0.0357314567565918, 0.035727359771728515, 0.035732479095458985, 0.035746814727783204, 0.035773441314697264, 0.03580518341064453, 0.03574784088134766, 0.03577446365356445, 0.03582566452026367, 0.035759105682373046, 0.035768318176269534, 0.03580518341064453, 0.03578060913085938, 0.035773441314697264, 0.035745792388916016, 0.0357498893737793, 0.035768318176269534, 0.035768318176269534, 0.03577241516113281, 0.035767295837402346, 0.03578572845458984, 0.03582463836669922, 0.0358021125793457, 0.03585126495361328, 0.03594342422485351, 0.035811328887939455, 0.03584000015258789, 0.03582668685913086, 0.03584000015258789, 0.035806209564208984, 0.03582361602783203, 0.035846145629882815, 0.03583180618286133, 0.03582156753540039, 0.035813377380371096, 0.0358389778137207, 0.03583283233642578, 0.035811328887939455, 0.03585433578491211, 0.13020364379882812, 0.03566284942626953, 0.035664894104003905, 0.03565670394897461, 0.03565158462524414, 0.03565465545654297, 0.03566899108886719, 0.03567103958129883, 0.035688449859619144, 0.035727359771728515, 0.035694591522216795, 0.03566694259643555, 0.03567001724243164, 0.03567103958129883, 0.03567001724243164, 0.03567308807373047, 0.03578675079345703, 0.036073471069335936, 0.03572326278686523, 0.03570687866210937, 0.03570995330810547, 0.035697662353515625, 0.035707904815673826, 0.035681278228759765, 0.035716094970703126, 0.035729408264160156, 0.035737598419189456, 0.03573452758789063, 0.035732479095458985, 0.035724288940429685, 0.03587481689453125, 0.035776512145996094, 0.03576115036010742, 0.035737598419189456, 0.03573964691162109, 0.0357386245727539, 0.035714046478271484, 0.035740673065185545, 0.03577241516113281, 0.035724288940429685, 0.035745792388916016, 0.03577036666870117, 0.035745792388916016, 0.03576319885253906, 0.03575807952880859, 0.03577036666870117, 0.035783679962158206, 0.03578060913085938, 0.035829761505126956, 0.035783679962158206, 0.03579596710205078, 0.035798015594482424, 0.03577958297729492, 0.03577139282226562, 0.03580825424194336, 0.03579904174804688, 0.035810302734375, 0.035773441314697264, 0.03583488082885742, 0.035776512145996094, 0.035814399719238284, 0.03582361602783203, 0.03582566452026367, 0.13012786865234374, 0.03561779022216797, 0.03565468978881836, 0.03567715072631836, 0.03570892715454101, 0.03566387176513672, 0.035659774780273434, 0.03562905502319336, 0.03569561767578125, 0.0356495361328125, 0.03569664001464844, 0.035717151641845704, 0.03570070266723633, 0.03565363311767578, 0.035642368316650394, 0.03568025588989258, 0.035716129302978517, 0.035691486358642575, 0.03568435287475586, 0.03567308807373047, 0.03567411041259766, 0.035745792388916016, 0.03569664001464844, 0.035676158905029294, 0.035675136566162106, 0.035699710845947266, 0.03580108642578125, 0.03570278549194336, 0.035765247344970705, 0.03571507263183594, 0.03572531127929687, 0.035759105682373046, 0.03572531127929687, 0.03571507263183594, 0.03572531127929687, 0.035746814727783204, 0.035746814727783204, 0.035745792388916016, 0.03577753448486328, 0.03574169540405273, 0.035745792388916016, 0.03575500869750976, 0.035746814727783204, 0.03574169540405273, 0.035727359771728515, 0.03577241516113281, 0.03578879928588867, 0.03579391860961914, 0.03588608169555664, 0.03607654571533203, 0.035915775299072264, 0.03580416107177734, 0.03579289627075195, 0.03583180618286133, 0.035757057189941405, 0.03579596710205078, 0.03581644821166992, 0.03581235122680664, 0.035846145629882815, 0.035806209564208984, 0.03584000015258789, 0.03584716796875, 0.03585638427734375]",tokens/s,26.93864149203844,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-180B,tiiuae/falcon-180B,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 403 Client Error: Forbidden for url: https://huggingface.co/tiiuae/falcon-180B/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-66932713-25deaa396cca7682646cd378;b5e31215-7453-4bae-90eb-55027004c9e1) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/tiiuae/falcon-180B/resolve/main/config.json. If you are trying to create or update content,make sure you have a token with the `write` role. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1826, in _raise_on_head_call_error raise LocalEntryNotFoundError( huggingface_hub.utils._errors.LocalEntryNotFoundError: An error happened while trying to locate the file on the Hub and we cannot find the requested files in the local cache. Please check your connection and try again or make sure your Internet connection is on. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 445, in cached_file raise EnvironmentError( OSError: We couldn't connect to 'https://huggingface.co' to load this file, couldn't find it in the cached files and it looks like tiiuae/falcon-180B is not the path to a directory containing a file named config.json. Checkout your internet connection or see how to run the library in offline mode at 'https://huggingface.co/docs/transformers/installation#offline-mode'. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,mistralai/Mistral-7B-v0.1,mistralai/Mistral-7B-v0.1,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.1,,,,MB,1271.13216,15987.113984,0.0,15340.66688,15114.865664,s,10,0.8037002868652342,0.08037002868652342,0.0003574703869442419,0.08033499145507812,0.08058157730102539,0.08096539497375488,0.08127244911193847,"[0.08134921264648437, 0.0801430435180664, 0.08035801696777344, 0.08001481628417968, 0.08018147277832031, 0.08037398529052735, 0.08031196594238281, 0.08039427185058594, 0.08049628448486328, 0.08007721710205078]",tokens/s,3185.266997956497,kWh,9.520324627558391e-07,5.215199259822839e-07,6.268378348032178e-06,7.741930736770301e-06,tokens/kWh,33066686.9420735,MB,1271.13216,15987.113984,0.0,15340.66688,15134.09792,s,10,23.448356689453124,2.3448356689453123,0.0035437284207401275,2.343679931640625,2.3452836914062503,2.3503582763671873,2.354417944335937,"[2.355432861328125, 2.343271728515625, 2.34377294921875, 2.343294677734375, 2.3435869140625, 2.344156005859375, 2.343521240234375, 2.34394873046875, 2.343412353515625, 2.343959228515625]",tokens/s,26.867554445014427,kWh,2.7613349633746683e-05,1.5133218477572519e-05,0.00011971988910916498,0.00016246645722048418,tokens/kWh,387772.3505381934,,s,629,24.132534248352027,0.03836650913887449,0.009602659628695916,0.037184513092041016,0.03728957366943359,0.03742986145019531,0.11792289367675782,"[0.0379607048034668, 0.03787571334838867, 0.037904384613037106, 0.03789004898071289, 0.03793100738525391, 0.03784601593017578, 0.03800064086914062, 0.037868545532226565, 0.037926910400390625, 0.03774771118164062, 0.03784396743774414, 0.03759001541137695, 0.03736678314208984, 0.03720499038696289, 0.03715686416625977, 0.03713536071777344, 0.03724595260620117, 0.03711078262329102, 0.037160961151123044, 0.037160961151123044, 0.03719987106323242, 0.03719270324707031, 0.037152767181396484, 0.03720499038696289, 0.03717529678344727, 0.03724595260620117, 0.03743231964111328, 0.037178432464599606, 0.03721721649169922, 0.03718143844604492, 0.03718041610717773, 0.03719782257080078, 0.037185535430908204, 0.03725823974609375, 0.037359615325927735, 0.03744255828857422, 0.03733606338500976, 0.037200958251953124, 0.03721823883056641, 0.037200897216796876, 0.03724185562133789, 0.037184513092041016, 0.03724288177490234, 0.03715891265869141, 0.037178398132324215, 0.03718857574462891, 0.037525505065917966, 0.037544960021972655, 0.03722751998901367, 0.037203968048095705, 0.03720601654052735, 0.03718041610717773, 0.03735756683349609, 0.037292030334472655, 0.037292030334472655, 0.03727462387084961, 0.037305343627929685, 0.03723571014404297, 0.03732787322998047, 0.03732787322998047, 0.03728384017944336, 0.037389312744140625, 0.11799346923828125, 0.03718348693847656, 0.0371517448425293, 0.03714355087280274, 0.037116928100585936, 0.03715584182739258, 0.037166080474853515, 0.03720499038696289, 0.03715689468383789, 0.037124065399169924, 0.03720908737182617, 0.03714559936523437, 0.037133312225341795, 0.037133312225341795, 0.03724288177490234, 0.03729919815063477, 0.03710976028442383, 0.03714355087280274, 0.03711283111572266, 0.037179393768310545, 0.03712921524047851, 0.03712307357788086, 0.037179393768310545, 0.037169185638427735, 0.03717014312744141, 0.037200897216796876, 0.03721420669555664, 0.03718041610717773, 0.03717529678344727, 0.03717529678344727, 0.0371486701965332, 0.0371701774597168, 0.03720908737182617, 0.03715891265869141, 0.0371517448425293, 0.037187583923339845, 0.037169151306152344, 0.037175361633300784, 0.03716396713256836, 0.037168128967285156, 0.03715686416625977, 0.03717123031616211, 0.037168094635009766, 0.03728486251831055, 0.03719372940063476, 0.03722547149658203, 0.0372305908203125, 0.037207073211669925, 0.03717424011230469, 0.037292030334472655, 0.03724697494506836, 0.03722444915771484, 0.03720908737182617, 0.03720806503295898, 0.037174270629882815, 0.03720908737182617, 0.03723775863647461, 0.03726847839355469, 0.037184513092041016, 0.0372674560546875, 0.03718656158447266, 0.0372408332824707, 0.037236736297607424, 0.11812249755859375, 0.037168128967285156, 0.03712307357788086, 0.03723980712890625, 0.03711385726928711, 0.03719168090820312, 0.037141502380371096, 0.037100543975830076, 0.03712921524047851, 0.0371701774597168, 0.03715686416625977, 0.03740364837646484, 0.037160961151123044, 0.03712614440917969, 0.037133312225341795, 0.037212158203125, 0.03712921524047851, 0.03714355087280274, 0.03714048004150391, 0.03718963241577149, 0.03715584182739258, 0.037235774993896485, 0.03716089630126953, 0.03715379333496094, 0.03709952163696289, 0.03715686416625977, 0.03717324829101563, 0.037176319122314457, 0.03718656158447266, 0.037171199798583986, 0.037131263732910154, 0.037166080474853515, 0.03723775863647461, 0.037195777893066405, 0.03724185562133789, 0.0371671028137207, 0.037174270629882815, 0.037190654754638675, 0.03726131057739258, 0.03722137451171875, 0.037190654754638675, 0.03721219253540039, 0.03716912078857422, 0.0372408332824707, 0.037223423004150394, 0.03726131057739258, 0.03720499038696289, 0.037256191253662106, 0.03718656158447266, 0.037253120422363284, 0.0372408332824707, 0.037392383575439454, 0.03720908737182617, 0.03720294570922852, 0.03724185562133789, 0.037215232849121094, 0.037185535430908204, 0.03719680023193359, 0.037223423004150394, 0.03724492645263672, 0.037203968048095705, 0.03721011352539062, 0.037233665466308595, 0.11782860565185548, 0.03710566329956055, 0.03710464096069336, 0.037114879608154294, 0.03715071868896484, 0.037146625518798826, 0.03718041610717773, 0.03712204742431641, 0.03711283111572266, 0.0371671028137207, 0.03709952163696289, 0.03718143844604492, 0.037187583923339845, 0.03724185562133789, 0.0371599349975586, 0.037147647857666014, 0.03713945770263672, 0.037179393768310545, 0.03719372940063476, 0.03725209426879883, 0.03719475173950195, 0.037154815673828126, 0.037130241394042966, 0.037168128967285156, 0.03715071868896484, 0.03715891265869141, 0.03723980712890625, 0.037200897216796876, 0.037207038879394534, 0.03742617416381836, 0.03723571014404297, 0.037168128967285156, 0.03718143844604492, 0.037169151306152344, 0.037171199798583986, 0.03716304016113281, 0.037168094635009766, 0.03718143844604492, 0.03714048004150391, 0.03721625518798828, 0.037138431549072266, 0.03718656158447266, 0.03718041610717773, 0.0372305908203125, 0.03718041610717773, 0.03718143844604492, 0.037207038879394534, 0.03723161697387695, 0.03720499038696289, 0.03721011352539062, 0.037168128967285156, 0.03719372940063476, 0.037166080474853515, 0.03730944061279297, 0.03726131057739258, 0.03718143844604492, 0.03717836761474609, 0.037177345275878904, 0.03726028823852539, 0.03723775863647461, 0.0371701774597168, 0.03722547149658203, 0.03719071960449219, 0.11793606567382812, 0.037101566314697264, 0.03714559936523437, 0.037106689453125, 0.037116928100585936, 0.037106689453125, 0.037154815673828126, 0.03709132766723633, 0.03712102508544922, 0.037168128967285156, 0.03712819290161133, 0.037163009643554686, 0.03713740921020508, 0.03715379333496094, 0.03709952163696289, 0.037106689453125, 0.03719475173950195, 0.037176319122314457, 0.037157886505126955, 0.03720297622680664, 0.037209056854248045, 0.037184513092041016, 0.03714355087280274, 0.037185535430908204, 0.03718143844604492, 0.03713433456420898, 0.03715584182739258, 0.037157886505126955, 0.037146625518798826, 0.037154815673828126, 0.03718656158447266, 0.037154815673828126, 0.03717836761474609, 0.03715071868896484, 0.037182464599609374, 0.03727052688598633, 0.037364734649658206, 0.03717529678344727, 0.037144577026367184, 0.03720601654052735, 0.03717324829101563, 0.037220352172851565, 0.0371701774597168, 0.03728076934814453, 0.03742617416381836, 0.037395454406738284, 0.03726335906982422, 0.037233665466308595, 0.037198848724365234, 0.03717222213745117, 0.037223423004150394, 0.03724288177490234, 0.037198848724365234, 0.037215232849121094, 0.03718860626220703, 0.03721932983398438, 0.03730739212036133, 0.037256191253662106, 0.037236736297607424, 0.03723775863647461, 0.03721420669555664, 0.03726540756225586, 0.03721113586425781, 0.11815321350097656, 0.03713945770263672, 0.03711897659301758, 0.03713945770263672, 0.03745792007446289, 0.037566463470458986, 0.03751731109619141, 0.03712307357788086, 0.03717324829101563, 0.03714048004150391, 0.037168128967285156, 0.0371486701965332, 0.037125118255615236, 0.03711795043945312, 0.03713740921020508, 0.03717836761474609, 0.037152767181396484, 0.037141502380371096, 0.037106719970703125, 0.037166046142578124, 0.03711590576171875, 0.03713945770263672, 0.03721113586425781, 0.037157920837402346, 0.0371363525390625, 0.037179393768310545, 0.037174270629882815, 0.03712614440917969, 0.037174270629882815, 0.037238784790039066, 0.03717324829101563, 0.03719993591308594, 0.03716499328613281, 0.037220352172851565, 0.037174270629882815, 0.03718143844604492, 0.03718041610717773, 0.037169151306152344, 0.03722137451171875, 0.03722137451171875, 0.037169151306152344, 0.03719375991821289, 0.03713942337036133, 0.03722956848144531, 0.037152767181396484, 0.0371701774597168, 0.03719987106323242, 0.037177345275878904, 0.037253120422363284, 0.03722137451171875, 0.03724595260620117, 0.03724185562133789, 0.03730643081665039, 0.03723667144775391, 0.0371671028137207, 0.037288959503173826, 0.03722547149658203, 0.03727052688598633, 0.03725107192993164, 0.03723775863647461, 0.03719475173950195, 0.037195777893066405, 0.03723775863647461, 0.11818905639648437, 0.03713849639892578, 0.03718751907348633, 0.03709542465209961, 0.037106689453125, 0.037168128967285156, 0.037174270629882815, 0.037223423004150394, 0.03712307357788086, 0.03726438522338867, 0.03716403198242187, 0.03713228988647461, 0.037163009643554686, 0.0371599349975586, 0.03711385726928711, 0.03718963241577149, 0.037179393768310545, 0.03719475173950195, 0.037154815673828126, 0.037176319122314457, 0.0371486701965332, 0.0371599349975586, 0.0372305908203125, 0.03724800109863281, 0.0371486701965332, 0.03714048004150391, 0.03715379333496094, 0.03715071868896484, 0.03712307357788086, 0.037200897216796876, 0.037131263732910154, 0.0371671028137207, 0.037147647857666014, 0.03718963241577149, 0.037177345275878904, 0.03720601654052735, 0.037266433715820314, 0.03716403198242187, 0.03715686416625977, 0.03720399856567383, 0.03726742553710938, 0.03720499038696289, 0.037133312225341795, 0.037238784790039066, 0.03724492645263672, 0.03719270324707031, 0.03722137451171875, 0.03720806503295898, 0.037179393768310545, 0.03718348693847656, 0.03719475173950195, 0.03721830368041992, 0.037138431549072266, 0.03723161697387695, 0.03722547149658203, 0.03721830368041992, 0.03719987106323242, 0.03720806503295898, 0.03734425735473633, 0.03746406555175781, 0.03721113586425781, 0.037203968048095705, 0.03720908737182617, 0.11819622039794922, 0.037106689453125, 0.0371599349975586, 0.03711078262329102, 0.037100543975830076, 0.03711897659301758, 0.0371517448425293, 0.037138431549072266, 0.037119998931884765, 0.037157886505126955, 0.037141502380371096, 0.03712409591674805, 0.037147647857666014, 0.03717324829101563, 0.03719475173950195, 0.03712204742431641, 0.037275646209716795, 0.03725516891479492, 0.0371517448425293, 0.03726847839355469, 0.03715686416625977, 0.03715891265869141, 0.03712204742431641, 0.037182464599609374, 0.037190654754638675, 0.03718143844604492, 0.037174270629882815, 0.037174270629882815, 0.037220352172851565, 0.03717529678344727, 0.03715891265869141, 0.037176319122314457, 0.037185535430908204, 0.03713945770263672, 0.0371671028137207, 0.037182464599609374, 0.03722649765014648, 0.03745382308959961, 0.037362720489501955, 0.0372889289855957, 0.03728793716430664, 0.0372408332824707, 0.03718656158447266, 0.03719782257080078, 0.037190654754638675, 0.037217281341552735, 0.03721420669555664, 0.03723263931274414, 0.03719372940063476, 0.037200897216796876, 0.037182464599609374, 0.03722649765014648, 0.037302272796630856, 0.03719680023193359, 0.03719475173950195, 0.037176319122314457, 0.03721219253540039, 0.03722646331787109, 0.037217342376708984, 0.03729401779174805, 0.037233665466308595, 0.03721420669555664, 0.03722854232788086, 0.11825049591064453, 0.03718656158447266, 0.03717836761474609, 0.0371701774597168, 0.037149696350097655, 0.037166080474853515, 0.03714355087280274, 0.03713740921020508, 0.03715686416625977, 0.03712204742431641, 0.037131263732910154, 0.03714252853393555, 0.03712204742431641, 0.03713228988647461, 0.037144577026367184, 0.03714355087280274, 0.037125118255615236, 0.03723161697387695, 0.03713433456420898, 0.03716403198242187, 0.037114879608154294, 0.037147647857666014, 0.037166080474853515, 0.03722444915771484, 0.037174270629882815, 0.03712102508544922, 0.0371701774597168, 0.03715379333496094, 0.03717836761474609, 0.03722550582885742, 0.037159904479980466, 0.03719270324707031, 0.03716505432128906, 0.03718041610717773, 0.037154815673828126, 0.0371517448425293, 0.037154815673828126, 0.03716505432128906, 0.037169151306152344, 0.03727667236328125, 0.03719372940063476, 0.037190654754638675, 0.03716713714599609, 0.037212127685546874, 0.03725823974609375, 0.037379070281982424, 0.03721318435668945, 0.037176319122314457, 0.037174270629882815, 0.037335041046142575, 0.03719372940063476, 0.037282817840576174, 0.037341182708740234, 0.03735859298706055, 0.03718963241577149, 0.03724800109863281, 0.037198848724365234, 0.037253120422363284, 0.037198848724365234, 0.03721420669555664, 0.037179393768310545, 0.03720915222167969, 0.03721721649169922, 0.11788902282714844, 0.037106689453125, 0.03714771270751953, 0.037101505279541015, 0.03709337615966797, 0.03712614440917969, 0.03717836761474609, 0.037136383056640625, 0.037108734130859376, 0.03719168090820312, 0.03716505432128906, 0.03709132766723633, 0.037147647857666014, 0.03764326477050781, 0.037602302551269534, 0.03722854232788086, 0.03712921524047851, 0.03718041610717773, 0.03715379333496094, 0.03719270324707031, 0.03719680023193359, 0.037168128967285156, 0.03722956848144531, 0.037141502380371096, 0.03715686416625977, 0.03718860626220703, 0.03717836761474609, 0.037157886505126955, 0.037138431549072266, 0.03724595260620117, 0.03715071868896484, 0.037171199798583986, 0.03717324829101563, 0.03715891265869141, 0.037166080474853515, 0.03723468780517578, 0.03717324829101563, 0.03716505432128906, 0.037163009643554686, 0.037185535430908204, 0.037198848724365234, 0.03735859298706055, 0.03720297622680664, 0.03718550491333008, 0.03718041610717773, 0.037223423004150394, 0.037217281341552735, 0.03721625518798828, 0.037152767181396484, 0.037195777893066405, 0.03719168090820312, 0.03721318435668945, 0.037190654754638675, 0.037223423004150394, 0.037222400665283206, 0.03721830368041992, 0.03722547149658203, 0.03720191955566406, 0.03719987106323242, 0.03729715347290039, 0.0372305908203125, 0.037195777893066405, 0.037174270629882815]",tokens/s,26.064398936591285,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,google/recurrentgemma-7b,google/recurrentgemma-7b,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/google/recurrentgemma-7b/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66947048-4f3ad67a68bafb75790312ff;cc8ec58f-e063-46b7-a33f-30aa26a60e77) Repository Not Found for url: https://huggingface.co/google/recurrentgemma-7b/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: google/recurrentgemma-7b is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,mistralai/Mixtral-8x7B-v0.1,mistralai/Mixtral-8x7B-v0.1,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ self.load_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 250, in load_model_with_no_weights self.pretrained_model = self.automodel_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3710, in from_pretrained model = cls(config, *model_args, **model_kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/mixtral/modeling_mixtral.py"", line 1323, in __init__ self.model = MixtralModel(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/mixtral/modeling_mixtral.py"", line 1087, in __init__ [MixtralDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/mixtral/modeling_mixtral.py"", line 1087, in [MixtralDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/mixtral/modeling_mixtral.py"", line 877, in __init__ self.block_sparse_moe = MixtralSparseMoeBlock(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/mixtral/modeling_mixtral.py"", line 824, in __init__ self.experts = nn.ModuleList([MixtralBlockSparseTop2MLP(config) for _ in range(self.num_experts)]) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/mixtral/modeling_mixtral.py"", line 824, in self.experts = nn.ModuleList([MixtralBlockSparseTop2MLP(config) for _ in range(self.num_experts)]) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/mixtral/modeling_mixtral.py"", line 791, in __init__ self.w2 = nn.Linear(self.ffn_dim, self.hidden_dim, bias=False) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/linear.py"", line 98, in __init__ self.weight = Parameter(torch.empty((out_features, in_features), **factory_kwargs)) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_device.py"", line 78, in __torch_function__ return func(*args, **kwargs) torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 112.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,t,t,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/t/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669473ad-586e3d0b72f94d346722e174;8c42337d-2e5a-4918-8fd9-d51a00703850) Repository Not Found for url: https://huggingface.co/t/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: t is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-72B,Qwen/Qwen-72B,cuda,0,42,,,True,,,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",qwen,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-30b,facebook/opt-30b,cuda,0,42,,,True,,,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",opt,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-MoE-A2.7B,Qwen/Qwen1.5-MoE-A2.7B,cuda,0,42,,,True,,,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",qwen2_moe,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,x,x,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/x/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66932474-2697e70375a1ad4076fb93f4;9c144aa1-d1f8-48b5-a046-941019f99e71) Repository Not Found for url: https://huggingface.co/x/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: x is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,01-ai/Yi-34B,01-ai/Yi-34B,cuda,0,42,,,True,,,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",llama,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/xglm-4.5B,facebook/xglm-4.5B,cuda,0,42,,,True,,,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",xglm,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/xglm-7.5B,facebook/xglm-7.5B,cuda,0,42,,,True,,,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",xglm,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-14B,Qwen/Qwen1.5-14B,cuda,0,42,,,True,,,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",qwen2,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,meta-llama/Llama-2-70b-hf,meta-llama/Llama-2-70b-hf,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ self.load_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 250, in load_model_with_no_weights self.pretrained_model = self.automodel_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3710, in from_pretrained model = cls(config, *model_args, **model_kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 1101, in __init__ self.model = LlamaModel(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 884, in __init__ [LlamaDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 884, in [LlamaDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 679, in __init__ self.mlp = LlamaMLP(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 192, in __init__ self.gate_proj = nn.Linear(self.hidden_size, self.intermediate_size, bias=config.mlp_bias) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/linear.py"", line 98, in __init__ self.weight = Parameter(torch.empty((out_features, in_features), **factory_kwargs)) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_device.py"", line 78, in __torch_function__ return func(*args, **kwargs) torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 448.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Deci/DeciCoder-1b,Deci/DeciCoder-1b,cuda,0,42,,,True,,,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",llama,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,meta-llama/Meta-Llama-3-70B,meta-llama/Meta-Llama-3-70B,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ self.load_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 250, in load_model_with_no_weights self.pretrained_model = self.automodel_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3710, in from_pretrained model = cls(config, *model_args, **model_kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 1101, in __init__ self.model = LlamaModel(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 884, in __init__ [LlamaDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 884, in [LlamaDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 679, in __init__ self.mlp = LlamaMLP(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 193, in __init__ self.up_proj = nn.Linear(self.hidden_size, self.intermediate_size, bias=config.mlp_bias) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/linear.py"", line 98, in __init__ self.weight = Parameter(torch.empty((out_features, in_features), **factory_kwargs)) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_device.py"", line 78, in __torch_function__ return func(*args, **kwargs) torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 448.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,r,r,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/r/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669473ea-69adecae312f9779412c9183;8023eb00-40b5-46cb-860d-fea063a86411) Repository Not Found for url: https://huggingface.co/r/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: r is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,togethercomputer/RedPajama-INCITE-Base-7B-v0.1,togethercomputer/RedPajama-INCITE-Base-7B-v0.1,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.42.3,,0.31.0,,,,1.20.0,,,,0.11.1,,,,MB,1265.905664,15150.350336,0.0,14503.903232,14287.962624,s,10,0.7792710342407229,0.07792710342407227,0.00019341401932078007,0.07786975860595703,0.07800535659790038,0.07824472465515137,0.07843621910095214,"[0.07848409271240234, 0.07794735717773438, 0.07778915405273437, 0.07780316925048827, 0.07787369537353515, 0.07786582183837891, 0.07790793609619141, 0.0778314208984375, 0.07781622314453125, 0.07795216369628906]",tokens/s,3285.1214628993853,kWh,9.202622435505692e-07,5.042643107365787e-07,4.819310487658845e-06,6.243837041945993e-06,tokens/kWh,41000429.428282045,MB,1265.905664,15150.350336,0.0,14503.903232,14465.569792,s,10,21.039862304687503,2.10398623046875,0.002584409884179392,2.1033356933593748,2.1053836425781247,2.10841630859375,2.11084244140625,"[2.111448974609375, 2.103057861328125, 2.102752197265625, 2.101876220703125, 2.1034814453125, 2.10337646484375, 2.10247802734375, 2.103294921875, 2.103386474609375, 2.104709716796875]",tokens/s,29.943161741112792,kWh,2.483937996340969e-05,1.3612726292859853e-05,0.00011535531342974012,0.00015380741968600966,tokens/kWh,409603.1266151622,,s,629,21.706926082611083,0.03451021634755339,0.009259646844571118,0.03338649749755859,0.033502630615234376,0.03359496307373047,0.11130109680175782,"[0.033825790405273434, 0.03389750289916992, 0.03418524932861328, 0.03389228820800781, 0.033841152191162106, 0.0338197135925293, 0.03377452850341797, 0.03378483200073242, 0.033805313110351565, 0.03379097747802735, 0.0337448959350586, 0.03364457702636719, 0.033802207946777345, 0.0335175666809082, 0.03330047988891602, 0.03328716659545899, 0.033291263580322264, 0.03368038558959961, 0.03336908721923828, 0.03328716659545899, 0.033354782104492185, 0.033595359802246094, 0.03340390396118164, 0.03329740905761719, 0.033360897064208986, 0.03338649749755859, 0.03331686401367188, 0.033304576873779294, 0.03335580825805664, 0.033385440826416014, 0.033350654602050785, 0.03336703872680664, 0.033491008758544924, 0.0333639030456543, 0.03343360137939453, 0.03336294555664063, 0.03339980697631836, 0.03338137435913086, 0.03338956832885742, 0.03337830352783203, 0.0334202880859375, 0.033432575225830076, 0.03347868728637695, 0.03344892883300781, 0.03343667221069336, 0.03348787307739258, 0.03355955123901367, 0.03342233657836914, 0.0334202880859375, 0.033465343475341795, 0.033438720703125, 0.0334571533203125, 0.03343360137939453, 0.033445888519287106, 0.033495040893554685, 0.03340492630004883, 0.03348889541625977, 0.033448959350585936, 0.03347353744506836, 0.0334919662475586, 0.03349401473999023, 0.0335206413269043, 0.11143679809570313, 0.03323392105102539, 0.03331686401367188, 0.033261566162109374, 0.033279998779296875, 0.03326464080810547, 0.033291263580322264, 0.03328102493286133, 0.033331199645996096, 0.033307647705078124, 0.033277950286865234, 0.033339393615722655, 0.033260543823242186, 0.033307647705078124, 0.03332505416870117, 0.03325747299194336, 0.033314815521240236, 0.03334656143188477, 0.0333496322631836, 0.03331891250610351, 0.03331686401367188, 0.033301631927490236, 0.03331161499023438, 0.033342464447021485, 0.033306625366210936, 0.033342464447021485, 0.03338649749755859, 0.03331891250610351, 0.033410049438476565, 0.03341209411621094, 0.033337345123291014, 0.033337345123291014, 0.03341721725463867, 0.03341628646850586, 0.03340585708618164, 0.03342131042480469, 0.033347583770751955, 0.03334348678588867, 0.03345724868774414, 0.03340278244018555, 0.03339878463745117, 0.03337318420410156, 0.03341107177734375, 0.03345305633544922, 0.03342233657836914, 0.03347558212280274, 0.03345100784301758, 0.03340902328491211, 0.03347353744506836, 0.03345510482788086, 0.03341926574707031, 0.033432575225830076, 0.03343667221069336, 0.033462272644042966, 0.03348992156982422, 0.03350527954101563, 0.033546241760253906, 0.03355033493041992, 0.033492992401123044, 0.03349708938598633, 0.033492992401123044, 0.03348070526123047, 0.03351039886474609, 0.11127603149414063, 0.033259521484375, 0.03321548843383789, 0.033235969543457033, 0.0332861442565918, 0.03326976013183594, 0.03330156707763672, 0.033297344207763674, 0.033314815521240236, 0.03330355072021484, 0.033307647705078124, 0.03326668930053711, 0.03325235366821289, 0.03328409576416016, 0.033277950286865234, 0.0332861442565918, 0.03328921508789062, 0.033309696197509765, 0.03330355072021484, 0.033326080322265625, 0.033307647705078124, 0.033293312072753906, 0.03332505416870117, 0.033375232696533204, 0.03331891250610351, 0.03328921508789062, 0.03334348678588867, 0.03333324813842774, 0.03336294555664063, 0.03335987091064453, 0.0333496322631836, 0.033364990234375, 0.03339571380615235, 0.0333496322631836, 0.033350654602050785, 0.033375232696533204, 0.03345305633544922, 0.0333834228515625, 0.03341721725463867, 0.033416191101074216, 0.03348275375366211, 0.03342339324951172, 0.03343356704711914, 0.03339878463745117, 0.033414142608642575, 0.033438720703125, 0.03344793701171875, 0.03343564987182617, 0.0334202880859375, 0.033455169677734375, 0.03346220779418945, 0.03342438507080078, 0.03343360137939453, 0.03347763061523437, 0.0334837760925293, 0.03347251129150391, 0.033476608276367184, 0.03342953491210938, 0.03349091339111328, 0.03347251129150391, 0.033729534149169925, 0.033535999298095705, 0.03350527954101563, 0.11126271820068359, 0.033261566162109374, 0.03324006271362305, 0.03325235366821289, 0.033242111206054685, 0.03327590560913086, 0.03324518585205078, 0.033290241241455076, 0.033320960998535154, 0.03329228973388672, 0.0333383674621582, 0.033276927947998046, 0.03328716659545899, 0.03338854217529297, 0.033291263580322264, 0.033271808624267575, 0.033291263580322264, 0.0333383674621582, 0.03330867385864258, 0.033309696197509765, 0.033276927947998046, 0.03332198333740234, 0.03331379318237305, 0.033304576873779294, 0.033304576873779294, 0.0333199348449707, 0.033339393615722655, 0.033364990234375, 0.03334656143188477, 0.033350654602050785, 0.03334348678588867, 0.033377281188964845, 0.033339393615722655, 0.03334143829345703, 0.03342233657836914, 0.03336294555664063, 0.03344179153442383, 0.03339059066772461, 0.033363967895507815, 0.03338956832885742, 0.03338137435913086, 0.03338649749755859, 0.03339059066772461, 0.033377281188964845, 0.03337318420410156, 0.03340800094604492, 0.03339775848388672, 0.03339468765258789, 0.03346432113647461, 0.03340390396118164, 0.0334202880859375, 0.03349094390869141, 0.03342540740966797, 0.03343769454956055, 0.03343155288696289, 0.033462272644042966, 0.03348275375366211, 0.03347148895263672, 0.033468414306640625, 0.033462272644042966, 0.03344384002685547, 0.033454078674316406, 0.033484798431396484, 0.11132927703857422, 0.033293312072753906, 0.03322880172729492, 0.03324415969848633, 0.03323904037475586, 0.033230846405029296, 0.03326259231567383, 0.033306625366210936, 0.03329945755004883, 0.03324825668334961, 0.033307647705078124, 0.03325132751464844, 0.03332716751098633, 0.033313728332519534, 0.033290241241455076, 0.033306625366210936, 0.033345535278320314, 0.033320960998535154, 0.033309696197509765, 0.033314815521240236, 0.03329945755004883, 0.03331174468994141, 0.03341721725463867, 0.03328102493286133, 0.033326080322265625, 0.03335475158691406, 0.033328128814697267, 0.03337123107910156, 0.03341814422607422, 0.0333568000793457, 0.03336703872680664, 0.03340390396118164, 0.033320960998535154, 0.03337830352783203, 0.03337932968139649, 0.033380352020263675, 0.033355777740478515, 0.03340083312988281, 0.03339980697631836, 0.03338137435913086, 0.03356364822387695, 0.033438720703125, 0.03347148895263672, 0.03341926574707031, 0.03338854217529297, 0.03340288162231445, 0.03340288162231445, 0.0334571533203125, 0.03345510482788086, 0.03347967910766601, 0.033460289001464846, 0.033455039978027346, 0.03345817565917969, 0.033462272644042966, 0.033448959350585936, 0.03345100784301758, 0.03383705520629883, 0.03354828643798828, 0.03350431823730469, 0.03344070434570313, 0.03360768127441406, 0.03367833709716797, 0.033548351287841796, 0.11136300659179688, 0.033203201293945314, 0.03320627212524414, 0.03328726577758789, 0.033288097381591795, 0.033304576873779294, 0.03326259231567383, 0.03332198333740234, 0.03325337600708008, 0.033279998779296875, 0.03325439834594727, 0.033279998779296875, 0.03343360137939453, 0.033247230529785156, 0.03327897644042969, 0.03338751983642578, 0.03338751983642578, 0.03334041595458984, 0.033337345123291014, 0.03329228973388672, 0.0333568000793457, 0.033355838775634764, 0.03326867294311524, 0.033342464447021485, 0.03335689544677734, 0.03332700729370117, 0.03333529663085937, 0.03335987091064453, 0.03331174468994141, 0.03337932968139649, 0.033445888519287106, 0.03333324813842774, 0.0333834228515625, 0.033535999298095705, 0.033344512939453126, 0.033442817687988284, 0.03340288162231445, 0.03339468765258789, 0.03342335891723633, 0.03343769454956055, 0.03336908721923828, 0.03346636962890625, 0.03340492630004883, 0.03340492630004883, 0.033465343475341795, 0.03344793701171875, 0.033438720703125, 0.03346739196777344, 0.033416191101074216, 0.033416191101074216, 0.03350527954101563, 0.033478656768798826, 0.03343462371826172, 0.03349606323242187, 0.03344384002685547, 0.03359436798095703, 0.03351039886474609, 0.033552383422851564, 0.03345817565917969, 0.03348070526123047, 0.03354214477539062, 0.033465343475341795, 0.03346944046020508, 0.11141532897949219, 0.033246177673339844, 0.03325235366821289, 0.03324313735961914, 0.033225727081298825, 0.03321548843383789, 0.033242111206054685, 0.03329536056518555, 0.033328128814697267, 0.033291263580322264, 0.03330047988891602, 0.033326080322265625, 0.033258495330810545, 0.033317886352539065, 0.03330252838134766, 0.03323494338989258, 0.03328102493286133, 0.03329536056518555, 0.033317886352539065, 0.03331071853637695, 0.033329151153564454, 0.033320960998535154, 0.033271808624267575, 0.033339393615722655, 0.033328128814697267, 0.03337625503540039, 0.03333631896972656, 0.03337011337280273, 0.033355777740478515, 0.033391616821289063, 0.0333496322631836, 0.033342464447021485, 0.03337011337280273, 0.03337420654296875, 0.0333383674621582, 0.03353395080566406, 0.03334656143188477, 0.03339263916015625, 0.033519615173339845, 0.033438720703125, 0.033463294982910154, 0.03338137435913086, 0.0334571533203125, 0.03337932968139649, 0.03340902328491211, 0.033432575225830076, 0.033462272644042966, 0.03340492630004883, 0.033393665313720705, 0.033462272644042966, 0.033413120269775394, 0.03343564987182617, 0.03344486236572266, 0.033430526733398434, 0.03347763061523437, 0.033432575225830076, 0.03343667221069336, 0.03349913787841797, 0.03348070526123047, 0.033476608276367184, 0.033484798431396484, 0.03356364822387695, 0.033481727600097655, 0.11131084442138672, 0.03323187255859375, 0.03325337600708008, 0.03328307342529297, 0.0332677116394043, 0.03335475158691406, 0.033304576873779294, 0.033261566162109374, 0.03330867385864258, 0.03334143829345703, 0.033293312072753906, 0.0332492790222168, 0.03328204727172852, 0.03328204727172852, 0.03329945755004883, 0.03330252838134766, 0.033312767028808594, 0.033347583770751955, 0.033288192749023435, 0.033323009490966796, 0.033285118103027346, 0.03331071853637695, 0.03336703872680664, 0.033347583770751955, 0.033317886352539065, 0.033339393615722655, 0.03343769454956055, 0.033345535278320314, 0.033393665313720705, 0.0333568000793457, 0.033329151153564454, 0.03335273742675781, 0.03339977645874023, 0.0333568000793457, 0.03362713623046875, 0.033430526733398434, 0.03334860610961914, 0.033539070129394534, 0.033358848571777344, 0.03340800094604492, 0.03342950439453125, 0.03342950439453125, 0.033436737060546874, 0.033400768280029296, 0.0334202880859375, 0.033465343475341795, 0.03344793701171875, 0.03346124649047852, 0.03344793701171875, 0.03345510482788086, 0.03347353744506836, 0.033481727600097655, 0.033465343475341795, 0.03343667221069336, 0.03347763061523437, 0.0334837760925293, 0.03349094390869141, 0.03347353744506836, 0.033454078674316406, 0.033527809143066405, 0.03347353744506836, 0.03349606323242187, 0.03347763061523437, 0.11139276885986328, 0.03325132751464844, 0.03321651077270508, 0.0332492790222168, 0.03329129409790039, 0.03330249786376953, 0.033304576873779294, 0.033306625366210936, 0.0332564468383789, 0.03331584167480469, 0.03325235366821289, 0.033309696197509765, 0.03327897644042969, 0.0332861442565918, 0.033312767028808594, 0.03334041595458984, 0.03332505416870117, 0.033306625366210936, 0.03330051040649414, 0.03329942321777344, 0.033329151153564454, 0.033331199645996096, 0.033331199645996096, 0.033342464447021485, 0.03334656143188477, 0.03334348678588867, 0.03337932968139649, 0.03337625503540039, 0.033468414306640625, 0.033355777740478515, 0.03336908721923828, 0.033396736145019534, 0.03338240051269531, 0.03340185546875, 0.03335987091064453, 0.03337625503540039, 0.033414142608642575, 0.033347583770751955, 0.03342438507080078, 0.0334571533203125, 0.03340595245361328, 0.03340185546875, 0.033463294982910154, 0.03339680099487305, 0.033430465698242186, 0.03363840103149414, 0.033484832763671875, 0.03347350311279297, 0.03356367874145508, 0.03343561553955078, 0.03343667221069336, 0.03345305633544922, 0.03343360137939453, 0.03347558212280274, 0.03345305633544922, 0.03343360137939453, 0.03352371215820313, 0.0334571533203125, 0.03346739196777344, 0.03346739196777344, 0.03352169418334961, 0.03350115203857422, 0.03349094390869141, 0.11144499206542968, 0.033242111206054685, 0.033274879455566404, 0.033248321533203125, 0.03321644973754883, 0.03331891250610351, 0.033258495330810545, 0.03327078247070313, 0.03326566314697266, 0.033301502227783206, 0.03331686401367188, 0.03332710266113281, 0.033363967895507815, 0.03337625503540039, 0.033331199645996096, 0.03334143829345703, 0.03332505416870117, 0.033306720733642575, 0.03330550384521484, 0.033331199645996096, 0.03333222579956055, 0.03331071853637695, 0.0333568000793457, 0.03331686401367188, 0.03335168075561523, 0.033345535278320314, 0.03339878463745117, 0.03335987091064453, 0.033391616821289063, 0.033976318359375, 0.033452030181884765, 0.033358848571777344, 0.03344486236572266, 0.03356671905517578, 0.03348889541625977, 0.03339676666259766, 0.03339772796630859, 0.03338751983642578, 0.03339775848388672, 0.03341209411621094, 0.0335175666809082, 0.03340390396118164, 0.03343360137939453, 0.033446910858154294, 0.0334284782409668, 0.03350732803344727, 0.0335022087097168, 0.03343667221069336, 0.033476608276367184, 0.033416191101074216, 0.033465343475341795, 0.033476608276367184, 0.03347455978393555, 0.03346432113647461, 0.03365273666381836, 0.033493022918701175, 0.03350012969970703, 0.03345100784301758, 0.03349606323242187, 0.03348992156982422, 0.033495040893554685, 0.033501182556152344, 0.033544193267822264]",tokens/s,28.9769264245976,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-72B,Qwen/Qwen1.5-72B,cuda,0,42,,,True,,,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",qwen2,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,internlm/internlm-20b,internlm/internlm-20b,cuda,0,42,,,True,,,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",internlm,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-110B,Qwen/Qwen1.5-110B,cuda,0,42,,,True,,,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",qwen2,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Salesforce/codegen-16B-nl,Salesforce/codegen-16B-nl,cuda,0,42,,,True,,,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",codegen,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,m,m,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/m/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669472ea-2818c28145d4623108d141f4;ae60b401-f432-4183-a446-e9a82bb9bb78) Repository Not Found for url: https://huggingface.co/m/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: m is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,openai-community/gpt2-large,openai-community/gpt2-large,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.42.3,,0.31.0,,,,1.20.0,,,,0.11.1,,,,MB,1188.225024,2663.907328,0.0,2017.460224,1692.386816,s,10,0.17705046463012697,0.017705046463012696,0.0021909820654718083,0.016859600067138672,0.018333903694152828,0.021273447704315177,0.02362508291244507,"[0.02421299171447754, 0.016793216705322266, 0.01768067169189453, 0.017442815780639647, 0.01678656005859375, 0.016977664947509765, 0.016746271133422853, 0.016848960876464845, 0.016691072463989258, 0.0168702392578125]",tokens/s,14459.154373573949,kWh,2.0028371286706513e-07,1.097463286653968e-07,8.32753491061022e-07,1.1427835325934839e-06,tokens/kWh,224014428.54100478,MB,1188.225024,2663.907328,0.0,2017.460224,1714.454528,s,10,11.065347900390625,1.1065347900390625,0.014787988543622706,1.1065360107421875,1.1290044921875,1.1293021118164064,1.1295402075195313,"[1.1289383544921876, 1.1060321044921875, 1.107697998046875, 1.106547119140625, 1.11025, 1.10652490234375, 1.1042779541015626, 1.1295997314453126, 1.08260546875, 1.082874267578125]",tokens/s,56.934495478245196,kWh,1.2995178994149135e-05,7.120967364178229e-06,2.9846382163337927e-05,4.996252852166528e-05,tokens/kWh,1260944.9894570743,,s,629,11.198115859985348,0.017803045882329654,0.0020628569017095127,0.017375232696533204,0.018036530685424802,0.01813831672668457,0.034187591247558595,"[0.018366464614868162, 0.018017280578613282, 0.017915903091430666, 0.018131967544555663, 0.01803059196472168, 0.018127872467041017, 0.018137088775634767, 0.018182144165039063, 0.01800601577758789, 0.018116607666015624, 0.01842585563659668, 0.018520063400268554, 0.018143232345581056, 0.018013216018676757, 0.018067487716674803, 0.01816671943664551, 0.018088991165161133, 0.017932256698608397, 0.01821900749206543, 0.017932287216186525, 0.018103296279907227, 0.01803980827331543, 0.01795686340332031, 0.01761177635192871, 0.01779817581176758, 0.018892768859863282, 0.0186429443359375, 0.018353151321411132, 0.018342912673950194, 0.017952768325805665, 0.017923072814941408, 0.01785753631591797, 0.018102272033691406, 0.017884159088134767, 0.017975296020507812, 0.017895423889160156, 0.017916927337646483, 0.017863679885864257, 0.017968128204345703, 0.01785241508483887, 0.01780633544921875, 0.017925119400024413, 0.017901567459106444, 0.01856515121459961, 0.017959903717041016, 0.017903615951538086, 0.017953792572021485, 0.017377279281616212, 0.0172042236328125, 0.01724313545227051, 0.017296384811401368, 0.017188863754272463, 0.01721036720275879, 0.017802240371704102, 0.01720319938659668, 0.017177600860595704, 0.017183744430541992, 0.017184768676757813, 0.017364992141723632, 0.017192960739135742, 0.017201183319091796, 0.01724515151977539, 0.03421491241455078, 0.017152000427246093, 0.01721548843383789, 0.017087488174438475, 0.017122304916381836, 0.01724825668334961, 0.017090560913085938, 0.01721139144897461, 0.017188863754272463, 0.017356800079345702, 0.017266687393188478, 0.017274879455566407, 0.017311744689941407, 0.017136640548706054, 0.017238016128540038, 0.017315872192382814, 0.017185760498046876, 0.01725542449951172, 0.017275903701782228, 0.017252351760864256, 0.01719193649291992, 0.017280000686645508, 0.017148927688598634, 0.01719193649291992, 0.017209344863891602, 0.01720012855529785, 0.017207328796386718, 0.017177568435668946, 0.01741926383972168, 0.01800601577758789, 0.01795686340332031, 0.01789132881164551, 0.01780531120300293, 0.017886207580566405, 0.01789132881164551, 0.01784115219116211, 0.017860671997070313, 0.017871871948242187, 0.01777862358093262, 0.01804800033569336, 0.01781760025024414, 0.017789951324462892, 0.0177828483581543, 0.018057184219360353, 0.017928159713745118, 0.017522687911987304, 0.01740598487854004, 0.017688543319702148, 0.017909759521484374, 0.017772544860839845, 0.017796096801757814, 0.017802240371704102, 0.017785856246948242, 0.017819648742675782, 0.017959936141967774, 0.017787904739379884, 0.017876991271972655, 0.01778483200073242, 0.017839103698730468, 0.01783500862121582, 0.017827840805053712, 0.01785241508483887, 0.017916927337646483, 0.03589542388916016, 0.017824640274047853, 0.01716223907470703, 0.017118207931518553, 0.01741004753112793, 0.01788313674926758, 0.017724416732788087, 0.01717452812194824, 0.017537023544311522, 0.01827123260498047, 0.018093055725097656, 0.018112512588500978, 0.017878015518188475, 0.017895423889160156, 0.01785753631591797, 0.01759846305847168, 0.017375232696533204, 0.017335296630859375, 0.01716633605957031, 0.0176363525390625, 0.01804800033569336, 0.017886240005493163, 0.018132959365844727, 0.017574911117553712, 0.017262592315673828, 0.01718169593811035, 0.017326080322265625, 0.017135616302490234, 0.017250303268432618, 0.01720832061767578, 0.01714995193481445, 0.017111040115356444, 0.0172728328704834, 0.017328128814697266, 0.0172359676361084, 0.01719603157043457, 0.017324031829833983, 0.017970176696777345, 0.01785139274597168, 0.017896448135375977, 0.017909759521484374, 0.017895423889160156, 0.017944576263427735, 0.017161216735839844, 0.017183744430541992, 0.017283071517944337, 0.017897472381591797, 0.017953792572021485, 0.01775103950500488, 0.017846271514892577, 0.01787392044067383, 0.017979391098022462, 0.017689599990844726, 0.01717862319946289, 0.017337343215942384, 0.017125375747680666, 0.01719398307800293, 0.01717452812194824, 0.017120319366455077, 0.01777964782714844, 0.017830911636352538, 0.017961984634399415, 0.01806540870666504, 0.03556147384643555, 0.017970176696777345, 0.017954816818237306, 0.017993728637695314, 0.01805721664428711, 0.01721446418762207, 0.017260543823242186, 0.017252351760864256, 0.017150976181030272, 0.01722470474243164, 0.01722572708129883, 0.017229824066162108, 0.0172359676361084, 0.01718169593811035, 0.017878015518188475, 0.018083839416503905, 0.017983488082885742, 0.01757695960998535, 0.017160192489624023, 0.017353727340698243, 0.017130495071411133, 0.017167360305786132, 0.01717452812194824, 0.017056768417358398, 0.017097728729248047, 0.01722675132751465, 0.0172391357421875, 0.01715500831604004, 0.01719702339172363, 0.017689599990844726, 0.017869823455810546, 0.017901567459106444, 0.017833984375, 0.018118656158447266, 0.017929216384887696, 0.017921024322509766, 0.017903615951538086, 0.01784012794494629, 0.017773567199707033, 0.01770086479187012, 0.017134592056274413, 0.017316928863525392, 0.01771513557434082, 0.017328128814697266, 0.017899551391601563, 0.01783600044250488, 0.017902591705322265, 0.018084863662719726, 0.017869823455810546, 0.017979391098022462, 0.01842483139038086, 0.017984512329101563, 0.017912832260131836, 0.017934335708618163, 0.0174704647064209, 0.01723289680480957, 0.017334272384643554, 0.01719193649291992, 0.017131519317626954, 0.01726361656188965, 0.017309823989868165, 0.017126304626464844, 0.017164255142211916, 0.03418931198120117, 0.01720319938659668, 0.017156095504760743, 0.017154048919677735, 0.017175552368164062, 0.01720217514038086, 0.017171455383300782, 0.01782579231262207, 0.01725951957702637, 0.017543167114257813, 0.017292287826538084, 0.017318975448608397, 0.017193920135498048, 0.01724825668334961, 0.01722675132751465, 0.017316864013671874, 0.017164287567138673, 0.018070528030395508, 0.018107391357421874, 0.017702911376953127, 0.017097728729248047, 0.017108991622924806, 0.01721446418762207, 0.017937408447265626, 0.018291711807250977, 0.0180633602142334, 0.018945024490356444, 0.01804083251953125, 0.017918975830078124, 0.017901567459106444, 0.0179487361907959, 0.017918912887573243, 0.017954816818237306, 0.017935359954833984, 0.017813600540161133, 0.018131872177124024, 0.017957887649536132, 0.018160640716552736, 0.017885183334350584, 0.01784115219116211, 0.017942527770996093, 0.01790771293640137, 0.017906688690185548, 0.01801113510131836, 0.018100223541259765, 0.018075647354125975, 0.018113536834716795, 0.017941503524780272, 0.017967103958129883, 0.017916927337646483, 0.01785753631591797, 0.01796505546569824, 0.017135616302490234, 0.017137664794921875, 0.017271808624267578, 0.017146879196166993, 0.017192960739135742, 0.017138687133789063, 0.017122304916381836, 0.01718988800048828, 0.017116159439086915, 0.017120256423950195, 0.0171909122467041, 0.034375679016113284, 0.017184768676757813, 0.017290239334106446, 0.01721036720275879, 0.017104896545410156, 0.01721343994140625, 0.017143808364868163, 0.01765068817138672, 0.01821286392211914, 0.0178657283782959, 0.01723084831237793, 0.017156095504760743, 0.017184768676757813, 0.017245183944702147, 0.017262592315673828, 0.01746329689025879, 0.017898496627807618, 0.01801420783996582, 0.017917951583862304, 0.01786675262451172, 0.017895423889160156, 0.01800396728515625, 0.01799884796142578, 0.017939456939697264, 0.017810432434082032, 0.017943552017211914, 0.017833984375, 0.017897472381591797, 0.017961984634399415, 0.01796403121948242, 0.017827840805053712, 0.017829887390136717, 0.01725644874572754, 0.017317888259887695, 0.01716531181335449, 0.017184768676757813, 0.01717350387573242, 0.017150976181030272, 0.017175552368164062, 0.017132575988769532, 0.017149919509887694, 0.017354751586914064, 0.017097728729248047, 0.017514495849609374, 0.017938528060913086, 0.019543968200683593, 0.017985536575317384, 0.017362943649291994, 0.017188863754272463, 0.017373184204101562, 0.017217536926269532, 0.017112064361572265, 0.017129472732543945, 0.01744691276550293, 0.017878015518188475, 0.01785139274597168, 0.017978368759155275, 0.01803264045715332, 0.018139135360717772, 0.017769472122192383, 0.01716633605957031, 0.01721343994140625, 0.017147903442382813, 0.03418316650390625, 0.01717865562438965, 0.017129472732543945, 0.017149919509887694, 0.017257535934448242, 0.01718675231933594, 0.01716633605957031, 0.01729443168640137, 0.017795999526977538, 0.017975296020507812, 0.017889280319213868, 0.017822719573974608, 0.01782067108154297, 0.01799577522277832, 0.017950719833374023, 0.01781452751159668, 0.01788313674926758, 0.017836032867431642, 0.017761344909667968, 0.017838016510009765, 0.017786880493164063, 0.017932287216186525, 0.020976640701293944, 0.0184453125, 0.018126848220825196, 0.017943552017211914, 0.01785036849975586, 0.01778278350830078, 0.017847295761108398, 0.017772544860839845, 0.01802137565612793, 0.017914880752563478, 0.017415168762207032, 0.017344512939453126, 0.017116159439086915, 0.017155071258544922, 0.01719398307800293, 0.0172042236328125, 0.017106943130493164, 0.017168384552001953, 0.017148927688598634, 0.017080320358276366, 0.017135616302490234, 0.017368064880371094, 0.017139711380004884, 0.017309696197509765, 0.017160192489624023, 0.01716531181335449, 0.01717862319946289, 0.017521663665771483, 0.017468416213989257, 0.017080320358276366, 0.017275903701782228, 0.017315839767456053, 0.017176576614379883, 0.017123327255249024, 0.017118207931518553, 0.017269760131835937, 0.017124351501464845, 0.01723494338989258, 0.017189983367919923, 0.01716316795349121, 0.017285120010375975, 0.03427328109741211, 0.017876991271972655, 0.017861631393432616, 0.017876991271972655, 0.017918975830078124, 0.018053119659423827, 0.017870847702026366, 0.01789030456542969, 0.017838079452514647, 0.01802649688720703, 0.018035711288452147, 0.017909759521484374, 0.017915903091430666, 0.017991680145263672, 0.017870847702026366, 0.017903615951538086, 0.01825484848022461, 0.017941503524780272, 0.017913856506347657, 0.017953792572021485, 0.018085887908935547, 0.017921119689941405, 0.017885087966918945, 0.017895423889160156, 0.01783193588256836, 0.018025472640991212, 0.01798860740661621, 0.0178187198638916, 0.01796188735961914, 0.017862655639648437, 0.01797427177429199, 0.017991680145263672, 0.01798041534423828, 0.017854463577270507, 0.018025472640991212, 0.01786675262451172, 0.017920000076293945, 0.01787494468688965, 0.017941503524780272, 0.017908735275268553, 0.017871871948242187, 0.017936384201049805, 0.017960960388183594, 0.017999872207641602, 0.017966079711914062, 0.017966079711914062, 0.01788211250305176, 0.017985536575317384, 0.017984512329101563, 0.017979391098022462, 0.017887231826782226, 0.017952831268310546, 0.017980384826660156, 0.01789027214050293, 0.017975296020507812, 0.017999872207641602, 0.017911808013916015, 0.017862655639648437, 0.017905664443969727, 0.01780633544921875, 0.017971200942993162, 0.0180163516998291, 0.017977247238159178, 0.034027584075927736, 0.017211328506469725, 0.0171561279296875, 0.01711017608642578, 0.017131328582763672, 0.017123327255249024, 0.017161279678344726, 0.017176511764526368, 0.017301504135131835, 0.017161216735839844, 0.017119232177734374, 0.017177600860595704, 0.017135616302490234, 0.017091583251953125, 0.01721855926513672, 0.01720729637145996, 0.01728102493286133, 0.01721139144897461, 0.017088575363159178, 0.017180608749389648, 0.017140735626220704, 0.017148927688598634, 0.017080320358276366, 0.017114112854003907, 0.017288192749023438, 0.017184768676757813, 0.017245248794555666, 0.017180608749389648, 0.01740595245361328, 0.0172410888671875, 0.01717350387573242, 0.017164287567138673, 0.017133567810058595, 0.01702092742919922, 0.01720627212524414, 0.017076223373413087, 0.017088512420654296, 0.0172410888671875, 0.01715814399719238, 0.0170700798034668, 0.017350656509399414, 0.017269760131835937, 0.0171909122467041, 0.017177600860595704, 0.017184768676757813, 0.017238016128540038, 0.017111040115356444, 0.017291328430175782, 0.017091520309448244, 0.017117183685302736, 0.01737215995788574, 0.017532928466796875, 0.017153024673461914, 0.017157119750976564, 0.017153024673461914, 0.01719398307800293, 0.017135616302490234, 0.017217536926269532, 0.017116159439086915, 0.017142784118652343, 0.01714080047607422, 0.01711916732788086, 0.017250303268432618, 0.03427635192871094, 0.01715814399719238, 0.017134592056274413, 0.017153024673461914, 0.017150976181030272, 0.017320959091186524, 0.01725951957702637, 0.017076223373413087, 0.017144832611083984, 0.01716633605957031, 0.017369087219238282, 0.017164287567138673, 0.01717350387573242, 0.01722265625, 0.01718681526184082, 0.017294336318969726, 0.017103872299194335, 0.01721241569519043, 0.017169408798217774, 0.017115135192871094, 0.017172479629516603, 0.017351680755615235, 0.01720012855529785, 0.01719398307800293, 0.01719910430908203, 0.017092607498168946, 0.01721651268005371, 0.01713974380493164, 0.017209312438964845, 0.017094655990600584, 0.01715814399719238, 0.017180671691894533, 0.017216543197631835, 0.017181663513183593, 0.01722060775756836, 0.017122304916381836, 0.017274879455566407, 0.017253376007080077, 0.017126399993896483, 0.017260543823242186, 0.01714995193481445, 0.01715814399719238, 0.0171909122467041, 0.017134592056274413, 0.017147903442382813, 0.0172677116394043, 0.017336320877075196, 0.017312768936157227, 0.017146944046020508, 0.01712531280517578, 0.017062911987304686, 0.017092607498168946, 0.017075199127197266, 0.017143808364868163, 0.017154048919677735, 0.017121280670166016, 0.01731283187866211, 0.017249311447143555, 0.017079200744628906, 0.017146879196166993, 0.017186880111694336, 0.017265600204467775, 0.01720832061767578]",tokens/s,56.17016361186522,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,huggyllama/llama-65b,huggyllama/llama-65b,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ self.load_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 250, in load_model_with_no_weights self.pretrained_model = self.automodel_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3710, in from_pretrained model = cls(config, *model_args, **model_kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 1101, in __init__ self.model = LlamaModel(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 884, in __init__ [LlamaDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 884, in [LlamaDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 677, in __init__ self.self_attn = LLAMA_ATTENTION_CLASSES[config._attn_implementation](config=config, layer_idx=layer_idx) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 264, in __init__ self.v_proj = nn.Linear(self.hidden_size, self.num_key_value_heads * self.head_dim, bias=config.attention_bias) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/linear.py"", line 98, in __init__ self.weight = Parameter(torch.empty((out_features, in_features), **factory_kwargs)) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_device.py"", line 78, in __torch_function__ return func(*args, **kwargs) torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 128.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,l,l,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/l/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6694745d-06c2c7f1543af17b66cd03f0;62749f78-948a-495f-b1a6-96a29d5cd358) Repository Not Found for url: https://huggingface.co/l/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: l is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,B,B,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/B/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6693250e-750907b31259ba2304b28a1f;6927d3ba-4b7a-461b-821d-7e1ac1052620) Repository Not Found for url: https://huggingface.co/B/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: B is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,huggyllama/llama-30b,huggyllama/llama-30b,cuda,0,42,,,True,,,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",llama,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,meta-llama/Llama-2-13b-hf,meta-llama/Llama-2-13b-hf,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ self.load_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 250, in load_model_with_no_weights self.pretrained_model = self.automodel_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3710, in from_pretrained model = cls(config, *model_args, **model_kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 1101, in __init__ self.model = LlamaModel(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 884, in __init__ [LlamaDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 884, in [LlamaDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 679, in __init__ self.mlp = LlamaMLP(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 192, in __init__ self.gate_proj = nn.Linear(self.hidden_size, self.intermediate_size, bias=config.mlp_bias) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/linear.py"", line 98, in __init__ self.weight = Parameter(torch.empty((out_features, in_features), **factory_kwargs)) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_device.py"", line 78, in __torch_function__ return func(*args, **kwargs) torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 136.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-7B,,cuda,0,42,,,,,True,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,217063f5c507ed7cc255df7e1f64c4333a0b4dfe,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,,MB,1281.077248,17102.798848,0.0,16456.351744,16359.853056,s,10,0.9702524795532226,0.09702524795532226,0.00025320433899165477,0.09696044540405274,0.09724756317138672,0.09748131790161133,0.09766832168579101,"[0.09771507263183593, 0.09698252868652343, 0.09697071838378907, 0.09687910461425782, 0.096891357421875, 0.0969501724243164, 0.09686860656738282, 0.09702301025390625, 0.09677629089355469, 0.09719561767578125]",tokens/s,2638.488490314209,kWh,1.1476153758994457e-06,6.288410415630634e-07,5.71933035755311e-06,7.495786775015618e-06,tokens/kWh,34152518.966158375,MB,1281.077248,17102.798848,0.0,16456.351744,16416.242176,s,10,23.38539697265625,2.3385396972656247,0.0029707981617589597,2.3381635742187497,2.3395967529296877,2.3432687133789063,2.346206281738281,"[2.346940673828125, 2.336866943359375, 2.335611083984375, 2.336256103515625, 2.338372314453125, 2.338546875, 2.337695068359375, 2.33810107421875, 2.33878076171875, 2.33822607421875]",tokens/s,26.93988905711704,kWh,2.7537362347593073e-05,1.5091464710438814e-05,0.00011439315462125123,0.00015702198167928308,tokens/kWh,401217.7105793844,,s,629,24.219824024200456,0.03850528461717081,0.011591252179652836,0.03709337615966797,0.037211750030517576,0.03740590057373047,0.1345780224609375,"[0.037854110717773434, 0.03778051376342773, 0.0378449592590332, 0.03768739318847656, 0.037798816680908204, 0.0376995849609375, 0.037776382446289065, 0.03777433776855469, 0.037754878997802735, 0.037515262603759765, 0.0376258544921875, 0.037651454925537106, 0.03721113586425781, 0.037026817321777344, 0.03700838470458984, 0.037026817321777344, 0.03706163024902344, 0.037029888153076174, 0.037010433197021485, 0.03705036926269531, 0.03703398513793945, 0.03711795043945312, 0.03704115295410156, 0.03706777572631836, 0.0370513916015625, 0.037043201446533204, 0.037131263732910154, 0.0370882568359375, 0.03702783966064453, 0.03705344009399414, 0.03727360153198242, 0.03744460678100586, 0.037152767181396484, 0.03709747314453125, 0.0370882568359375, 0.037103614807128905, 0.037087230682373046, 0.037157886505126955, 0.03709542465209961, 0.03708006286621094, 0.03709337615966797, 0.03710771179199219, 0.03706880187988281, 0.0371599349975586, 0.03728486251831055, 0.037116928100585936, 0.03711795043945312, 0.03723571014404297, 0.03726335906982422, 0.03715584182739258, 0.03713536071777344, 0.03712819290161133, 0.03712409591674805, 0.037099582672119144, 0.037134273529052735, 0.03709337615966797, 0.03708108901977539, 0.037144577026367184, 0.037149696350097655, 0.037130241394042966, 0.037116928100585936, 0.037166080474853515, 0.13465087890625, 0.03696230316162109, 0.03696640014648438, 0.036947967529296875, 0.03693670272827149, 0.03698995208740234, 0.03701145553588867, 0.03700735855102539, 0.03697868728637695, 0.036959232330322264, 0.03699814224243164, 0.037015552520751956, 0.037005313873291014, 0.036982784271240236, 0.03711897659301758, 0.037512191772460936, 0.03710771179199219, 0.03706675338745117, 0.03701145553588867, 0.03703302383422852, 0.037025726318359375, 0.03700838470458984, 0.03701760101318359, 0.03699814224243164, 0.03700534439086914, 0.03702985763549805, 0.03704217529296875, 0.03707392120361328, 0.03705036926269531, 0.03700735855102539, 0.03700838470458984, 0.03727155303955078, 0.037498878479003905, 0.03712716674804688, 0.03704422378540039, 0.03708927917480469, 0.037087230682373046, 0.037059585571289064, 0.037032958984375, 0.03714355087280274, 0.037294078826904296, 0.0370964469909668, 0.037177345275878904, 0.03710774230957031, 0.037061599731445315, 0.0370709114074707, 0.03721516799926758, 0.037103614807128905, 0.037108734130859376, 0.037119998931884765, 0.037116928100585936, 0.037136383056640625, 0.03711180877685547, 0.037082111358642575, 0.03712102508544922, 0.03713433456420898, 0.03710771179199219, 0.037163009643554686, 0.03712409591674805, 0.037133312225341795, 0.03714976119995117, 0.03714041519165039, 0.03712211227416992, 0.13467027282714844, 0.03700735855102539, 0.036983806610107424, 0.03702374267578125, 0.03700838470458984, 0.036944896697998046, 0.03693772888183594, 0.03698995208740234, 0.037032958984375, 0.03700940704345703, 0.03698688125610351, 0.037026817321777344, 0.03703398513793945, 0.03702169418334961, 0.037002239227294925, 0.03697459030151367, 0.03705241775512695, 0.03711283111572266, 0.037026817321777344, 0.037032958984375, 0.03701760101318359, 0.03724800109863281, 0.0370247688293457, 0.037116928100585936, 0.036994049072265625, 0.03702272033691406, 0.03704012680053711, 0.03703705596923828, 0.037028865814208986, 0.037048320770263675, 0.03709235382080078, 0.03709132766723633, 0.03703398513793945, 0.03706265640258789, 0.03708108901977539, 0.03707494354248047, 0.03706367874145508, 0.037032958984375, 0.03711180877685547, 0.03706572723388672, 0.03703398513793945, 0.03708313751220703, 0.03709542465209961, 0.03708313751220703, 0.03706265640258789, 0.03706367874145508, 0.03712409591674805, 0.037114879608154294, 0.03711897659301758, 0.03717324829101563, 0.0370964469909668, 0.037085182189941404, 0.03709036636352539, 0.0371578254699707, 0.03714048004150391, 0.03710771179199219, 0.03709030532836914, 0.03710367965698242, 0.03713222503662109, 0.037119998931884765, 0.03712307357788086, 0.03711180877685547, 0.03712307357788086, 0.13458432006835938, 0.036961280822753906, 0.03693670272827149, 0.037028865814208986, 0.03695820617675781, 0.036980735778808595, 0.03699302291870117, 0.03704422378540039, 0.03700121688842774, 0.03701356887817383, 0.037033920288085935, 0.03703603363037109, 0.03701862335205078, 0.03699715042114258, 0.037002273559570316, 0.03705440139770508, 0.037032958984375, 0.037028865814208986, 0.037048320770263675, 0.03703500747680664, 0.037013504028320314, 0.037031936645507815, 0.03700838470458984, 0.037031936645507815, 0.03704524612426758, 0.037029888153076174, 0.037010433197021485, 0.03706572723388672, 0.03712819290161133, 0.03706982421875, 0.03708931350708008, 0.037086177825927734, 0.03706572723388672, 0.03707494354248047, 0.037131263732910154, 0.03709132766723633, 0.03708006286621094, 0.037106689453125, 0.037186622619628906, 0.03711174392700195, 0.0370882568359375, 0.03712614440917969, 0.03708927917480469, 0.03708108901977539, 0.037100543975830076, 0.03712307357788086, 0.03710771179199219, 0.03709439849853516, 0.0370964469909668, 0.037104671478271484, 0.03710563278198242, 0.037114879608154294, 0.037125118255615236, 0.03713945770263672, 0.0371517448425293, 0.03712921524047851, 0.03716505432128906, 0.0371486701965332, 0.03713433456420898, 0.03718963241577149, 0.03713129425048828, 0.037109729766845706, 0.03719782257080078, 0.1347071990966797, 0.03697155380249023, 0.03693257522583008, 0.037010433197021485, 0.03698688125610351, 0.0370063362121582, 0.03709542465209961, 0.037000190734863284, 0.03700940704345703, 0.03705855941772461, 0.036977664947509765, 0.03699612808227539, 0.03700527954101562, 0.03698688125610351, 0.037026817321777344, 0.036988929748535154, 0.037007423400878904, 0.0370656623840332, 0.0370247688293457, 0.03705241775512695, 0.03698175811767578, 0.03700428771972656, 0.037016609191894534, 0.03706876754760742, 0.037026817321777344, 0.03710976028442383, 0.0370513916015625, 0.03708927917480469, 0.037032993316650394, 0.03708208084106445, 0.03703705596923828, 0.03713740921020508, 0.037028865814208986, 0.037061695098876954, 0.03702880096435547, 0.03703705596923828, 0.037048320770263675, 0.03706880187988281, 0.0372490234375, 0.0377968635559082, 0.037154815673828126, 0.0372674560546875, 0.03717324829101563, 0.03723571014404297, 0.0370964469909668, 0.037125118255615236, 0.03711078262329102, 0.037171199798583986, 0.03709439849853516, 0.03712819290161133, 0.037119998931884765, 0.037169151306152344, 0.037177345275878904, 0.03715584182739258, 0.037130241394042966, 0.037272575378417966, 0.03749068832397461, 0.0374015998840332, 0.03717836761474609, 0.03717228698730469, 0.03718956756591797, 0.03724800109863281, 0.03721420669555664, 0.13471539306640626, 0.036983806610107424, 0.037015552520751956, 0.036972545623779295, 0.03695820617675781, 0.03699097442626953, 0.03697049713134765, 0.03702783966064453, 0.036997119903564454, 0.037000190734863284, 0.036983806610107424, 0.03704115295410156, 0.03701561737060547, 0.037039039611816406, 0.037032958984375, 0.03703705596923828, 0.037138431549072266, 0.03708108901977539, 0.03706675338745117, 0.037054462432861326, 0.037028865814208986, 0.037215232849121094, 0.03749273681640625, 0.03716198348999023, 0.03709952163696289, 0.03718963241577149, 0.037116928100585936, 0.037157886505126955, 0.03707904052734375, 0.03708006286621094, 0.037084159851074217, 0.03710771179199219, 0.03712409591674805, 0.037141502380371096, 0.037157886505126955, 0.037101566314697264, 0.03711385726928711, 0.037182464599609374, 0.03722547149658203, 0.03729100799560547, 0.037343231201171875, 0.03715584182739258, 0.03713536071777344, 0.037119998931884765, 0.0370882568359375, 0.037101566314697264, 0.03717222213745117, 0.03718348693847656, 0.03716819381713867, 0.037178302764892576, 0.03712716674804688, 0.03711590576171875, 0.03711385726928711, 0.037146625518798826, 0.037163009643554686, 0.03712102508544922, 0.037125118255615236, 0.03714767837524414, 0.03712815856933594, 0.037100543975830076, 0.03711590576171875, 0.03712102508544922, 0.037116928100585936, 0.13456182861328125, 0.03697046279907226, 0.036975616455078124, 0.037013504028320314, 0.03701657485961914, 0.03698995208740234, 0.03698483276367188, 0.03703398513793945, 0.03703603363037109, 0.03699507141113281, 0.03696640014648438, 0.037026817321777344, 0.037029888153076174, 0.036969470977783206, 0.03699100875854492, 0.03699299240112305, 0.036999168395996096, 0.03706982421875, 0.037005313873291014, 0.03706572723388672, 0.0370247688293457, 0.0371486701965332, 0.03702579116821289, 0.03722547149658203, 0.03751321411132812, 0.037084159851074217, 0.03704217529296875, 0.03707187271118164, 0.03705855941772461, 0.03707904052734375, 0.037101566314697264, 0.03706780624389648, 0.03703907012939453, 0.03707699203491211, 0.037059585571289064, 0.03707494354248047, 0.03710771179199219, 0.03717222213745117, 0.037084159851074217, 0.037106689453125, 0.03711180877685547, 0.037147647857666014, 0.03708313751220703, 0.03714252853393555, 0.037130241394042966, 0.03716505432128906, 0.03712102508544922, 0.03716815948486328, 0.03715887832641602, 0.0371671028137207, 0.03711795043945312, 0.03719270324707031, 0.03713228988647461, 0.03715071868896484, 0.037133312225341795, 0.03719680023193359, 0.03715071868896484, 0.03715689468383789, 0.03718038558959961, 0.037294078826904296, 0.037182464599609374, 0.037223457336425785, 0.037200862884521485, 0.13478501892089845, 0.03696332931518555, 0.036969470977783206, 0.036947967529296875, 0.037378047943115236, 0.03700940704345703, 0.03700940704345703, 0.03699609756469727, 0.03700428771972656, 0.03703091049194336, 0.03701964950561523, 0.036983806610107424, 0.036985855102539066, 0.03702272033691406, 0.036982784271240236, 0.03700428771972656, 0.0370780143737793, 0.03708313751220703, 0.03701356887817383, 0.03705344009399414, 0.03708102416992187, 0.03709235382080078, 0.03712716674804688, 0.03706367874145508, 0.03708620834350586, 0.03706675338745117, 0.03703091049194336, 0.03710675048828125, 0.03711174392700195, 0.03706777572631836, 0.03705344009399414, 0.03709542465209961, 0.03713433456420898, 0.03713945770263672, 0.03719372940063476, 0.03722137451171875, 0.03713945770263672, 0.03709132766723633, 0.03712819290161133, 0.03712307357788086, 0.03711180877685547, 0.03721932983398438, 0.03712204742431641, 0.037141502380371096, 0.03725721740722656, 0.037185535430908204, 0.0372305908203125, 0.03719475173950195, 0.037131263732910154, 0.037144577026367184, 0.0370964469909668, 0.037111839294433596, 0.03718038558959961, 0.0371610221862793, 0.03713222503662109, 0.03715891265869141, 0.03721011352539062, 0.03718963241577149, 0.037116928100585936, 0.03715584182739258, 0.03715379333496094, 0.03713433456420898, 0.037174270629882815, 0.13449215698242187, 0.03707494354248047, 0.037082111358642575, 0.036980735778808595, 0.03698995208740234, 0.03716505432128906, 0.03707292938232422, 0.03710460662841797, 0.03713433456420898, 0.03703910446166992, 0.03699507141113281, 0.0370063362121582, 0.03706982421875, 0.03717222213745117, 0.0370513916015625, 0.03697663879394531, 0.03704422378540039, 0.03705753707885742, 0.03709337615966797, 0.0371486701965332, 0.037059585571289064, 0.0370513916015625, 0.03706163024902344, 0.037100543975830076, 0.03709747314453125, 0.037054462432861326, 0.03700735855102539, 0.037103614807128905, 0.037073951721191406, 0.03710051345825195, 0.03706880187988281, 0.037108734130859376, 0.03710566329956055, 0.03712102508544922, 0.03703910446166992, 0.037157886505126955, 0.03708006286621094, 0.03705548858642578, 0.03703500747680664, 0.03709952163696289, 0.03707699203491211, 0.03709030532836914, 0.0370964469909668, 0.037108734130859376, 0.03708313751220703, 0.03744870376586914, 0.03763407897949219, 0.03717532730102539, 0.03707795333862305, 0.0371517448425293, 0.037100543975830076, 0.03718656158447266, 0.03730636978149414, 0.03715071868896484, 0.03710771179199219, 0.037125118255615236, 0.03713945770263672, 0.03716403198242187, 0.037119998931884765, 0.03720502471923828, 0.03716195297241211, 0.037160961151123044, 0.03721113586425781, 0.13518643188476562, 0.03713740921020508, 0.0370063362121582, 0.03700944137573242, 0.03694998550415039, 0.03706163024902344, 0.03701657485961914, 0.036997119903564454, 0.03697356796264648, 0.0370247688293457, 0.03704729461669922, 0.03702272033691406, 0.03705344009399414, 0.03707596969604492, 0.03704012680053711, 0.03701964950561523, 0.03702272033691406, 0.03702169418334961, 0.037000190734863284, 0.03702169418334961, 0.03701964950561523, 0.03707699203491211, 0.03705753707885742, 0.03704934310913086, 0.0370821762084961, 0.03707795333862305, 0.03707392120361328, 0.03703500747680664, 0.03727769470214844, 0.03741593551635742, 0.03714355087280274, 0.037195777893066405, 0.03704422378540039, 0.03713945770263672, 0.0371486701965332, 0.0370780143737793, 0.03707494354248047, 0.03706675338745117, 0.03709756851196289, 0.03709328079223633, 0.037082111358642575, 0.03720294570922852, 0.03712819290161133, 0.0371671028137207, 0.03721529769897461, 0.037120960235595704, 0.03711795043945312, 0.037185535430908204, 0.037292030334472655, 0.03715071868896484, 0.037108734130859376, 0.03708927917480469, 0.03709952163696289, 0.03716198348999023, 0.03712614440917969, 0.037166080474853515, 0.0371599349975586, 0.037408767700195314, 0.03712204742431641, 0.03715891265869141, 0.037146625518798826, 0.03711590576171875, 0.03715379333496094]",tokens/s,25.970461196229316,,,main,False,False,,,,,,,,,,,,,,,,,,,,,,, bfloat16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,M,M,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/M/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669322a2-2c0256da31407d2c44b35095;79a623c7-5e7f-4b80-9de0-5510b71b1df3) Repository Not Found for url: https://huggingface.co/M/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: M is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,stabilityai/stablelm-base-alpha-3b,stabilityai/stablelm-base-alpha-3b,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.42.3,,0.31.0,,,,1.20.0,,,,0.11.1,,,,MB,1277.231104,8529.641472,0.0,7883.194368,7764.230656,s,10,0.3879643821716308,0.03879643821716308,0.00015240225015702527,0.03874294471740723,0.03893232421875,0.039075905227661134,0.03919077003479004,"[0.03921948623657227, 0.038780990600585936, 0.038749855041503904, 0.03874272155761719, 0.038675552368164064, 0.03871920013427734, 0.03869424057006836, 0.03874316787719727, 0.03873875045776367, 0.03890041732788086]",tokens/s,6598.543880936695,kWh,4.5770069321354595e-07,2.507342432158816e-07,2.3605681690192826e-06,3.06900310544871e-06,tokens/kWh,83414708.68683627,MB,1277.558784,8529.641472,0.0,7883.194368,7840.651264,s,10,11.064323608398437,1.1064323608398436,0.0035680246495570192,1.1051912231445313,1.1068262939453126,1.111966796875,1.11607919921875,"[1.1171072998046876, 1.1052010498046876, 1.1050965576171874, 1.1047894287109374, 1.105181396484375, 1.10504931640625, 1.1056839599609376, 1.1054593505859376, 1.10510888671875, 1.1056463623046875]",tokens/s,56.939766252118204,kWh,1.3054254912399315e-05,7.15343414828663e-06,5.988542607218363e-05,8.009311513286957e-05,tokens/kWh,786584.4635395547,,s,629,11.395188726425157,0.018116357275715693,0.004620663668178502,0.017546239852905272,0.01760870361328125,0.017806949996948245,0.056395695037841795,"[0.01801625633239746, 0.017967103958129883, 0.01800601577758789, 0.01800499153137207, 0.01798246383666992, 0.01800601577758789, 0.017966079711914062, 0.018001920700073244, 0.017864704132080078, 0.017941503524780272, 0.018035711288452147, 0.018020351409912108, 0.018115583419799804, 0.01801625633239746, 0.017984512329101563, 0.01797427177429199, 0.017942527770996093, 0.01800396728515625, 0.018036735534667968, 0.018019327163696287, 0.018027519226074217, 0.017656831741333007, 0.01784012794494629, 0.01763430404663086, 0.017757183074951173, 0.017888256072998047, 0.017666048049926757, 0.017538047790527343, 0.01754521560668945, 0.017556480407714844, 0.017505279541015627, 0.017517568588256836, 0.017531904220581054, 0.017535999298095704, 0.017527807235717775, 0.017554431915283202, 0.017537023544311522, 0.017525760650634766, 0.017562623977661132, 0.017547264099121093, 0.017538047790527343, 0.017526784896850587, 0.017566719055175782, 0.01755340766906738, 0.017567743301391603, 0.017570816040039062, 0.017566719055175782, 0.017565696716308594, 0.017547264099121093, 0.017550336837768556, 0.01759436798095703, 0.01761177635192871, 0.017567743301391603, 0.017571840286254883, 0.01757900810241699, 0.017551359176635743, 0.01759846305847168, 0.017574911117553712, 0.01760051155090332, 0.01756159973144531, 0.01759436798095703, 0.01760870361328125, 0.05664767837524414, 0.017522687911987304, 0.017472511291503907, 0.01747865676879883, 0.01746329689025879, 0.017512447357177736, 0.017641471862792968, 0.017497087478637697, 0.017477632522583008, 0.017511423110961915, 0.017500160217285156, 0.01747865676879883, 0.01748684883117676, 0.017505279541015627, 0.017484800338745117, 0.017511423110961915, 0.017498111724853514, 0.01765273666381836, 0.017526784896850587, 0.017504255294799806, 0.017524736404418945, 0.017513471603393553, 0.017513471603393553, 0.017528831481933595, 0.017492992401123047, 0.01756876754760742, 0.017535999298095704, 0.017532928466796875, 0.017538047790527343, 0.017522687911987304, 0.017528831481933595, 0.017574911117553712, 0.017504255294799806, 0.01754521560668945, 0.017534975051879884, 0.017539072036743163, 0.017551359176635743, 0.017520639419555666, 0.017533952713012696, 0.017565696716308594, 0.017537023544311522, 0.017529855728149413, 0.01755340766906738, 0.017539072036743163, 0.017543167114257813, 0.017538047790527343, 0.017543167114257813, 0.017562623977661132, 0.017543167114257813, 0.017555456161499023, 0.017550336837768556, 0.017555456161499023, 0.01758515167236328, 0.017563648223876953, 0.017549312591552735, 0.01756876754760742, 0.017573888778686524, 0.017581056594848633, 0.01757900810241699, 0.01760051155090332, 0.017622016906738282, 0.01759436798095703, 0.01760767936706543, 0.05645312118530273, 0.017487871170043946, 0.017474559783935546, 0.017458175659179686, 0.017489919662475584, 0.017500160217285156, 0.017491968154907226, 0.017518592834472657, 0.017477632522583008, 0.017480703353881837, 0.017496063232421876, 0.017512447357177736, 0.017491968154907226, 0.017512447357177736, 0.017511423110961915, 0.017503231048583985, 0.01749504089355469, 0.017492992401123047, 0.017525760650634766, 0.017516544342041016, 0.017492992401123047, 0.017529855728149413, 0.017500160217285156, 0.017532928466796875, 0.017514495849609374, 0.017508352279663086, 0.017517568588256836, 0.017531904220581054, 0.017529855728149413, 0.017566719055175782, 0.017527807235717775, 0.017524736404418945, 0.017539072036743163, 0.017521663665771483, 0.017546239852905272, 0.017542144775390626, 0.017542144775390626, 0.017565696716308594, 0.017533952713012696, 0.017526784896850587, 0.017575935363769533, 0.01754521560668945, 0.01757695960998535, 0.0175861759185791, 0.017542144775390626, 0.017564672470092774, 0.017580032348632812, 0.017590272903442384, 0.01759129524230957, 0.01761177635192871, 0.017574911117553712, 0.017588224411010742, 0.01759129524230957, 0.017550336837768556, 0.017613824844360353, 0.017557504653930665, 0.01757900810241699, 0.01762099266052246, 0.017584127426147463, 0.017584127426147463, 0.01760153579711914, 0.017588224411010742, 0.0175994873046875, 0.05641625595092774, 0.01744691276550293, 0.017489919662475584, 0.01746124839782715, 0.017466367721557616, 0.017524736404418945, 0.017473535537719728, 0.017482751846313475, 0.017492992401123047, 0.017482751846313475, 0.017477632522583008, 0.017494016647338868, 0.017508352279663086, 0.017518592834472657, 0.017490943908691405, 0.017491968154907226, 0.017551359176635743, 0.017503231048583985, 0.017515520095825195, 0.017491968154907226, 0.017492992401123047, 0.017546239852905272, 0.017512447357177736, 0.017505279541015627, 0.017518592834472657, 0.017508352279663086, 0.017525760650634766, 0.017521663665771483, 0.017520639419555666, 0.017534975051879884, 0.017500160217285156, 0.017533952713012696, 0.017541120529174805, 0.017531904220581054, 0.017529855728149413, 0.017565696716308594, 0.017539072036743163, 0.0175861759185791, 0.017550336837768556, 0.017540096282958984, 0.017542144775390626, 0.017549312591552735, 0.017543167114257813, 0.017547264099121093, 0.017582080841064454, 0.017572864532470703, 0.01758720016479492, 0.0175861759185791, 0.017549312591552735, 0.017584127426147463, 0.017574911117553712, 0.017570816040039062, 0.017565696716308594, 0.01756876754760742, 0.01759129524230957, 0.01756979179382324, 0.01757798385620117, 0.01759539222717285, 0.017566719055175782, 0.017574911117553712, 0.01759231948852539, 0.017583103179931642, 0.01760767936706543, 0.056369152069091794, 0.017480703353881837, 0.01744486427307129, 0.017485824584960938, 0.017466367721557616, 0.01747148895263672, 0.017490943908691405, 0.01748684883117676, 0.017525760650634766, 0.017497087478637697, 0.017522687911987304, 0.017535999298095704, 0.017475584030151366, 0.017509376525878906, 0.017515520095825195, 0.017507328033447265, 0.017518592834472657, 0.017489919662475584, 0.017556480407714844, 0.017580032348632812, 0.017490943908691405, 0.017515520095825195, 0.017519615173339845, 0.017500160217285156, 0.017558528900146485, 0.017517568588256836, 0.017514495849609374, 0.017538047790527343, 0.017546239852905272, 0.01755340766906738, 0.017531904220581054, 0.017516544342041016, 0.01755340766906738, 0.017548288345336914, 0.017532928466796875, 0.017547264099121093, 0.017556480407714844, 0.017565696716308594, 0.017550336837768556, 0.017559551239013673, 0.017535999298095704, 0.017547264099121093, 0.017572864532470703, 0.017527807235717775, 0.017572864532470703, 0.017564672470092774, 0.017571840286254883, 0.017548288345336914, 0.017573888778686524, 0.017574911117553712, 0.017583103179931642, 0.017570816040039062, 0.017605632781982423, 0.017589248657226563, 0.017559551239013673, 0.0175861759185791, 0.017590272903442384, 0.01760870361328125, 0.017589248657226563, 0.017563648223876953, 0.01759231948852539, 0.0175994873046875, 0.01760767936706543, 0.056406017303466796, 0.017475584030151366, 0.017468416213989257, 0.01747865676879883, 0.01747865676879883, 0.017498111724853514, 0.017491968154907226, 0.01747865676879883, 0.017481727600097655, 0.017511423110961915, 0.017509376525878906, 0.017504255294799806, 0.017498111724853514, 0.017501184463500977, 0.017514495849609374, 0.017472511291503907, 0.017508352279663086, 0.017516544342041016, 0.017479679107666016, 0.017551359176635743, 0.017508352279663086, 0.017506303787231444, 0.017485824584960938, 0.017521663665771483, 0.017490943908691405, 0.017514495849609374, 0.017528831481933595, 0.01754521560668945, 0.017574911117553712, 0.017521663665771483, 0.017522687911987304, 0.017539072036743163, 0.017547264099121093, 0.017541120529174805, 0.017538047790527343, 0.017533952713012696, 0.017550336837768556, 0.01757798385620117, 0.017555456161499023, 0.017556480407714844, 0.017534975051879884, 0.017562623977661132, 0.017565696716308594, 0.017539072036743163, 0.017550336837768556, 0.01760870361328125, 0.0175861759185791, 0.017540096282958984, 0.01757695960998535, 0.017590272903442384, 0.017617919921875, 0.017564672470092774, 0.01759129524230957, 0.01757900810241699, 0.017571840286254883, 0.017590272903442384, 0.017571840286254883, 0.01759231948852539, 0.0175861759185791, 0.017571840286254883, 0.017596416473388672, 0.017588224411010742, 0.017632255554199217, 0.056368129730224606, 0.017458175659179686, 0.01744691276550293, 0.017472511291503907, 0.017468416213989257, 0.017501184463500977, 0.017498111724853514, 0.017498111724853514, 0.017503231048583985, 0.017485824584960938, 0.017504255294799806, 0.017521663665771483, 0.0174704647064209, 0.017497087478637697, 0.017497087478637697, 0.017468416213989257, 0.017509376525878906, 0.017518592834472657, 0.017479679107666016, 0.017534975051879884, 0.01761484718322754, 0.017532928466796875, 0.017535999298095704, 0.017520639419555666, 0.017508352279663086, 0.017537023544311522, 0.017526784896850587, 0.017552383422851564, 0.01762303924560547, 0.01765068817138672, 0.01757695960998535, 0.01757695960998535, 0.0176312313079834, 0.017531904220581054, 0.017547264099121093, 0.017564672470092774, 0.017552383422851564, 0.017549312591552735, 0.01755340766906738, 0.017559551239013673, 0.017549312591552735, 0.017547264099121093, 0.017572864532470703, 0.01756876754760742, 0.017567743301391603, 0.017558528900146485, 0.01760256004333496, 0.0175861759185791, 0.017555456161499023, 0.017550336837768556, 0.01757798385620117, 0.017581056594848633, 0.01757798385620117, 0.01757695960998535, 0.017589248657226563, 0.01759334373474121, 0.017604608535766602, 0.0175994873046875, 0.017583103179931642, 0.01764556884765625, 0.01769267272949219, 0.017581056594848633, 0.01760051155090332, 0.056438785552978515, 0.017469440460205078, 0.017512447357177736, 0.017530879974365234, 0.017498111724853514, 0.017507328033447265, 0.017474559783935546, 0.017483776092529296, 0.017490943908691405, 0.017477632522583008, 0.017504255294799806, 0.017487871170043946, 0.017480703353881837, 0.017485824584960938, 0.017508352279663086, 0.017502208709716797, 0.017514495849609374, 0.017511423110961915, 0.017507328033447265, 0.017506303787231444, 0.017507328033447265, 0.017544191360473634, 0.017514495849609374, 0.017515520095825195, 0.01760870361328125, 0.017570816040039062, 0.017562623977661132, 0.017549312591552735, 0.017507328033447265, 0.017554431915283202, 0.017513471603393553, 0.017542144775390626, 0.017537023544311522, 0.017541120529174805, 0.01756876754760742, 0.017538047790527343, 0.017526784896850587, 0.017565696716308594, 0.017562623977661132, 0.017544191360473634, 0.017566719055175782, 0.017554431915283202, 0.017571840286254883, 0.017567743301391603, 0.01754521560668945, 0.017559551239013673, 0.017575935363769533, 0.017583103179931642, 0.017583103179931642, 0.01760767936706543, 0.017559551239013673, 0.017575935363769533, 0.01760767936706543, 0.017570816040039062, 0.017571840286254883, 0.01756876754760742, 0.017712127685546874, 0.017612800598144532, 0.017583103179931642, 0.01759846305847168, 0.017589248657226563, 0.017596416473388672, 0.01760256004333496, 0.056441856384277345, 0.017473535537719728, 0.0174653434753418, 0.017480703353881837, 0.017492992401123047, 0.017512447357177736, 0.017500160217285156, 0.017489919662475584, 0.017524736404418945, 0.017488895416259767, 0.017498111724853514, 0.01748684883117676, 0.017494016647338868, 0.017503231048583985, 0.017490943908691405, 0.01747148895263672, 0.017509376525878906, 0.017494016647338868, 0.017498111724853514, 0.017508352279663086, 0.017516544342041016, 0.017517568588256836, 0.017516544342041016, 0.017518592834472657, 0.017527807235717775, 0.017551359176635743, 0.017541120529174805, 0.017511423110961915, 0.017527807235717775, 0.01760051155090332, 0.017533952713012696, 0.017566719055175782, 0.017531904220581054, 0.017534975051879884, 0.017557504653930665, 0.017526784896850587, 0.017529855728149413, 0.017539072036743163, 0.017541120529174805, 0.017546239852905272, 0.017570816040039062, 0.017571840286254883, 0.017566719055175782, 0.017563648223876953, 0.017571840286254883, 0.017546239852905272, 0.017622016906738282, 0.017588224411010742, 0.017546239852905272, 0.017555456161499023, 0.01757900810241699, 0.017581056594848633, 0.01759334373474121, 0.017581056594848633, 0.017573888778686524, 0.017570816040039062, 0.017597440719604493, 0.017589248657226563, 0.01757695960998535, 0.0175994873046875, 0.0175994873046875, 0.01757798385620117, 0.017597440719604493, 0.056594432830810545, 0.017483776092529296, 0.0174653434753418, 0.01746329689025879, 0.01747148895263672, 0.017497087478637697, 0.017482751846313475, 0.017467391967773437, 0.017494016647338868, 0.017500160217285156, 0.017511423110961915, 0.017481727600097655, 0.017507328033447265, 0.017507328033447265, 0.017557504653930665, 0.017499135971069335, 0.01761689567565918, 0.017516544342041016, 0.017514495849609374, 0.017497087478637697, 0.017494016647338868, 0.017546239852905272, 0.017534975051879884, 0.017530879974365234, 0.01748684883117676, 0.017513471603393553, 0.017678335189819337, 0.017550336837768556, 0.017502208709716797, 0.017549312591552735, 0.017527807235717775, 0.017542144775390626, 0.017552383422851564, 0.017532928466796875, 0.017547264099121093, 0.017520639419555666, 0.017542144775390626, 0.017546239852905272, 0.017544191360473634, 0.01760665512084961, 0.017589248657226563, 0.017560575485229494, 0.017564672470092774, 0.017556480407714844, 0.017572864532470703, 0.01756979179382324, 0.017555456161499023, 0.017554431915283202, 0.017563648223876953, 0.017613824844360353, 0.017722368240356445, 0.01764249610900879, 0.01761075210571289, 0.01760767936706543, 0.01759129524230957, 0.017572864532470703, 0.017573888778686524, 0.01760256004333496, 0.017570816040039062, 0.01757798385620117, 0.017575935363769533, 0.017597440719604493, 0.01765478324890137]",tokens/s,55.198734755604704,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,google/recurrentgemma-9b,google/recurrentgemma-9b,cuda,0,42,,,True,,,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",recurrent_gemma,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,mistralai/Mixtral-8x22B-v0.1,mistralai/Mixtral-8x22B-v0.1,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/base.py"", line 43, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 54, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1021, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 403 Client Error: Forbidden for url: https://huggingface.co/mistralai/Mixtral-8x22B-v0.1/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 398, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 321, in hf_raise_for_status raise GatedRepoError(message, response) from e huggingface_hub.utils._errors.GatedRepoError: 403 Client Error. (Request ID: Root=1-66475089-642ac3c560550424022093b6;ff7740ca-5e44-404c-8a1a-a09eec9a2a44) Cannot access gated repo for url https://huggingface.co/mistralai/Mixtral-8x22B-v0.1/resolve/main/config.json. Access to model mistralai/Mixtral-8x22B-v0.1 is restricted and you are not in the authorized list. Visit https://huggingface.co/mistralai/Mixtral-8x22B-v0.1 to ask for access. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 87, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/base.py"", line 56, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 37, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 62, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 928, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 631, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 686, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 416, in cached_file raise EnvironmentError( OSError: You are trying to access a gated repo. Make sure to have access to it at https://huggingface.co/mistralai/Mixtral-8x22B-v0.1. 403 Client Error. (Request ID: Root=1-66475089-642ac3c560550424022093b6;ff7740ca-5e44-404c-8a1a-a09eec9a2a44) Cannot access gated repo for url https://huggingface.co/mistralai/Mixtral-8x22B-v0.1/resolve/main/config.json. Access to model mistralai/Mixtral-8x22B-v0.1 is restricted and you are not in the authorized list. Visit https://huggingface.co/mistralai/Mixtral-8x22B-v0.1 to ask for access. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,01-ai/Yi-6B,01-ai/Yi-6B,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1263.321088,13032.226816,0.0,12385.779712,12315.692032,s,10,0.852141990661621,0.08521419906616212,0.0003264795887369383,0.08511492919921876,0.08529609451293944,0.085736429977417,0.08608869834899903,"[0.08617676544189454, 0.08503734588623046, 0.08510870361328125, 0.08512115478515625, 0.0850223388671875, 0.08517027282714844, 0.08509820556640625, 0.08502777862548828, 0.08518118286132813, 0.0851982421875]",tokens/s,3004.194169580074,kWh,1.0085146656705387e-06,5.526200274227634e-07,5.132174538881362e-06,6.693309231974664e-06,tokens/kWh,38247149.67255065,MB,1263.648768,13032.226816,0.0,12385.779712,12315.694592,s,10,20.959052001953125,2.0959052001953125,0.022875737660153668,2.1014481201171877,2.121259033203125,2.1222686767578125,2.1230763916015625,"[2.096263671875, 2.077789794921875, 2.04854736328125, 2.06796923828125, 2.0999189453125, 2.1232783203125, 2.111183349609375, 2.11008935546875, 2.102977294921875, 2.12103466796875]",tokens/s,30.058611426761658,kWh,2.436010769363177e-05,1.3348122345784559e-05,0.00010214391128191866,0.00013985214132133497,tokens/kWh,450475.7625072496,,s,629,21.690339378356928,0.034483846388484785,0.010119238945336317,0.03354009628295898,0.033925927734375,0.03412561950683594,0.11806199676513672,"[0.03441049575805664, 0.033947647094726564, 0.0335206413269043, 0.033529857635498046, 0.033947647094726564, 0.03375820922851563, 0.034016254425048825, 0.03482316970825195, 0.033882110595703126, 0.034344959259033206, 0.033649665832519535, 0.033516544342041016, 0.03419033432006836, 0.03458259201049805, 0.034310081481933596, 0.034028545379638675, 0.03345414352416992, 0.03346425628662109, 0.033742847442626955, 0.032484352111816404, 0.032502784729003906, 0.03257753753662109, 0.032363521575927735, 0.03240755081176758, 0.03252633666992188, 0.032553985595703126, 0.03264102554321289, 0.032535552978515625, 0.03251814270019531, 0.032478206634521486, 0.03249667358398438, 0.03271676635742188, 0.03446988677978516, 0.03287756729125976, 0.03255807876586914, 0.03237376022338867, 0.03302912139892578, 0.03244543838500977, 0.03261849594116211, 0.03244236755371094, 0.032606208801269534, 0.03253964614868164, 0.03335168075561523, 0.03272294235229492, 0.03251200103759765, 0.03250995254516602, 0.0324771842956543, 0.03387596893310547, 0.03418316650390625, 0.033995777130126956, 0.03272806549072266, 0.03322675323486328, 0.03362713623046875, 0.03335782241821289, 0.03369881439208984, 0.03284889602661133, 0.03234406280517578, 0.03338547134399414, 0.033699840545654294, 0.03354111862182617, 0.033710079193115236, 0.03359436798095703, 0.11807231903076172, 0.033532928466796875, 0.03234099197387695, 0.03234304046630859, 0.03265024185180664, 0.033498111724853515, 0.033565696716308595, 0.03298406219482422, 0.033547264099121094, 0.033591327667236326, 0.03352163314819336, 0.033486846923828126, 0.0334284782409668, 0.032486400604248046, 0.03257753753662109, 0.03407769775390625, 0.03396710586547851, 0.0336814079284668, 0.03354214477539062, 0.033535999298095705, 0.03346636962890625, 0.033391616821289063, 0.03354111862182617, 0.03347148895263672, 0.03235846328735351, 0.03241567993164062, 0.03258060836791992, 0.03244543838500977, 0.032484352111816404, 0.03222630310058594, 0.03250483322143555, 0.03241779327392578, 0.032467967987060545, 0.032519168853759765, 0.032702529907226566, 0.03259590530395508, 0.032672767639160154, 0.03256115341186523, 0.03403776168823242, 0.03258163070678711, 0.03341107177734375, 0.03345817565917969, 0.03355033493041992, 0.033539070129394534, 0.0334284782409668, 0.033800193786621094, 0.03357388687133789, 0.03361587142944336, 0.03353497695922852, 0.03279667282104492, 0.032540672302246096, 0.032573440551757815, 0.03257651138305664, 0.03246899032592773, 0.03260518264770508, 0.032473087310791016, 0.03231846237182617, 0.032502784729003906, 0.0324956169128418, 0.03239321517944336, 0.03239424133300781, 0.032494590759277346, 0.03244236755371094, 0.11931033325195313, 0.0326297607421875, 0.03233491134643555, 0.03239110565185547, 0.0323768310546875, 0.03234304046630859, 0.032360511779785155, 0.03227436828613281, 0.032279552459716795, 0.03251507186889648, 0.032467967987060545, 0.032508926391601564, 0.0324505615234375, 0.03238092803955078, 0.03242598342895508, 0.03240556716918945, 0.032871360778808596, 0.03243215942382813, 0.03235939025878906, 0.03261439895629883, 0.03237478256225586, 0.03243315124511719, 0.03238304138183594, 0.032462783813476566, 0.03246284866333008, 0.03254886245727539, 0.03254476928710937, 0.03252633666992188, 0.0325928955078125, 0.032440319061279296, 0.0325591049194336, 0.03261337661743164, 0.03260115051269531, 0.032803775787353516, 0.032889854431152346, 0.032517120361328124, 0.032456703186035156, 0.032524288177490236, 0.032489471435546875, 0.03261654281616211, 0.03251395034790039, 0.032430145263671876, 0.03249964904785156, 0.0323768310546875, 0.03267686462402344, 0.032540672302246096, 0.03242496109008789, 0.032285694122314454, 0.03237478256225586, 0.03250175857543945, 0.03259187316894531, 0.032517120361328124, 0.032489471435546875, 0.03241779327392578, 0.032435199737548825, 0.03243008041381836, 0.032415744781494144, 0.03242496109008789, 0.03250688171386719, 0.0325928955078125, 0.0325928955078125, 0.032500736236572264, 0.03256115341186523, 0.11815628814697265, 0.03239731216430664, 0.032492542266845705, 0.03240959930419922, 0.03250380706787109, 0.032499713897705076, 0.032456703186035156, 0.03245568084716797, 0.03249049758911133, 0.032473087310791016, 0.032481281280517575, 0.03245772933959961, 0.032473087310791016, 0.032464897155761716, 0.03235430526733398, 0.03246387100219727, 0.032478271484375, 0.032575424194335935, 0.03244144058227539, 0.03253647994995117, 0.032402431488037106, 0.03259084701538086, 0.03247411346435547, 0.032648193359375, 0.032553985595703126, 0.032623615264892575, 0.03265945434570312, 0.03284479904174804, 0.03266559982299805, 0.03340595245361328, 0.03393843078613281, 0.03295743942260742, 0.03367833709716797, 0.03367427062988281, 0.0336844482421875, 0.03377766418457031, 0.03260825729370117, 0.03254483032226563, 0.03251705551147461, 0.03246899032592773, 0.0326195182800293, 0.032451583862304685, 0.032541694641113283, 0.032478206634521486, 0.032467967987060545, 0.03251814270019531, 0.03242803192138672, 0.032502784729003906, 0.03245772933959961, 0.03256217575073242, 0.03256422424316406, 0.032525310516357424, 0.0326195182800293, 0.032717823028564456, 0.03399174499511719, 0.03411654281616211, 0.033713153839111325, 0.033620990753173825, 0.03398252868652344, 0.0337407341003418, 0.03367833709716797, 0.033568767547607424, 0.03345305633544922, 0.11823104095458985, 0.03256729507446289, 0.03268198394775391, 0.03331584167480469, 0.033230846405029296, 0.03344384002685547, 0.03346124649047852, 0.03361280059814453, 0.0336363525390625, 0.03365478515625, 0.03560857772827149, 0.03378688049316406, 0.033132545471191405, 0.0332410888671875, 0.033500160217285156, 0.03349606323242187, 0.03359641647338867, 0.033568767547607424, 0.033346622467041016, 0.03324716949462891, 0.03351039886474609, 0.03381452941894531, 0.03379302215576172, 0.03265433502197266, 0.03303424072265625, 0.0330618896484375, 0.03254476928710937, 0.032894977569580076, 0.03252121734619141, 0.03258163070678711, 0.03233894348144531, 0.03242496109008789, 0.03335475158691406, 0.03372032165527344, 0.03422208023071289, 0.03313663864135742, 0.0336517105102539, 0.033293312072753906, 0.033438785552978516, 0.03357487869262695, 0.033166366577148436, 0.03317343902587891, 0.03251302337646484, 0.03281100845336914, 0.032947200775146485, 0.03300556945800781, 0.0331038703918457, 0.033309696197509765, 0.03243417739868164, 0.033221694946289064, 0.03350624084472656, 0.03355852890014648, 0.03377459335327149, 0.03355340957641602, 0.03359743881225586, 0.033529857635498046, 0.033532928466796875, 0.033565696716308595, 0.03362713623046875, 0.03349606323242187, 0.03423231887817383, 0.033683456420898435, 0.03373056030273437, 0.11855462646484374, 0.033585151672363284, 0.03360768127441406, 0.03354214477539062, 0.033880062103271484, 0.033807361602783206, 0.03344998550415039, 0.03261747360229492, 0.03256729507446289, 0.03403059387207031, 0.03424665451049805, 0.03367833709716797, 0.03363123321533203, 0.03360665512084961, 0.0335175666809082, 0.033702911376953124, 0.03361587142944336, 0.03363020706176758, 0.03337932968139649, 0.03370700836181641, 0.03379097747802735, 0.03348992156982422, 0.033999870300292966, 0.03508428955078125, 0.03435520172119141, 0.03405311965942383, 0.033849342346191406, 0.033068031311035154, 0.03397635269165039, 0.03383804702758789, 0.0338155517578125, 0.03372544097900391, 0.034253822326660154, 0.033620990753173825, 0.03384012985229492, 0.03396198272705078, 0.03364659118652344, 0.033672191619873046, 0.033957889556884766, 0.03373673629760742, 0.03371414566040039, 0.033568767547607424, 0.03355136108398438, 0.03351244735717773, 0.03369062423706055, 0.033567745208740236, 0.03365990447998047, 0.03362303924560547, 0.03360768127441406, 0.033721343994140625, 0.03358924865722656, 0.033116161346435545, 0.033552383422851564, 0.033683456420898435, 0.03359334564208984, 0.03449856185913086, 0.03407769775390625, 0.03382374572753906, 0.03399782562255859, 0.03321651077270508, 0.033742847442626955, 0.033713153839111325, 0.03357183837890625, 0.11795661163330078, 0.03266867065429688, 0.033554431915283206, 0.0336517105102539, 0.03365683364868164, 0.03373056030273437, 0.033605663299560544, 0.03360764694213867, 0.03379916763305664, 0.03370598220825195, 0.03364352035522461, 0.033571872711181644, 0.0336926383972168, 0.03344998550415039, 0.033753089904785157, 0.033726463317871096, 0.03371417617797851, 0.0337336311340332, 0.03372748947143555, 0.03362201690673828, 0.033650688171386715, 0.03354627227783203, 0.03352880096435547, 0.03370393753051758, 0.03364761734008789, 0.033186817169189455, 0.03371929550170898, 0.03321548843383789, 0.03334041595458984, 0.03407769775390625, 0.033770496368408204, 0.0337367057800293, 0.03376537704467773, 0.03375001525878906, 0.03372236633300781, 0.033723392486572266, 0.03363430404663086, 0.033701889038085936, 0.03372032165527344, 0.03417190551757812, 0.03371212768554688, 0.03361996841430664, 0.03369779205322266, 0.03356364822387695, 0.03360665512084961, 0.033685504913330076, 0.03356979370117188, 0.03361177444458008, 0.033588222503662106, 0.03369472122192383, 0.03362201690673828, 0.03360358428955078, 0.033017856597900394, 0.03335782241821289, 0.03379814529418945, 0.03369676971435547, 0.033638465881347654, 0.03247507095336914, 0.03256934356689453, 0.03260518264770508, 0.03257958221435547, 0.03263078308105469, 0.03250688171386719, 0.11803545379638672, 0.032557056427001956, 0.03269734573364258, 0.03253760147094727, 0.03264716720581055, 0.03355955123901367, 0.03386368179321289, 0.03404800033569336, 0.03388927841186523, 0.03369267272949219, 0.033549312591552735, 0.033567745208740236, 0.033588222503662106, 0.033629215240478516, 0.03362198257446289, 0.03354828643798828, 0.03362508773803711, 0.033742847442626955, 0.033721343994140625, 0.03364659118652344, 0.03364556884765625, 0.033568767547607424, 0.03362617492675781, 0.033646526336669924, 0.033699840545654294, 0.03362508773803711, 0.033772544860839845, 0.03374899291992187, 0.03360358428955078, 0.03297894287109375, 0.033723392486572266, 0.0337305908203125, 0.03404079818725586, 0.03371523284912109, 0.03379203033447266, 0.03372947311401367, 0.033713153839111325, 0.03370905685424805, 0.03364659118652344, 0.033699840545654294, 0.03367628860473633, 0.03359539031982422, 0.03371417617797851, 0.03311001586914063, 0.03352371215820313, 0.03362611389160156, 0.033462272644042966, 0.033627166748046874, 0.03344073486328125, 0.03366912078857422, 0.03361996841430664, 0.03265433502197266, 0.032547870635986326, 0.03252220916748047, 0.03262771224975586, 0.03343564987182617, 0.033576961517333984, 0.03358316802978516, 0.033605567932128905, 0.03354009628295898, 0.03359747314453125, 0.03355849456787109, 0.03363942337036133, 0.11819213104248047, 0.03261337661743164, 0.03236556625366211, 0.03247923278808594, 0.032543743133544925, 0.03249049758911133, 0.032440319061279296, 0.03263385772705078, 0.03325030517578125, 0.033546241760253906, 0.033567745208740236, 0.033554431915283206, 0.033598464965820314, 0.032697376251220704, 0.03262563323974609, 0.03246080017089844, 0.03252326583862305, 0.03256115341186523, 0.033263614654541016, 0.033497150421142576, 0.033610687255859376, 0.03365478515625, 0.03342745590209961, 0.03318374252319336, 0.033432575225830076, 0.03448729705810547, 0.03411763381958008, 0.03381350326538086, 0.0337448959350586, 0.03356364822387695, 0.03372748947143555, 0.03322470474243164, 0.03366809463500976, 0.03294515228271484, 0.032694271087646484, 0.033649665832519535, 0.03413094329833984, 0.03380428695678711, 0.0336363525390625, 0.03363123321533203, 0.03359641647338867, 0.03361382293701172, 0.033667072296142575, 0.03346022415161133, 0.0336445426940918, 0.033547264099121094, 0.033650688171386715, 0.03402035140991211, 0.033535999298095705, 0.03371110534667969, 0.03338854217529297, 0.033127422332763674, 0.03356364822387695, 0.033670143127441404, 0.03357491302490234, 0.03359436798095703, 0.033667072296142575, 0.03339775848388672, 0.03364147186279297, 0.03383193588256836, 0.033788928985595705, 0.033808383941650394, 0.033650688171386715, 0.11842150115966797, 0.03256934356689453, 0.033050624847412106, 0.033740833282470704, 0.03381039810180664, 0.03369676971435547, 0.033726463317871096, 0.03376025772094727, 0.03375820922851563, 0.03389952087402344, 0.03358924865722656, 0.033964031219482424, 0.03365273666381836, 0.0336824951171875, 0.03364448165893555, 0.03366092681884766, 0.03372851181030274, 0.033775615692138675, 0.03367424011230469, 0.033591327667236326, 0.03352265548706055, 0.033616897583007815, 0.033584129333496096, 0.033691646575927735, 0.033672191619873046, 0.03279052734375, 0.032764926910400394, 0.032723968505859374, 0.03267071914672852, 0.03256217575073242, 0.03319507217407226, 0.033929153442382814, 0.0337786865234375, 0.03387596893310547, 0.03369580841064453, 0.03384313583374023, 0.033925121307373046, 0.03408998489379883, 0.03364659118652344, 0.03383295822143555, 0.033893375396728515, 0.03357388687133789, 0.0337336311340332, 0.03359539031982422, 0.034083839416503905, 0.03389440155029297, 0.03491027069091797, 0.03429574584960938, 0.03386880111694336, 0.03379916763305664, 0.033982463836669925, 0.03372748947143555, 0.03366502380371094, 0.033649665832519535, 0.03377664184570312, 0.03362406539916992, 0.03346944046020508, 0.03360563278198242, 0.03347148895263672, 0.034252799987792966, 0.034977825164794925, 0.03405923080444336, 0.033532928466796875]",tokens/s,28.999085216141395,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,/,/,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 373, in cached_file raise EnvironmentError( OSError: / does not appear to have a file named config.json. Checkout 'https://huggingface.co///tree/None' for available files. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,TencentARC/Mistral_Pro_8B_v0.1,,cuda,0,42,,,,,True,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,d35829e539df8480b726c647eeabf91e41eae047,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,,MB,1266.393088,19665.518592,0.0,19019.071488,18747.268096,s,10,0.9985231781005859,0.09985231781005859,0.0003397017608118708,0.09978948593139648,0.10010437698364258,0.1004238697052002,0.1006794638824463,"[0.10074336242675781, 0.099523681640625, 0.09960009765625, 0.09980255889892578, 0.10003337860107422, 0.09977641296386719, 0.09961027526855469, 0.09960505676269531, 0.10001417541503907, 0.09981417846679687]",tokens/s,2563.7862556878167,kWh,1.1808174753477601e-06,6.470316232471395e-07,7.209813248594475e-06,9.037662347189373e-06,tokens/kWh,28325908.865096465,MB,1266.393088,19665.518592,0.0,19019.071488,18774.921728,s,10,28.944365234374995,2.8944365234374994,0.0028451327263455473,2.893563232421875,2.8956759765625,2.8991606201171876,2.9019483349609376,"[2.902645263671875, 2.89277685546875, 2.894901611328125, 2.8943369140625, 2.893554931640625, 2.893571533203125, 2.893523193359375, 2.894090576171875, 2.891999755859375, 2.892964599609375]",tokens/s,21.765894497896866,kWh,3.4121333750987344e-05,1.8700035649691004e-05,0.00014284441790580617,0.00019566578730648447,tokens/kWh,321977.59693839,,s,629,29.795304481506335,0.04736932350001009,0.011931217203109688,0.045911041259765625,0.046023681640625,0.04617359313964844,0.14624359619140626,"[0.04660838317871094, 0.0465428466796875, 0.04661663818359375, 0.0465601921081543, 0.04660531234741211, 0.046614528656005856, 0.04659097671508789, 0.0465530891418457, 0.046440448760986325, 0.04643430328369141, 0.04605449676513672, 0.045857696533203124, 0.045859840393066405, 0.04583225631713867, 0.04585977554321289, 0.04588544082641602, 0.0459683837890625, 0.0458752326965332, 0.04595094299316406, 0.04588851165771484, 0.04601139068603516, 0.04627763366699219, 0.046069759368896485, 0.04588339233398438, 0.04593766403198242, 0.045985790252685545, 0.04590694427490234, 0.04590694427490234, 0.04591001510620117, 0.04585062408447266, 0.04589875030517578, 0.045884414672851564, 0.04591206359863281, 0.04597350311279297, 0.046296062469482424, 0.0464005126953125, 0.045969406127929685, 0.04587519836425781, 0.04590796661376953, 0.04589363098144531, 0.04588041687011719, 0.045894561767578126, 0.04591820907592774, 0.04599603271484375, 0.04589977645874024, 0.04620697784423828, 0.0459683837890625, 0.04600934219360352, 0.04589977645874024, 0.04597350311279297, 0.04589670562744141, 0.046102527618408204, 0.045992961883544924, 0.04607590484619141, 0.045927425384521485, 0.04596428680419922, 0.045864959716796876, 0.04595302581787109, 0.04591820907592774, 0.045927425384521485, 0.04592230224609375, 0.0459315185546875, 0.14600498962402345, 0.04586700820922852, 0.04580147171020508, 0.04583935928344727, 0.04583119964599609, 0.045924320220947265, 0.04578201675415039, 0.045970497131347654, 0.04591404724121094, 0.04582912063598633, 0.04587724685668945, 0.04589977645874024, 0.04586598587036133, 0.04595609664916992, 0.045878273010253906, 0.045818878173828126, 0.04602675247192383, 0.045830142974853515, 0.04609638214111328, 0.04597964859008789, 0.045908992767333984, 0.04588236618041992, 0.04585779190063476, 0.04586905670166016, 0.045856769561767576, 0.04585785675048828, 0.04588127899169922, 0.045848575592041016, 0.0459048957824707, 0.045917182922363284, 0.04584243011474609, 0.045884414672851564, 0.04586393737792969, 0.045914112091064455, 0.04585779190063476, 0.04588236618041992, 0.04588748931884765, 0.04592947387695313, 0.045897727966308595, 0.04595609664916992, 0.04588339233398438, 0.045884414672851564, 0.04593561553955078, 0.045900833129882815, 0.045962207794189455, 0.045886463165283206, 0.0459417610168457, 0.04593868637084961, 0.045969406127929685, 0.04591206359863281, 0.04600729751586914, 0.04595609664916992, 0.04590694427490234, 0.04593971252441406, 0.04595097732543945, 0.04595404815673828, 0.04599097442626953, 0.04598886489868164, 0.04595507049560547, 0.04592736053466797, 0.04595199966430664, 0.045914112091064455, 0.04594790267944336, 0.14631015014648438, 0.045868030548095705, 0.04579328155517578, 0.045840385437011716, 0.045948928833007815, 0.04584755325317383, 0.04582195281982422, 0.04592127990722656, 0.04586393737792969, 0.04600831985473633, 0.045848575592041016, 0.045879295349121094, 0.04584755325317383, 0.04587212753295898, 0.04587724685668945, 0.04589670562744141, 0.045871105194091794, 0.04585779190063476, 0.045841407775878903, 0.04621619033813477, 0.04594278335571289, 0.0459090576171875, 0.046051265716552735, 0.045868030548095705, 0.0459048957824707, 0.04600012969970703, 0.046101566314697265, 0.0458935661315918, 0.04588339233398438, 0.045902847290039066, 0.04592639923095703, 0.045917182922363284, 0.045917182922363284, 0.04585472106933594, 0.045873153686523435, 0.04592947387695313, 0.04589977645874024, 0.04592025756835937, 0.0459048957824707, 0.04595199966430664, 0.046023681640625, 0.04600630569458008, 0.045962207794189455, 0.045914112091064455, 0.04596121597290039, 0.04595097732543945, 0.04592435073852539, 0.04614963150024414, 0.045894657135009766, 0.046018558502197264, 0.04591308975219727, 0.04601958465576172, 0.04592230224609375, 0.04593664169311523, 0.04591001510620117, 0.046023681640625, 0.04590796661376953, 0.04598067092895508, 0.04602163314819336, 0.0459683837890625, 0.045969406127929685, 0.04598988723754883, 0.04596025466918945, 0.14676985168457032, 0.04588032150268555, 0.04586598587036133, 0.04586905670166016, 0.045862911224365234, 0.04584243011474609, 0.04588851165771484, 0.04584447860717773, 0.04586598587036133, 0.045859840393066405, 0.045894657135009766, 0.0458342399597168, 0.04592540740966797, 0.04582499313354492, 0.045861888885498046, 0.04583321762084961, 0.04589567947387695, 0.045884414672851564, 0.04588652801513672, 0.04590687942504883, 0.04589363098144531, 0.04588748931884765, 0.04585369491577149, 0.04588544082641602, 0.04581990432739258, 0.04589567947387695, 0.045930496215820314, 0.04600227355957031, 0.04601129531860351, 0.04594483184814453, 0.04586393737792969, 0.04595507049560547, 0.04586393737792969, 0.04588236618041992, 0.04590387344360351, 0.045892608642578124, 0.045932544708251956, 0.04596121597290039, 0.04594278335571289, 0.04591308975219727, 0.04600320053100586, 0.04619161605834961, 0.045902847290039066, 0.04598374557495117, 0.046265342712402346, 0.04605235290527344, 0.04595302581787109, 0.04593664169311523, 0.04598886489868164, 0.04596633529663086, 0.045992961883544924, 0.04595199966430664, 0.045900798797607424, 0.04596223831176758, 0.045894657135009766, 0.04593459320068359, 0.04592230224609375, 0.045892608642578124, 0.04604006576538086, 0.04618547058105469, 0.046189567565917966, 0.04598988723754883, 0.04596633529663086, 0.14672691345214844, 0.045824001312255856, 0.045818878173828126, 0.04602470397949219, 0.04591308975219727, 0.04587724685668945, 0.04591308975219727, 0.045846527099609374, 0.04586700820922852, 0.045911041259765625, 0.04585580825805664, 0.04584748840332031, 0.04582809448242187, 0.045835296630859376, 0.04579119873046875, 0.045902847290039066, 0.045870079040527346, 0.045859840393066405, 0.04584864044189453, 0.045975486755371095, 0.04595206451416016, 0.045902782440185544, 0.04587318420410156, 0.045861888885498046, 0.04588028717041016, 0.04589670562744141, 0.04587519836425781, 0.04581785583496094, 0.04594790267944336, 0.04596121597290039, 0.04597555160522461, 0.04591308975219727, 0.04590796661376953, 0.04588851165771484, 0.04591206359863281, 0.04592025756835937, 0.04590387344360351, 0.045911041259765625, 0.04591820907592774, 0.04598169708251953, 0.046129150390625, 0.04610867309570312, 0.045960193634033204, 0.04593561553955078, 0.04596223831176758, 0.04591308975219727, 0.04588339233398438, 0.04602265548706055, 0.04590387344360351, 0.04595097732543945, 0.04591001510620117, 0.045963264465332034, 0.04592230224609375, 0.04592844772338867, 0.04592537689208984, 0.04593971252441406, 0.04591820907592774, 0.04609331130981445, 0.04588339233398438, 0.04595609664916992, 0.04591206359863281, 0.04597964859008789, 0.04599398422241211, 0.14631219482421876, 0.045827072143554685, 0.04583731079101563, 0.04583116912841797, 0.04602163314819336, 0.045946880340576174, 0.045859840393066405, 0.045846527099609374, 0.04588032150268555, 0.04591206359863281, 0.04585779190063476, 0.04593664169311523, 0.04581478500366211, 0.04589567947387695, 0.04588339233398438, 0.04582809448242187, 0.04588748931884765, 0.04592230224609375, 0.045843456268310545, 0.045902847290039066, 0.045894657135009766, 0.045862911224365234, 0.0458230094909668, 0.04591510391235352, 0.045846527099609374, 0.04592537689208984, 0.04590387344360351, 0.045946880340576174, 0.04585062408447266, 0.0459950065612793, 0.04589670562744141, 0.04592947387695313, 0.045840385437011716, 0.045871105194091794, 0.045900798797607424, 0.04605952072143555, 0.04588032150268555, 0.045897727966308595, 0.045862911224365234, 0.04590796661376953, 0.04591923141479492, 0.04592230224609375, 0.045987838745117186, 0.04589056015014648, 0.04592435073852539, 0.04595609664916992, 0.045963264465332034, 0.04602880096435547, 0.04623155212402344, 0.04600223922729492, 0.046054336547851564, 0.04595302581787109, 0.04599091339111328, 0.0460052490234375, 0.04596121597290039, 0.04592947387695313, 0.04593459320068359, 0.0459417610168457, 0.045963264465332034, 0.04591513442993164, 0.04595711898803711, 0.045948928833007815, 0.04596428680419922, 0.14618832397460937, 0.0459007682800293, 0.04588032150268555, 0.04600320053100586, 0.04598988723754883, 0.04590796661376953, 0.04583628845214844, 0.04587417602539062, 0.04584243011474609, 0.045843456268310545, 0.045870079040527346, 0.04592435073852539, 0.0458526725769043, 0.04585990524291992, 0.04589049530029297, 0.04580966567993164, 0.045878273010253906, 0.04586393737792969, 0.045902847290039066, 0.0459417610168457, 0.04593356704711914, 0.045862911224365234, 0.04587724685668945, 0.04595507049560547, 0.045886463165283206, 0.04588851165771484, 0.04592639923095703, 0.04592435073852539, 0.04586393737792969, 0.04589056015014648, 0.04592230224609375, 0.04590796661376953, 0.04581273651123047, 0.045970497131347654, 0.04586284637451172, 0.04593868637084961, 0.04591820907592774, 0.04591206359863281, 0.04588544082641602, 0.04598886489868164, 0.04589670562744141, 0.04595507049560547, 0.04602675247192383, 0.04594483184814453, 0.045878273010253906, 0.0459417610168457, 0.04593664169311523, 0.04600729751586914, 0.04600831985473633, 0.04599808120727539, 0.04601036834716797, 0.04610867309570312, 0.04589670562744141, 0.04594790267944336, 0.045914112091064455, 0.04599193572998047, 0.04588339233398438, 0.04593561553955078, 0.04597248077392578, 0.0459233283996582, 0.04594278335571289, 0.04593459320068359, 0.0459417610168457, 0.14627839660644532, 0.04617113494873047, 0.04599193572998047, 0.04583116912841797, 0.045843456268310545, 0.04584755325317383, 0.04583628845214844, 0.045843456268310545, 0.045813758850097655, 0.045856769561767576, 0.04581891250610352, 0.04590179061889649, 0.04583939361572266, 0.04585161590576172, 0.04584243011474609, 0.04590800094604492, 0.04590383911132812, 0.04603596878051758, 0.0458260498046875, 0.04610867309570312, 0.04593664169311523, 0.045864959716796876, 0.04583321762084961, 0.045876224517822264, 0.04585881423950195, 0.04589977645874024, 0.04604108810424805, 0.04588236618041992, 0.04587212753295898, 0.04589363098144531, 0.04589977645874024, 0.04601651382446289, 0.0462366714477539, 0.046018558502197264, 0.04591001510620117, 0.045862911224365234, 0.04589567947387695, 0.04583935928344727, 0.04592025756835937, 0.04589670562744141, 0.045943809509277345, 0.045946880340576174, 0.045982719421386715, 0.04592844772338867, 0.045934654235839846, 0.045921215057373045, 0.045985790252685545, 0.045948928833007815, 0.045897727966308595, 0.04595404815673828, 0.04593868637084961, 0.045948928833007815, 0.045856769561767576, 0.045886463165283206, 0.04590182495117188, 0.04593971252441406, 0.04594073486328125, 0.04592025756835937, 0.04592639923095703, 0.04597145462036133, 0.04600115203857422, 0.04602675247192383, 0.04603903961181641, 0.1466808319091797, 0.04581894302368164, 0.045831104278564454, 0.045825023651123044, 0.04585779190063476, 0.045900798797607424, 0.045892608642578124, 0.04585884857177734, 0.04583830261230469, 0.045841438293457035, 0.045823966979980466, 0.045830142974853515, 0.045856769561767576, 0.04584755325317383, 0.04581273651123047, 0.045852737426757814, 0.04587411117553711, 0.04589363098144531, 0.04587519836425781, 0.045871105194091794, 0.04582403182983399, 0.04592534255981445, 0.04580659103393555, 0.04584243011474609, 0.04583833694458008, 0.04591923141479492, 0.04588851165771484, 0.04581990432739258, 0.04584755325317383, 0.045884414672851564, 0.04583833694458008, 0.04597760009765625, 0.04583731079101563, 0.04585881423950195, 0.045864959716796876, 0.045871105194091794, 0.04600627136230469, 0.04587417602539062, 0.045911041259765625, 0.045894657135009766, 0.04591308975219727, 0.04596121597290039, 0.045897727966308595, 0.04590182495117188, 0.04592025756835937, 0.04588748931884765, 0.04602982330322265, 0.04589158248901367, 0.045934654235839846, 0.04590591812133789, 0.045939647674560546, 0.04598886489868164, 0.045870079040527346, 0.04592639923095703, 0.04589875030517578, 0.045946880340576174, 0.0459417610168457, 0.04592947387695313, 0.0459315185546875, 0.04603903961181641, 0.046148609161376954, 0.04608204650878906, 0.045932544708251956, 0.14626509094238282, 0.04580659103393555, 0.045825023651123044, 0.04580659103393555, 0.04617523193359375, 0.04644147109985351, 0.04586905670166016, 0.045879295349121094, 0.04585574340820313, 0.045876224517822264, 0.045835262298583986, 0.0458076171875, 0.045871105194091794, 0.04585369491577149, 0.045878273010253906, 0.04583731079101563, 0.045861888885498046, 0.045846527099609374, 0.045832191467285156, 0.04589977645874024, 0.045873153686523435, 0.04592435073852539, 0.04590796661376953, 0.045878273010253906, 0.04588032150268555, 0.04588544082641602, 0.04596736145019531, 0.04594483184814453, 0.045873153686523435, 0.045884414672851564, 0.04587519836425781, 0.045897727966308595, 0.045886463165283206, 0.04593971252441406, 0.04586393737792969, 0.04593664169311523, 0.04585574340820313, 0.045908992767333984, 0.04590694427490234, 0.04589158248901367, 0.04593561553955078, 0.045985790252685545, 0.04595302581787109, 0.045988929748535155, 0.045907936096191405, 0.04590998458862305, 0.04592844772338867, 0.04595199966430664, 0.04593868637084961, 0.0459233283996582, 0.04598169708251953, 0.0459233283996582, 0.045930496215820314, 0.04592127990722656, 0.045932544708251956, 0.045914112091064455, 0.04592745590209961, 0.04592022323608398, 0.04594278335571289, 0.04590591812133789, 0.04593561553955078, 0.04591820907592774, 0.04592435073852539]",tokens/s,21.110708916917233,,,main,False,False,,,,,,,,,,,,,,,,,,,,,,, bfloat16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-6.7b,EleutherAI/pythia-6.7b,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.42.1,,0.31.0,,,,1.20.0,,,,0.11.1,,,,MB,1265.577984,15104.212992,0.0,14457.765888,14237.630976,s,10,0.74734521484375,0.07473452148437501,0.00029993811276126254,0.07465035247802734,0.0747626350402832,0.07519656410217285,0.07554370735168457,"[0.0756304931640625, 0.07465331268310547, 0.07457756805419923, 0.07465213012695313, 0.07466620635986328, 0.07462438201904296, 0.07459468841552734, 0.07464857482910156, 0.07466403198242187, 0.07463382720947266]",tokens/s,3425.4584750840045,kWh,8.814724152954064e-07,4.83006719018346e-07,4.619448140000016e-06,5.983927274313769e-06,tokens/kWh,42781268.59911709,MB,1265.577984,15104.212992,0.0,14457.765888,14415.238144,s,10,21.24931665039062,2.124931665039062,0.0026433175539294805,2.1242171630859374,2.1261835205078126,2.129321984863281,2.1318327563476562,"[2.13246044921875, 2.124260986328125, 2.12389599609375, 2.125486083984375, 2.125090087890625, 2.122947021484375, 2.12264111328125, 2.12417333984375, 2.12465283203125, 2.123708740234375]",tokens/s,29.64801223329781,kWh,2.5081762022541136e-05,1.3745567048835965e-05,0.00011635381530519993,0.00015518114437657702,tokens/kWh,405977.15819853946,,s,629,21.887022132873522,0.03479653757213598,0.008873146792458899,0.03372236633300781,0.033831118011474606,0.0339038215637207,0.10839699768066406,"[0.03409408187866211, 0.03411558532714844, 0.03413913726806641, 0.034065406799316404, 0.03419340896606445, 0.03416985702514649, 0.03438387298583984, 0.03419136047363281, 0.03420467376708984, 0.034141185760498044, 0.034124801635742184, 0.03406950378417969, 0.034157569885253904, 0.033955841064453124, 0.033688575744628906, 0.03365478515625, 0.03366195297241211, 0.033633281707763675, 0.03362406539916992, 0.033683456420898435, 0.03373158264160156, 0.03363225555419922, 0.03369472122192383, 0.03368243026733399, 0.033699840545654294, 0.033710079193115236, 0.033713153839111325, 0.033710079193115236, 0.03369267272949219, 0.03372748947143555, 0.03367628860473633, 0.03373056030273437, 0.03372851181030274, 0.03376025772094727, 0.03374899291992187, 0.033729534149169925, 0.03378073501586914, 0.03370393753051758, 0.03376025772094727, 0.03370393753051758, 0.03378585433959961, 0.03374387359619141, 0.03374899291992187, 0.03376435089111328, 0.03376435089111328, 0.033718273162841796, 0.033756160736083986, 0.033797119140625, 0.033791999816894534, 0.03379609680175781, 0.033890304565429685, 0.03382476806640625, 0.033808383941650394, 0.033906688690185545, 0.03389952087402344, 0.03380633544921875, 0.03382271957397461, 0.03381760025024414, 0.033849342346191406, 0.033876991271972655, 0.033838081359863284, 0.0338524169921875, 0.10844671630859375, 0.033554431915283206, 0.033616897583007815, 0.033588222503662106, 0.03359743881225586, 0.03359436798095703, 0.03361177444458008, 0.03360563278198242, 0.03360768127441406, 0.03364044952392578, 0.033567745208740236, 0.03365785598754883, 0.03359027099609375, 0.03361382293701172, 0.03365478515625, 0.03365478515625, 0.03363840103149414, 0.03364147186279297, 0.03364352035522461, 0.03365785598754883, 0.033753089904785157, 0.0336363525390625, 0.03368447875976562, 0.03369267272949219, 0.03369779205322266, 0.033688575744628906, 0.03372748947143555, 0.03371724700927734, 0.03378176116943359, 0.03370598220825195, 0.03374694442749023, 0.033686527252197264, 0.0337367057800293, 0.033724414825439454, 0.03374899291992187, 0.03371110534667969, 0.033751041412353515, 0.03369676971435547, 0.033753089904785157, 0.033767425537109375, 0.03376639938354492, 0.033737728118896484, 0.03376435089111328, 0.03374694442749023, 0.03380326461791992, 0.033740798950195314, 0.03379404830932617, 0.03376947021484375, 0.033838081359863284, 0.03379916763305664, 0.03382374572753906, 0.03377459335327149, 0.0337786865234375, 0.03377664184570312, 0.03383193588256836, 0.03382271957397461, 0.03383091354370117, 0.03380326461791992, 0.033842174530029294, 0.03379097747802735, 0.03387187194824219, 0.033811454772949216, 0.03383705520629883, 0.10836582183837891, 0.03351551818847656, 0.03357593536376953, 0.03355033493041992, 0.033560577392578124, 0.0335810546875, 0.033604606628417966, 0.03357491302490234, 0.033598464965820314, 0.03362713623046875, 0.03364352035522461, 0.03361996841430664, 0.033585151672363284, 0.033633281707763675, 0.03363020706176758, 0.033658878326416015, 0.03359436798095703, 0.03365785598754883, 0.033617919921875, 0.03364863967895508, 0.03365785598754883, 0.03366912078857422, 0.03365273666381836, 0.03367116928100586, 0.03365785598754883, 0.03369267272949219, 0.033683456420898435, 0.033710079193115236, 0.033653759002685545, 0.03372236633300781, 0.033675262451171875, 0.03373567962646484, 0.033691646575927735, 0.033716224670410154, 0.033704959869384765, 0.03371724700927734, 0.033685504913330076, 0.033739776611328126, 0.03370700836181641, 0.03374796676635742, 0.03370700836181641, 0.03381043243408203, 0.03376947021484375, 0.03381350326538086, 0.033732608795166014, 0.03375718307495117, 0.033745918273925785, 0.033788928985595705, 0.03386880111694336, 0.0339159049987793, 0.03378790283203125, 0.03379814529418945, 0.03374899291992187, 0.03378790283203125, 0.03386470413208008, 0.03380428695678711, 0.03379097747802735, 0.03382476806640625, 0.034103294372558594, 0.03390771102905273, 0.03381043243408203, 0.03382886505126953, 0.03380428695678711, 0.10846617889404297, 0.03362611389160156, 0.033598464965820314, 0.03364556884765625, 0.03358924865722656, 0.03357798385620117, 0.03362508773803711, 0.0335912971496582, 0.033620990753173825, 0.03361177444458008, 0.0336445426940918, 0.03362815856933594, 0.03363430404663086, 0.033658878326416015, 0.03364556884765625, 0.03365273666381836, 0.03369062423706055, 0.03362713623046875, 0.03365683364868164, 0.03374694442749023, 0.03360153579711914, 0.03369881439208984, 0.0337520637512207, 0.03370086288452148, 0.03369267272949219, 0.0337448959350586, 0.033701889038085936, 0.033716224670410154, 0.03369472122192383, 0.03374387359619141, 0.033723392486572266, 0.033723392486572266, 0.03378073501586914, 0.033745918273925785, 0.03371212768554688, 0.03374387359619141, 0.03387801742553711, 0.03379916763305664, 0.03372748947143555, 0.03381043243408203, 0.033759231567382815, 0.03378278350830078, 0.03377971267700195, 0.033754112243652344, 0.033729534149169925, 0.03433779144287109, 0.03383295822143555, 0.03381862258911133, 0.033754112243652344, 0.03381350326538086, 0.03374899291992187, 0.03381964874267578, 0.033833984375, 0.03382476806640625, 0.03379097747802735, 0.03379609680175781, 0.033808383941650394, 0.03384524917602539, 0.03381760025024414, 0.033825790405273434, 0.03381248092651367, 0.03381964874267578, 0.033833984375, 0.10847539520263672, 0.03355648040771484, 0.033600513458251956, 0.033620990753173825, 0.03359436798095703, 0.03362918472290039, 0.03363430404663086, 0.033649665832519535, 0.033672191619873046, 0.033667072296142575, 0.033759231567382815, 0.03375001525878906, 0.03367628860473633, 0.0336814079284668, 0.033653759002685545, 0.03370700836181641, 0.03366400146484375, 0.03365683364868164, 0.03365785598754883, 0.03366912078857422, 0.033688575744628906, 0.03366809463500976, 0.0336814079284668, 0.033649665832519535, 0.033699840545654294, 0.03373875045776367, 0.033708030700683594, 0.033726463317871096, 0.03369574356079102, 0.033702911376953124, 0.03374694442749023, 0.03371212768554688, 0.03382271957397461, 0.03377151870727539, 0.03374899291992187, 0.033732608795166014, 0.033767425537109375, 0.033759231567382815, 0.033716224670410154, 0.03372544097900391, 0.03373567962646484, 0.03378176116943359, 0.033737728118896484, 0.03379916763305664, 0.033906688690185545, 0.03377971267700195, 0.033737728118896484, 0.03377766418457031, 0.03376844787597656, 0.0337786865234375, 0.03377664184570312, 0.03381452941894531, 0.03376639938354492, 0.03381657409667969, 0.03380428695678711, 0.03382476806640625, 0.03385651016235352, 0.03377664184570312, 0.033808383941650394, 0.03378585433959961, 0.033821697235107424, 0.03383603286743164, 0.03386777496337891, 0.10843033599853516, 0.03359027099609375, 0.033562625885009766, 0.033555454254150394, 0.03354009628295898, 0.033567745208740236, 0.033576961517333984, 0.033620990753173825, 0.0335912971496582, 0.03361894226074219, 0.03360870361328125, 0.03360358428955078, 0.03359231948852539, 0.033617919921875, 0.03362611389160156, 0.03361075210571289, 0.03362815856933594, 0.03364556884765625, 0.03375513458251953, 0.0336445426940918, 0.03363123321533203, 0.03376639938354492, 0.03362508773803711, 0.033655807495117186, 0.03367731094360352, 0.03366604614257813, 0.03366604614257813, 0.03367628860473633, 0.0336732177734375, 0.03367116928100586, 0.033693695068359376, 0.03369267272949219, 0.03365990447998047, 0.03367833709716797, 0.03372748947143555, 0.03371929550170898, 0.03368243026733399, 0.03371417617797851, 0.033723392486572266, 0.033756160736083986, 0.033721343994140625, 0.03376947021484375, 0.03374694442749023, 0.03376844787597656, 0.03378483200073242, 0.033772544860839845, 0.0337448959350586, 0.033770496368408204, 0.033770496368408204, 0.033745918273925785, 0.033791999816894534, 0.033788928985595705, 0.03377664184570312, 0.03380940628051758, 0.03374387359619141, 0.03376639938354492, 0.033797119140625, 0.03378073501586914, 0.03377356719970703, 0.03379916763305664, 0.03382271957397461, 0.03378483200073242, 0.033827838897705076, 0.10837709045410156, 0.033544193267822264, 0.03356671905517578, 0.03357183837890625, 0.033584129333496096, 0.03361382293701172, 0.033614849090576174, 0.03356364822387695, 0.033568767547607424, 0.03361587142944336, 0.03364863967895508, 0.03357593536376953, 0.03362406539916992, 0.03363020706176758, 0.03360358428955078, 0.03362918472290039, 0.03359539031982422, 0.033633281707763675, 0.0336629753112793, 0.03365683364868164, 0.03361280059814453, 0.03365683364868164, 0.033633281707763675, 0.033617919921875, 0.03364147186279297, 0.03365785598754883, 0.0336363525390625, 0.033691646575927735, 0.03377971267700195, 0.03369062423706055, 0.033688575744628906, 0.03367833709716797, 0.03369472122192383, 0.03369267272949219, 0.033723392486572266, 0.033688575744628906, 0.03370598220825195, 0.033713153839111325, 0.033693695068359376, 0.033753089904785157, 0.033701889038085936, 0.0337448959350586, 0.03371110534667969, 0.0337336311340332, 0.03369881439208984, 0.033753089904785157, 0.03370598220825195, 0.03377766418457031, 0.03377971267700195, 0.03376639938354492, 0.03374796676635742, 0.03376537704467773, 0.033887233734130856, 0.03378790283203125, 0.03377664184570312, 0.033805313110351565, 0.033783809661865234, 0.03381043243408203, 0.03378073501586914, 0.03382374572753906, 0.03379302215576172, 0.03381452941894531, 0.03379097747802735, 0.10845798492431641, 0.03358003234863281, 0.0335728645324707, 0.03356979370117188, 0.03358003234863281, 0.033582080841064454, 0.03359231948852539, 0.03360768127441406, 0.033598464965820314, 0.03363225555419922, 0.03359641647338867, 0.03364659118652344, 0.03360153579711914, 0.03364659118652344, 0.03368960189819336, 0.03361587142944336, 0.03366604614257813, 0.033616897583007815, 0.03366195297241211, 0.0336363525390625, 0.033718273162841796, 0.033683456420898435, 0.033637374877929685, 0.033649665832519535, 0.03362918472290039, 0.03367833709716797, 0.03369062423706055, 0.033691646575927735, 0.03371417617797851, 0.033821697235107424, 0.03371929550170898, 0.03366400146484375, 0.03370905685424805, 0.033718273162841796, 0.03371724700927734, 0.03372032165527344, 0.033805313110351565, 0.03378278350830078, 0.03374182510375977, 0.03377151870727539, 0.033770496368408204, 0.03376537704467773, 0.03378995132446289, 0.03376947021484375, 0.03376435089111328, 0.03376537704467773, 0.03381452941894531, 0.0337786865234375, 0.03378790283203125, 0.03379609680175781, 0.03381452941894531, 0.033805313110351565, 0.033808383941650394, 0.03376844787597656, 0.03384729766845703, 0.03378278350830078, 0.033805313110351565, 0.03380121612548828, 0.03382476806640625, 0.033842174530029294, 0.03384524917602539, 0.033825790405273434, 0.03386265563964844, 0.10851430511474609, 0.033568767547607424, 0.03362713623046875, 0.0335912971496582, 0.033582080841064454, 0.033675262451171875, 0.0336517105102539, 0.033723392486572266, 0.033616897583007815, 0.03363532638549805, 0.03360255813598633, 0.03364352035522461, 0.03362611389160156, 0.03361587142944336, 0.03364147186279297, 0.033655807495117186, 0.03377766418457031, 0.033675262451171875, 0.03363840103149414, 0.03363123321533203, 0.033672191619873046, 0.03364659118652344, 0.033685504913330076, 0.03365785598754883, 0.03374694442749023, 0.03364863967895508, 0.03370086288452148, 0.033683456420898435, 0.03371212768554688, 0.03371929550170898, 0.03368755340576172, 0.033723392486572266, 0.033670143127441404, 0.033702911376953124, 0.03367833709716797, 0.033718273162841796, 0.033691646575927735, 0.03371724700927734, 0.03372236633300781, 0.03375001525878906, 0.03373056030273437, 0.03372851181030274, 0.0337151985168457, 0.03386470413208008, 0.03377151870727539, 0.03375718307495117, 0.03395891189575195, 0.03401318359375, 0.033791999816894534, 0.033775615692138675, 0.03379916763305664, 0.03389132690429687, 0.033805313110351565, 0.033775615692138675, 0.03381657409667969, 0.033825790405273434, 0.03385343933105469, 0.03377971267700195, 0.03382886505126953, 0.033811454772949216, 0.03381657409667969, 0.03379404830932617, 0.03383193588256836, 0.1084047393798828, 0.03357388687133789, 0.0336517105102539, 0.033649665832519535, 0.033584129333496096, 0.033633281707763675, 0.0335912971496582, 0.0335994873046875, 0.0335810546875, 0.03362713623046875, 0.03360870361328125, 0.03362508773803711, 0.03364352035522461, 0.033616897583007815, 0.03363225555419922, 0.034065406799316404, 0.033724414825439454, 0.033670143127441404, 0.03362815856933594, 0.03366604614257813, 0.03361280059814453, 0.0336445426940918, 0.033675262451171875, 0.03366604614257813, 0.033653759002685545, 0.033658878326416015, 0.03365273666381836, 0.03370393753051758, 0.033675262451171875, 0.033691646575927735, 0.03367116928100586, 0.03374694442749023, 0.033685504913330076, 0.033710079193115236, 0.03370598220825195, 0.03371724700927734, 0.033737728118896484, 0.03367628860473633, 0.033701889038085936, 0.033754112243652344, 0.03375513458251953, 0.03372851181030274, 0.03369676971435547, 0.03378073501586914, 0.03374796676635742, 0.03374387359619141, 0.033729534149169925, 0.03380121612548828, 0.03378483200073242, 0.033745918273925785, 0.03377356719970703, 0.03378176116943359, 0.03378278350830078, 0.03376128005981445, 0.033753089904785157, 0.03381452941894531, 0.0338155517578125, 0.03378176116943359, 0.03379814529418945, 0.03383603286743164, 0.03382067108154297, 0.033827838897705076, 0.03379302215576172]",tokens/s,28.738491521661338,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-13b,facebook/opt-13b,cuda,0,42,,,True,,,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",opt,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-6.7b,facebook/opt-6.7b,cuda,0,42,,,True,,,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",opt,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,microsoft/rho-math-1b-v0.1,microsoft/rho-math-1b-v0.1,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,MB,1258.74176,3041.394688,0.0,2394.947584,2279.417856,s,10,0.24129708862304688,0.02412970886230469,0.0009275755891401461,0.024045184135437013,0.024764550018310547,0.025596467399597166,0.026262001304626463,"[0.026428384780883787, 0.023323360443115233, 0.02338118362426758, 0.024444223403930664, 0.023216384887695313, 0.023652992248535155, 0.024437376022338867, 0.023386943817138673, 0.02444655990600586, 0.024579679489135742]",tokens/s,10609.328171378062,kWh,2.829735362897333e-07,1.550558548124651e-07,1.0543529168468965e-06,1.4923823079490949e-06,tokens/kWh,171537814.8323185,MB,1259.06944,3041.394688,0.0,2394.947584,2279.420416,s,10,14.706535156249998,1.470653515625,0.02462968819862708,1.4804291381835937,1.4855615844726564,1.4865984191894532,1.4874278869628907,"[1.440400390625, 1.478913818359375, 1.4812783203125, 1.4824281005859374, 1.4075567626953125, 1.4783734130859374, 1.4850379638671876, 1.4795799560546874, 1.48763525390625, 1.4853311767578126]",tokens/s,42.83809839003865,kWh,1.6815152510897653e-05,9.214636577634288e-06,3.707637203215265e-05,6.310616112068459e-05,tokens/kWh,998317.7376218217,,s,629,14.8957398891449,0.023681621445381395,0.002902418586720406,0.023476224899291992,0.023700889205932617,0.023891558837890625,0.04635525131225587,"[0.02434764862060547, 0.02344550323486328, 0.02329702377319336, 0.0230830078125, 0.023096319198608398, 0.023059455871582032, 0.022979583740234375, 0.023313440322875977, 0.023432159423828126, 0.023243776321411135, 0.023096319198608398, 0.02310860824584961, 0.023026687622070312, 0.02312499237060547, 0.023179264068603517, 0.023175167083740233, 0.02310860824584961, 0.023164928436279295, 0.02305843162536621, 0.023142400741577147, 0.023002111434936523, 0.023181312561035155, 0.023236608505249022, 0.023367679595947266, 0.023235584259033205, 0.023150592803955077, 0.022930431365966796, 0.022755327224731444, 0.022408191680908202, 0.022520832061767578, 0.022561792373657227, 0.022385663986206054, 0.02243174362182617, 0.022386688232421875, 0.022478847503662108, 0.022436864852905275, 0.02246451187133789, 0.02249625587463379, 0.022394880294799805, 0.02229248046875, 0.022417407989501953, 0.022486015319824217, 0.02246451187133789, 0.02243174362182617, 0.022368255615234374, 0.022372352600097657, 0.022329343795776366, 0.022417407989501953, 0.022475807189941407, 0.02237129592895508, 0.022436864852905275, 0.022335487365722655, 0.022540288925170897, 0.022395904541015626, 0.02244095993041992, 0.02246143913269043, 0.022441984176635742, 0.022420480728149415, 0.022418432235717774, 0.02227712059020996, 0.024041471481323243, 0.02351513671875, 0.04617830276489258, 0.022366207122802736, 0.022373376846313478, 0.02330624008178711, 0.023311359405517578, 0.023335935592651368, 0.02347417640686035, 0.023459840774536132, 0.023857152938842774, 0.023444480895996093, 0.02352742385864258, 0.023427072525024413, 0.023436288833618164, 0.02342911911010742, 0.02350492858886719, 0.023503839492797853, 0.02346598434448242, 0.023595008850097656, 0.023435264587402343, 0.02350284767150879, 0.023427072525024413, 0.0234967041015625, 0.023796735763549806, 0.023385087966918947, 0.023446527481079102, 0.023417856216430662, 0.023431167602539063, 0.023413759231567383, 0.023549951553344727, 0.02350182342529297, 0.02353561592102051, 0.02350284767150879, 0.02349158477783203, 0.02343731117248535, 0.023558143615722657, 0.023428096771240234, 0.023537664413452147, 0.023553024291992186, 0.023571456909179687, 0.023558208465576172, 0.023587776184082032, 0.023434240341186522, 0.023508991241455078, 0.02345062446594238, 0.023421951293945312, 0.023483455657958983, 0.02340447998046875, 0.023491615295410155, 0.02347007942199707, 0.02358268737792969, 0.02369843292236328, 0.023563264846801758, 0.02350284767150879, 0.02349772834777832, 0.023409664154052736, 0.023464960098266603, 0.02341891288757324, 0.02349769592285156, 0.023417856216430662, 0.02345369529724121, 0.023545856475830077, 0.025411584854125976, 0.023580671310424805, 0.04688588714599609, 0.023653375625610353, 0.02348748779296875, 0.02351513671875, 0.023414783477783203, 0.02351820755004883, 0.023434240341186522, 0.02351103973388672, 0.023517183303833008, 0.02350387191772461, 0.023516159057617187, 0.023946239471435548, 0.02411315155029297, 0.02370047950744629, 0.023623680114746092, 0.023617536544799804, 0.02349567985534668, 0.023399423599243165, 0.02352025604248047, 0.023548927307128906, 0.023592960357666014, 0.02352332878112793, 0.023415807723999024, 0.02349772834777832, 0.02367283248901367, 0.023411712646484374, 0.02349158477783203, 0.023433216094970705, 0.023545856475830077, 0.02345267105102539, 0.023777280807495117, 0.02349465560913086, 0.02346291160583496, 0.0234833927154541, 0.023593984603881835, 0.02346291160583496, 0.02348953628540039, 0.023497760772705076, 0.023456735610961912, 0.02349875259399414, 0.02350182342529297, 0.023412736892700195, 0.02332979202270508, 0.02346291160583496, 0.023366655349731445, 0.023409664154052736, 0.02349158477783203, 0.023555072784423828, 0.02343731117248535, 0.023588863372802735, 0.023418880462646483, 0.02346598434448242, 0.023441408157348635, 0.02347417640686035, 0.023414783477783203, 0.023417856216430662, 0.023431167602539063, 0.023409664154052736, 0.02345369529724121, 0.023504896163940428, 0.023446527481079102, 0.023619583129882812, 0.02346086311340332, 0.048465919494628903, 0.023447551727294923, 0.023423999786376954, 0.023398399353027344, 0.02346598434448242, 0.023441408157348635, 0.02346086311340332, 0.023636991500854493, 0.02389606475830078, 0.023550975799560548, 0.023388160705566406, 0.023459840774536132, 0.023406591415405274, 0.023541759490966797, 0.02350284767150879, 0.023588863372802735, 0.02349875259399414, 0.02368716812133789, 0.023344127655029297, 0.02326323127746582, 0.023436288833618164, 0.023627775192260742, 0.02348646354675293, 0.02347110366821289, 0.02352025604248047, 0.023488512039184572, 0.023464960098266603, 0.02353561592102051, 0.02347110366821289, 0.02345267105102539, 0.023571456909179687, 0.023532543182373047, 0.02348236846923828, 0.023518272399902344, 0.023555007934570313, 0.023534656524658203, 0.023495616912841796, 0.02352639961242676, 0.02437222480773926, 0.023982080459594726, 0.02350387191772461, 0.023568384170532225, 0.02345881652832031, 0.02347724723815918, 0.023412736892700195, 0.0276889591217041, 0.024598527908325195, 0.02370355224609375, 0.023488512039184572, 0.02346598434448242, 0.02390019226074219, 0.023435232162475585, 0.023396352767944335, 0.02349465560913086, 0.023434240341186522, 0.023488512039184572, 0.023359487533569336, 0.023178239822387696, 0.023190528869628906, 0.02229964828491211, 0.022338560104370117, 0.02234060859680176, 0.02230271911621094, 0.04605644989013672, 0.02233241653442383, 0.022225984573364256, 0.022307775497436524, 0.022171648025512695, 0.02230681610107422, 0.02225971221923828, 0.0222873592376709, 0.022227968215942383, 0.02226790428161621, 0.022228992462158204, 0.022231039047241212, 0.022312959671020507, 0.02224332809448242, 0.02230067253112793, 0.022289407730102538, 0.022239231109619142, 0.022339584350585938, 0.02225766372680664, 0.02225868797302246, 0.022335487365722655, 0.0223191032409668, 0.022361087799072265, 0.022244352340698242, 0.02232524871826172, 0.022213632583618165, 0.02226688003540039, 0.022252544403076172, 0.022346752166748047, 0.0222423038482666, 0.022611967086791994, 0.022323200225830078, 0.022370304107666016, 0.02239286422729492, 0.022396928787231447, 0.022254560470581056, 0.022387712478637696, 0.02235699272155762, 0.022386688232421875, 0.022377471923828125, 0.022388736724853517, 0.022452224731445314, 0.02230886459350586, 0.022366207122802736, 0.022338560104370117, 0.02245631980895996, 0.022508544921875, 0.02230784034729004, 0.022326303482055665, 0.022441951751708985, 0.022213632583618165, 0.022360063552856444, 0.022348831176757813, 0.022378463745117188, 0.0223191032409668, 0.022468608856201173, 0.022352895736694335, 0.02246143913269043, 0.022387712478637696, 0.0224532470703125, 0.022494207382202147, 0.022375423431396483, 0.022380544662475587, 0.04667084884643555, 0.02244095993041992, 0.022312959671020507, 0.022444032669067384, 0.023698463439941406, 0.023480287551879885, 0.023618560791015625, 0.02349977684020996, 0.023416831970214845, 0.023427072525024413, 0.0235100154876709, 0.023860223770141603, 0.023397375106811523, 0.0234833927154541, 0.02345062446594238, 0.02348646354675293, 0.023980031967163085, 0.02409779167175293, 0.023796735763549806, 0.023441408157348635, 0.023567359924316408, 0.023451648712158202, 0.02351513671875, 0.02350182342529297, 0.023413759231567383, 0.02346291160583496, 0.02342911911010742, 0.023498783111572264, 0.023575551986694337, 0.023567327499389647, 0.02368307113647461, 0.02352742385864258, 0.02345062446594238, 0.023407615661621094, 0.0235284481048584, 0.02348543930053711, 0.02351820755004883, 0.02348543930053711, 0.023623680114746092, 0.023556095123291015, 0.023493696212768554, 0.02359187126159668, 0.023408672332763673, 0.023400415420532228, 0.02343731117248535, 0.023372800827026367, 0.023372800827026367, 0.02353459167480469, 0.02346188735961914, 0.023439359664916993, 0.02348236846923828, 0.02348134422302246, 0.023513088226318358, 0.023784511566162108, 0.02346182441711426, 0.023446527481079102, 0.02344550323486328, 0.023601152420043944, 0.02346188735961914, 0.023602176666259765, 0.023524351119995117, 0.02347417640686035, 0.02349260711669922, 0.04845363235473633, 0.02351206398010254, 0.02345779228210449, 0.023384063720703126, 0.023342079162597656, 0.023385087966918947, 0.023463935852050782, 0.023431167602539063, 0.023386112213134767, 0.023566335678100587, 0.02343734359741211, 0.02350588798522949, 0.023451648712158202, 0.023399423599243165, 0.023629823684692384, 0.0235100154876709, 0.023439359664916993, 0.02349363136291504, 0.02352332878112793, 0.023402496337890624, 0.023553024291992186, 0.02395955276489258, 0.024360960006713867, 0.023872512817382813, 0.023620607376098633, 0.023513088226318358, 0.023617536544799804, 0.023734272003173826, 0.02376192092895508, 0.023395328521728515, 0.023517183303833008, 0.023559167861938478, 0.023826431274414063, 0.02346700859069824, 0.023630847930908205, 0.02349977684020996, 0.023442432403564452, 0.02354380798339844, 0.023451648712158202, 0.023418880462646483, 0.023521280288696288, 0.02349056053161621, 0.023830528259277343, 0.023536640167236327, 0.023587839126586914, 0.023843839645385743, 0.023639072418212892, 0.02358268737792969, 0.0235284481048584, 0.023645183563232423, 0.023635967254638672, 0.023714815139770508, 0.023757823944091795, 0.023448575973510744, 0.023463935852050782, 0.02345881652832031, 0.02348134422302246, 0.02352230453491211, 0.023636991500854493, 0.023618560791015625, 0.023581695556640626, 0.023576576232910155, 0.023478271484375, 0.04642406463623047, 0.0224901123046875, 0.023410688400268553, 0.023560192108154295, 0.023390207290649414, 0.023657472610473632, 0.023405567169189453, 0.023468032836914062, 0.02347007942199707, 0.02368921661376953, 0.023809024810791016, 0.02349363136291504, 0.02349567985534668, 0.02352230453491211, 0.023610368728637695, 0.023553024291992186, 0.02350182342529297, 0.02358892822265625, 0.023443391799926758, 0.02346086311340332, 0.022725631713867187, 0.02353152084350586, 0.023367679595947266, 0.023666688919067383, 0.02392985534667969, 0.023682048797607422, 0.023432191848754884, 0.023590911865234376, 0.023605247497558594, 0.023430143356323242, 0.02353971290588379, 0.02349465560913086, 0.023373823165893554, 0.02345372772216797, 0.023540704727172852, 0.02364723205566406, 0.023456768035888673, 0.023607295989990236, 0.023524351119995117, 0.023706623077392578, 0.023500799179077148, 0.02388479995727539, 0.023567359924316408, 0.02343731117248535, 0.02364825630187988, 0.023368703842163087, 0.02347417640686035, 0.02346188735961914, 0.02342095947265625, 0.023388128280639648, 0.023388160705566406, 0.023472127914428712, 0.023413759231567383, 0.02392166328430176, 0.02345267105102539, 0.023398399353027344, 0.023459840774536132, 0.023386112213134767, 0.023517183303833008, 0.02346598434448242, 0.02341993522644043, 0.02337174415588379, 0.023434240341186522, 0.04841471862792969, 0.023625728607177734, 0.023516159057617187, 0.023827455520629884, 0.023547903060913086, 0.023661567687988282, 0.023480319976806642, 0.023603200912475586, 0.023702527999877928, 0.023559167861938478, 0.023617536544799804, 0.02359609603881836, 0.02362668800354004, 0.023773183822631837, 0.023508991241455078, 0.023589887619018556, 0.023542783737182618, 0.023610368728637695, 0.02350796890258789, 0.023500799179077148, 0.023654399871826173, 0.023577600479125976, 0.023669759750366212, 0.023431167602539063, 0.023416831970214845, 0.023435264587402343, 0.02353561592102051, 0.023472127914428712, 0.023508991241455078, 0.023610368728637695, 0.023349248886108398, 0.023838720321655273, 0.023574527740478517, 0.023665664672851562, 0.023545856475830077, 0.02347007942199707, 0.02351923179626465, 0.02348441505432129, 0.023556095123291015, 0.023553024291992186, 0.023809024810791016, 0.023447551727294923, 0.02365132713317871, 0.02346291160583496, 0.023472127914428712, 0.023814144134521483, 0.023644159317016602, 0.02351103973388672, 0.023439359664916993, 0.02353561592102051, 0.023576576232910155, 0.02348134422302246, 0.023403520584106444, 0.02349260711669922, 0.023463935852050782, 0.024232959747314452, 0.024779775619506835, 0.0238786563873291, 0.023740415573120118, 0.02353561592102051, 0.023565311431884766, 0.023536640167236327, 0.02392780876159668, 0.048901119232177735, 0.02348134422302246, 0.02345267105102539, 0.02350284767150879, 0.023435264587402343, 0.02349567985534668, 0.02346086311340332, 0.023742528915405275, 0.023418815612792968, 0.023916608810424806, 0.02359391975402832, 0.023771135330200196, 0.023630847930908205, 0.02353049659729004, 0.023763967514038087, 0.02369638442993164, 0.023540735244750977, 0.02351206398010254, 0.02410086441040039, 0.0234967041015625, 0.023630847930908205, 0.02355200004577637, 0.02347520065307617, 0.023739456176757812, 0.02362668800354004, 0.023443456649780273, 0.02349772834777832, 0.023350271224975586, 0.02365132713317871, 0.023480319976806642, 0.02349977684020996, 0.023561216354370116, 0.023613439559936524, 0.023544832229614256, 0.02345779228210449, 0.023640064239501952, 0.02358787155151367, 0.023503839492797853, 0.023582719802856447, 0.023504896163940428, 0.023393280029296876, 0.023476224899291992, 0.02352230453491211, 0.02345267105102539, 0.023410688400268553, 0.02344550323486328, 0.023578624725341796, 0.023750656127929686, 0.02352025604248047, 0.023536640167236327, 0.023419904708862304, 0.024018943786621092, 0.02368409538269043, 0.023427072525024413, 0.023582719802856447, 0.023547935485839843, 0.02356937599182129, 0.023626752853393555, 0.023521280288696288, 0.0234967041015625, 0.02345267105102539, 0.023432191848754884, 0.02369126319885254]",tokens/s,42.22683832297425,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-2.7b,EleutherAI/pythia-2.7b,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.42.1,,0.31.0,,,,1.20.0,,,,0.11.1,,,,MB,1285.603328,6755.45088,0.0,6109.003776,5988.3136,s,10,0.304234079360962,0.030423407936096193,0.00021981722519289707,0.030311151504516604,0.030751270675659182,0.030845763397216796,0.03092135757446289,"[0.030940256118774413, 0.0303287353515625, 0.030730272293090822, 0.030293567657470703, 0.030261632919311523, 0.030276447296142577, 0.03034137535095215, 0.03048819160461426, 0.030281375885009766, 0.030292224884033204]",tokens/s,8414.573427727863,kWh,3.575588831288497e-07,1.9592575379471176e-07,1.8344073084108751e-06,2.3878919453344365e-06,tokens/kWh,107207531.10298125,MB,1285.603328,6755.45088,0.0,6109.003776,6091.200512,s,10,12.9281962890625,1.29281962890625,0.006292437642451194,1.2917204589843752,1.301793408203125,1.3036004394531249,1.305046064453125,"[1.301391845703125, 1.2949488525390624, 1.293156494140625, 1.2854097900390624, 1.2844527587890624, 1.2898765869140625, 1.288757080078125, 1.305407470703125, 1.294510986328125, 1.290284423828125]",tokens/s,48.73069575320356,kWh,1.526683735493327e-05,8.365995351810453e-06,5.5276805047189126e-05,7.890963775393286e-05,tokens/kWh,798381.5639409658,,s,629,13.180049415588377,0.020953973633685816,0.0036592025841014433,0.02044313621520996,0.020877721405029297,0.02112450523376465,0.0510394369506836,"[0.021996543884277343, 0.02128179168701172, 0.021211135864257814, 0.021158912658691405, 0.020793344497680662, 0.020900863647460938, 0.020810752868652343, 0.02060697555541992, 0.020583423614501953, 0.020700159072875975, 0.020543487548828124, 0.02046976089477539, 0.020653055191040038, 0.02090598487854004, 0.02087936019897461, 0.020797439575195312, 0.02120806312561035, 0.020753408432006838, 0.020523008346557618, 0.020814847946166993, 0.02046771240234375, 0.020220928192138672, 0.02032640075683594, 0.020937728881835937, 0.021797887802124022, 0.021562368392944335, 0.02069811248779297, 0.020296703338623046, 0.020335615158081053, 0.02032640075683594, 0.020444160461425782, 0.020541439056396483, 0.020347904205322266, 0.020378623962402344, 0.020519935607910156, 0.020468736648559572, 0.020323328018188477, 0.02047385597229004, 0.02044825553894043, 0.020649984359741212, 0.020405248641967775, 0.020229120254516602, 0.020289535522460937, 0.020189184188842774, 0.020254720687866212, 0.02026188850402832, 0.020197376251220703, 0.020291584014892578, 0.020478975296020507, 0.020398080825805662, 0.02027212715148926, 0.020287488937377928, 0.020289535522460937, 0.02030899238586426, 0.02106368064880371, 0.020993024826049804, 0.02165247917175293, 0.02048307228088379, 0.020321279525756835, 0.020525056838989256, 0.020612096786499022, 0.02068992042541504, 0.05112627029418945, 0.02044927978515625, 0.02044211196899414, 0.02172211265563965, 0.020562944412231447, 0.020313087463378905, 0.02027622413635254, 0.020281343460083007, 0.02044927978515625, 0.020594688415527345, 0.020497407913208008, 0.02024345588684082, 0.020274175643920898, 0.02043289566040039, 0.020571168899536134, 0.020465631484985352, 0.020454399108886717, 0.020486143112182616, 0.02048204803466797, 0.020518911361694335, 0.02043084716796875, 0.020533248901367186, 0.02047590446472168, 0.020454399108886717, 0.020485120773315428, 0.020576255798339844, 0.020192256927490236, 0.020426752090454102, 0.02044313621520996, 0.020406272888183592, 0.020397056579589845, 0.020426752090454102, 0.020336639404296874, 0.020414464950561522, 0.02044211196899414, 0.02040115165710449, 0.020487167358398437, 0.02046668815612793, 0.020340736389160157, 0.021433343887329103, 0.020528127670288086, 0.02019430351257324, 0.020344831466674804, 0.020552703857421875, 0.020364288330078126, 0.020418560028076172, 0.020402175903320312, 0.020413440704345705, 0.020479999542236327, 0.020327423095703127, 0.020183040618896485, 0.020208639144897463, 0.02023219108581543, 0.020197376251220703, 0.020739072799682616, 0.02023526382446289, 0.02024345588684082, 0.020191232681274415, 0.02020249557495117, 0.020220928192138672, 0.0212490234375, 0.024069120407104492, 0.022716415405273437, 0.05203353500366211, 0.02046156883239746, 0.020345855712890625, 0.020220928192138672, 0.020281343460083007, 0.02023628807067871, 0.020321279525756835, 0.020254720687866212, 0.020200447082519533, 0.020175872802734376, 0.020212736129760742, 0.020331520080566406, 0.02027008056640625, 0.020288511276245116, 0.020363264083862305, 0.02038374328613281, 0.020246528625488282, 0.02031820869445801, 0.020738048553466795, 0.020809728622436522, 0.02062848091125488, 0.020377599716186523, 0.020436992645263673, 0.021507072448730468, 0.020699136734008788, 0.020435968399047853, 0.020509695053100584, 0.02045747184753418, 0.02049843215942383, 0.020271104812622072, 0.020839424133300782, 0.020756479263305663, 0.02102374458312988, 0.02165555191040039, 0.021354496002197267, 0.020984832763671874, 0.02100223922729492, 0.020907007217407226, 0.020580352783203124, 0.02063564872741699, 0.0209039363861084, 0.020373504638671876, 0.020247552871704103, 0.02018611145019531, 0.020741119384765624, 0.020571136474609376, 0.02043084716796875, 0.020521984100341797, 0.020571136474609376, 0.02033459281921387, 0.020699136734008788, 0.020370431900024414, 0.02040729522705078, 0.02031718444824219, 0.020231168746948244, 0.020412416458129884, 0.02049126434326172, 0.020575231552124023, 0.020444160461425782, 0.020444160461425782, 0.020351999282836913, 0.02030284881591797, 0.020320255279541014, 0.050767871856689455, 0.020215808868408205, 0.02047283172607422, 0.020323328018188477, 0.020426752090454102, 0.020256767272949217, 0.02021785545349121, 0.021115903854370118, 0.0206878719329834, 0.02087936019897461, 0.020756479263305663, 0.02020147132873535, 0.020122623443603514, 0.020262912750244142, 0.020207616806030275, 0.02011136054992676, 0.020145151138305666, 0.02024550437927246, 0.020175872802734376, 0.020174848556518556, 0.020168703079223634, 0.020182016372680665, 0.02022400093078613, 0.020312063217163084, 0.02045952033996582, 0.020405248641967775, 0.020427776336669923, 0.02040934371948242, 0.020444160461425782, 0.020405248641967775, 0.02045747184753418, 0.02044211196899414, 0.020397056579589845, 0.020384767532348632, 0.02042163276672363, 0.02024550437927246, 0.020644863128662108, 0.020410367965698242, 0.0204400634765625, 0.020420608520507814, 0.02045747184753418, 0.020505599975585938, 0.020502527236938475, 0.020577280044555665, 0.020536319732666015, 0.02042572784423828, 0.020455423355102538, 0.020471807479858398, 0.020420608520507814, 0.02048307228088379, 0.020434944152832032, 0.020487167358398437, 0.020510719299316405, 0.02046566390991211, 0.020462591171264647, 0.02026905632019043, 0.020468736648559572, 0.02047590446472168, 0.020494335174560546, 0.02041548728942871, 0.02042265510559082, 0.020256767272949217, 0.020265983581542968, 0.05107097625732422, 0.020489215850830078, 0.02046668815612793, 0.020529151916503906, 0.0204769287109375, 0.020273151397705077, 0.020262912750244142, 0.0202608642578125, 0.02020966339111328, 0.020504575729370117, 0.020478975296020507, 0.0204769287109375, 0.020281343460083007, 0.020402175903320312, 0.0204400634765625, 0.020338687896728515, 0.02020351982116699, 0.020335615158081053, 0.02027827262878418, 0.020396032333374024, 0.0204083194732666, 0.020446207046508787, 0.020323328018188477, 0.02036735916137695, 0.020366336822509764, 0.020156415939331054, 0.020167680740356447, 0.020333568572998048, 0.02021887969970703, 0.02045030403137207, 0.020265983581542968, 0.020238336563110353, 0.020222976684570314, 0.020403200149536133, 0.020526079177856444, 0.020389888763427736, 0.020567039489746093, 0.02045030403137207, 0.020471807479858398, 0.020404224395751954, 0.02043391990661621, 0.020403200149536133, 0.02046771240234375, 0.020426752090454102, 0.02033459281921387, 0.020446207046508787, 0.020338687896728515, 0.02050048065185547, 0.020371456146240235, 0.02042982482910156, 0.020501504898071288, 0.02041548728942871, 0.020666368484497072, 0.02042265510559082, 0.020358144760131838, 0.020374528884887694, 0.020378623962402344, 0.020488191604614257, 0.02040729522705078, 0.02043084716796875, 0.020214784622192384, 0.020254720687866212, 0.020323328018188477, 0.051095550537109374, 0.020516864776611327, 0.02038374328613281, 0.020361215591430663, 0.02043289566040039, 0.020363264083862305, 0.02043801689147949, 0.020452352523803712, 0.020454399108886717, 0.020497407913208008, 0.020454399108886717, 0.02040729522705078, 0.02046668815612793, 0.020412416458129884, 0.02036735916137695, 0.020396032333374024, 0.020566015243530272, 0.020257791519165038, 0.020403200149536133, 0.020477951049804686, 0.020658176422119142, 0.020555776596069338, 0.020396032333374024, 0.020419584274291993, 0.020546560287475587, 0.02042265510559082, 0.020463615417480468, 0.02045337677001953, 0.020508672714233397, 0.020402175903320312, 0.020445184707641603, 0.020406272888183592, 0.02069606399536133, 0.02070425605773926, 0.020761600494384767, 0.020676607131958007, 0.020460544586181642, 0.02041651153564453, 0.020462591171264647, 0.020411392211914063, 0.020470815658569334, 0.020429792404174803, 0.020561920166015626, 0.020824064254760744, 0.02046668815612793, 0.020612096786499022, 0.0204083194732666, 0.020446207046508787, 0.020410367965698242, 0.02044927978515625, 0.020405248641967775, 0.020426752090454102, 0.020569087982177735, 0.02040934371948242, 0.02044927978515625, 0.020413440704345705, 0.02040115165710449, 0.020427776336669923, 0.020428800582885744, 0.020392959594726562, 0.020452352523803712, 0.02048307228088379, 0.020384767532348632, 0.05111808013916016, 0.020410367965698242, 0.020377599716186523, 0.020402175903320312, 0.020427776336669923, 0.02052403259277344, 0.02046774482727051, 0.020470752716064453, 0.02021785545349121, 0.020207616806030275, 0.02041753578186035, 0.02106675148010254, 0.02041753578186035, 0.0204769287109375, 0.020397056579589845, 0.020222976684570314, 0.02043903923034668, 0.020445184707641603, 0.020327423095703127, 0.02045030403137207, 0.020426752090454102, 0.02046463966369629, 0.02046156883239746, 0.020387840270996094, 0.02045747184753418, 0.020426752090454102, 0.02045849609375, 0.02048102378845215, 0.020434944152832032, 0.02040729522705078, 0.0204769287109375, 0.02042265510559082, 0.02062745666503906, 0.02041548728942871, 0.020525056838989256, 0.020370431900024414, 0.02044313621520996, 0.02043903923034668, 0.02045030403137207, 0.020452352523803712, 0.02046668815612793, 0.02043187141418457, 0.02047488021850586, 0.02032640075683594, 0.020305919647216796, 0.0204769287109375, 0.020428800582885744, 0.020295679092407228, 0.02084351921081543, 0.02069196891784668, 0.020479999542236327, 0.020288511276245116, 0.02043801689147949, 0.020414464950561522, 0.020676607131958007, 0.020714496612548827, 0.020333568572998048, 0.020462591171264647, 0.02039910316467285, 0.020427776336669923, 0.020539392471313478, 0.020463615417480468, 0.020436992645263673, 0.05120409774780273, 0.020888576507568358, 0.02106675148010254, 0.020960256576538085, 0.020974592208862306, 0.020875263214111327, 0.02085990333557129, 0.020567039489746093, 0.020447231292724608, 0.0206059513092041, 0.02042470359802246, 0.020564992904663085, 0.0206059513092041, 0.020774911880493165, 0.020382720947265624, 0.02031820869445801, 0.020331520080566406, 0.02031001663208008, 0.02042163276672363, 0.02040934371948242, 0.02044825553894043, 0.020586496353149415, 0.020530176162719727, 0.020566015243530272, 0.020364288330078126, 0.020717567443847656, 0.020793344497680662, 0.021994495391845705, 0.022196224212646484, 0.021148672103881837, 0.020748287200927733, 0.020563968658447264, 0.020732927322387695, 0.020520959854125977, 0.020727807998657227, 0.020403200149536133, 0.02050764846801758, 0.020579328536987306, 0.020746240615844725, 0.02048409652709961, 0.02050048065185547, 0.02040934371948242, 0.020554784774780274, 0.020790239334106446, 0.021114879608154297, 0.02082815933227539, 0.02089779281616211, 0.02087731170654297, 0.021013504028320314, 0.02083430480957031, 0.02086092758178711, 0.020908031463623047, 0.020989952087402345, 0.020787200927734374, 0.02064076805114746, 0.02059878349304199, 0.020707328796386718, 0.020517887115478514, 0.020858879089355468, 0.02086502456665039, 0.020766719818115235, 0.02067353630065918, 0.020564992904663085, 0.05146623992919922, 0.02089779281616211, 0.020932607650756836, 0.02084659194946289, 0.020799488067626954, 0.02103091239929199, 0.021130239486694336, 0.02087936019897461, 0.020788223266601562, 0.020739072799682616, 0.020909055709838868, 0.02089779281616211, 0.021029888153076173, 0.02083020782470703, 0.020728832244873048, 0.02024140739440918, 0.02039091110229492, 0.02022809600830078, 0.020312063217163084, 0.020291584014892578, 0.020330495834350586, 0.020190208435058594, 0.020322303771972656, 0.020150272369384766, 0.02022707176208496, 0.02031820869445801, 0.02043391990661621, 0.02043391990661621, 0.02044108772277832, 0.020256767272949217, 0.02042163276672363, 0.020513792037963868, 0.02036735916137695, 0.020517887115478514, 0.02048307228088379, 0.02041753578186035, 0.020337663650512695, 0.020387840270996094, 0.020351999282836913, 0.020546560287475587, 0.021292032241821288, 0.02064691162109375, 0.02045747184753418, 0.02106265640258789, 0.020787200927734374, 0.02040115165710449, 0.020542463302612304, 0.02045132827758789, 0.020404224395751954, 0.020539392471313478, 0.020392959594726562, 0.020414464950561522, 0.02042982482910156, 0.02042470359802246, 0.020541439056396483, 0.020405248641967775, 0.020327423095703127, 0.02038374328613281, 0.020411392211914063, 0.020387840270996094, 0.02048204803466797, 0.020554752349853517, 0.020273151397705077, 0.050958335876464846, 0.020487167358398437, 0.020273151397705077, 0.020143104553222657, 0.020171775817871093, 0.020162559509277343, 0.020336639404296874, 0.020364288330078126, 0.020456480026245116, 0.02031510353088379, 0.020340736389160157, 0.020410367965698242, 0.02043801689147949, 0.02035916709899902, 0.020392959594726562, 0.020342784881591795, 0.020392959594726562, 0.020420608520507814, 0.02039091110229492, 0.020351999282836913, 0.020446207046508787, 0.02041651153564453, 0.020371456146240235, 0.02045337677001953, 0.020545536041259766, 0.020426752090454102, 0.020392959594726562, 0.020412416458129884, 0.02040012741088867, 0.02045747184753418, 0.02047283172607422, 0.02045132827758789, 0.020463615417480468, 0.020427776336669923, 0.021992448806762696, 0.022226943969726562, 0.02086604881286621, 0.020485120773315428, 0.0204769287109375, 0.020494335174560546, 0.020447231292724608, 0.02043187141418457, 0.020540416717529295, 0.02043187141418457, 0.02049945640563965, 0.02045747184753418, 0.020413440704345705, 0.020412416458129884, 0.020568063735961914, 0.020446207046508787, 0.02039193534851074, 0.020323328018188477, 0.020411392211914063, 0.02042367935180664, 0.0202926082611084, 0.020758527755737305, 0.0204400634765625, 0.020494335174560546, 0.020314111709594726, 0.020402175903320312, 0.020516864776611327, 0.020753408432006838, 0.020246528625488282]",tokens/s,47.723645046130535,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen2-beta-72B,Qwen/Qwen2-beta-72B,cuda,0,42,,,True,,,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",qwen2,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-0.5B,,cuda,0,42,,,,,True,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,217063f5c507ed7cc255df7e1f64c4333a0b4dfe,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,,MB,1298.550784,2227.699712,0.0,1581.252608,1426.269184,s,10,0.20156473731994629,0.02015647373199463,0.0008666305191518691,0.019835792541503906,0.02048620777130127,0.02160451211929321,0.022499155597686768,"[0.022722816467285155, 0.019900352478027343, 0.01988812828063965, 0.020237695693969725, 0.019778528213500977, 0.019749055862426756, 0.019768991470336915, 0.019783456802368163, 0.01977996826171875, 0.019955743789672853]",tokens/s,12700.63421825852,kWh,2.3695549941546032e-07,1.298399891515554e-07,6.32767707282591e-07,9.995631958496068e-07,tokens/kWh,256111870.728099,MB,1298.550784,2229.796864,0.0,1583.34976,1426.271744,s,10,11.993829833984373,1.1993829833984373,0.014708442993412695,1.1939595947265627,1.206061279296875,1.2242870483398438,1.2388676635742188,"[1.2425128173828126, 1.1940203857421876, 1.2020111083984375, 1.195488525390625, 1.1934176025390626, 1.19032177734375, 1.191555419921875, 1.1975396728515626, 1.193063720703125, 1.1938988037109375]",tokens/s,52.52700836349225,kWh,1.4095023689238696e-05,7.721985782225197e-06,2.5064530628314826e-05,4.6881540099778724e-05,tokens/kWh,1343812.508418369,,s,629,12.152046592712408,0.019319628923231164,0.002443785106077763,0.01892767906188965,0.019181977462768555,0.020120576095581055,0.039161247406005864,"[0.020102144241333008, 0.02108518409729004, 0.020915199279785156, 0.01984921646118164, 0.021485567092895508, 0.021000192642211913, 0.021011455535888672, 0.021086208343505858, 0.02082508850097656, 0.02086911964416504, 0.021078016281127928, 0.02123980712890625, 0.020939775466918945, 0.0212541446685791, 0.021111808776855468, 0.02104115104675293, 0.020956159591674805, 0.021112831115722656, 0.021204992294311522, 0.01984716796875, 0.01970790481567383, 0.019598335266113282, 0.01947648048400879, 0.020592639923095703, 0.02019430351257324, 0.020182016372680665, 0.019544063568115236, 0.01945088005065918, 0.019180543899536134, 0.018950143814086915, 0.01902694320678711, 0.018921472549438476, 0.018925567626953126, 0.018938880920410156, 0.018982912063598634, 0.018967552185058592, 0.018939903259277344, 0.019080192565917968, 0.018856960296630858, 0.018936832427978514, 0.018927616119384767, 0.01905561637878418, 0.018900991439819336, 0.018970624923706055, 0.018944000244140623, 0.018934783935546876, 0.01908121681213379, 0.018907136917114258, 0.018962432861328125, 0.018949119567871094, 0.018918399810791017, 0.018909183502197266, 0.018959360122680666, 0.018951168060302736, 0.018922496795654296, 0.018964479446411133, 0.01903206443786621, 0.018972671508789063, 0.018983936309814452, 0.018959360122680666, 0.018932735443115235, 0.01900032043457031, 0.03935846328735351, 0.018915327072143554, 0.018945024490356444, 0.01899212837219238, 0.018907136917114258, 0.018927616119384767, 0.018948095321655273, 0.018924543380737305, 0.018950143814086915, 0.01901055908203125, 0.01900339126586914, 0.01899622344970703, 0.018936832427978514, 0.01887948799133301, 0.018969600677490234, 0.018894847869873048, 0.018913280487060546, 0.018885631561279297, 0.018907136917114258, 0.018899967193603515, 0.018932735443115235, 0.018899999618530273, 0.01888355255126953, 0.018874368667602538, 0.018937856674194335, 0.01886617660522461, 0.01886617660522461, 0.018925567626953126, 0.018946048736572265, 0.019137535095214844, 0.019143680572509765, 0.019188735961914064, 0.019070976257324217, 0.019017728805541992, 0.018885631561279297, 0.01899929618835449, 0.018934783935546876, 0.01902387237548828, 0.018928640365600585, 0.018892799377441406, 0.018921472549438476, 0.019086336135864256, 0.01903308868408203, 0.019021823883056642, 0.018903039932250978, 0.018885631561279297, 0.018897920608520507, 0.018919424057006837, 0.018926591873168946, 0.018967552185058592, 0.018876415252685547, 0.018900064468383788, 0.018922399520874024, 0.018932735443115235, 0.018902048110961914, 0.01891836738586426, 0.01887027168273926, 0.018869247436523438, 0.018931711196899414, 0.018909183502197266, 0.01899622344970703, 0.018890752792358398, 0.018946048736572265, 0.039636993408203126, 0.020348928451538087, 0.019594240188598632, 0.01903513526916504, 0.01899212837219238, 0.019005439758300782, 0.01904844856262207, 0.019556352615356445, 0.01903615951538086, 0.018916351318359375, 0.019165184020996092, 0.019075071334838867, 0.019277824401855468, 0.019125247955322267, 0.019260416030883788, 0.019293184280395507, 0.018942975997924806, 0.018972671508789063, 0.018973695755004884, 0.018946048736572265, 0.019126272201538085, 0.01899519920349121, 0.018919424057006837, 0.018982912063598634, 0.01900851249694824, 0.019009536743164062, 0.018980863571166993, 0.018952192306518553, 0.018929664611816405, 0.01899212837219238, 0.019006464004516603, 0.01904537582397461, 0.018988031387329102, 0.019094528198242186, 0.018985984802246093, 0.018955327987670897, 0.019052480697631834, 0.018983936309814452, 0.018930688858032226, 0.01892563247680664, 0.0189418888092041, 0.019144704818725586, 0.01921023941040039, 0.01899212837219238, 0.01899212837219238, 0.0190382080078125, 0.01899622344970703, 0.019005439758300782, 0.018947071075439453, 0.01902387237548828, 0.0188723201751709, 0.018963455200195312, 0.018947071075439453, 0.018927616119384767, 0.018980863571166993, 0.018948095321655273, 0.019002368927001953, 0.01893587112426758, 0.01891935920715332, 0.018911231994628908, 0.019124223709106446, 0.01927884864807129, 0.019721216201782226, 0.03974246215820312, 0.01886720085144043, 0.018965503692626954, 0.018922496795654296, 0.019325952529907226, 0.019380224227905272, 0.019075071334838867, 0.018928640365600585, 0.01909350395202637, 0.01902079963684082, 0.018964479446411133, 0.018905088424682616, 0.018970624923706055, 0.01905971145629883, 0.018958335876464845, 0.018953216552734374, 0.01900851249694824, 0.018953216552734374, 0.018979839324951172, 0.018972671508789063, 0.01904128074645996, 0.01877299118041992, 0.018942975997924806, 0.018969600677490234, 0.018972671508789063, 0.018928640365600585, 0.018892799377441406, 0.01888768005371094, 0.018975744247436522, 0.01903718376159668, 0.01902079963684082, 0.018917375564575196, 0.018977792739868164, 0.018944000244140623, 0.018989055633544923, 0.018945024490356444, 0.018891775131225585, 0.01889593505859375, 0.018829248428344728, 0.0189040641784668, 0.018892799377441406, 0.018962432861328125, 0.018964479446411133, 0.01889587211608887, 0.018903039932250978, 0.018928640365600585, 0.018937856674194335, 0.01889587211608887, 0.019019775390625, 0.019174400329589843, 0.018979839324951172, 0.018898944854736328, 0.018960384368896483, 0.018946048736572265, 0.01887948799133301, 0.018959360122680666, 0.019072000503540038, 0.019127296447753905, 0.018881536483764647, 0.018919424057006837, 0.018994176864624023, 0.018937856674194335, 0.01887846374511719, 0.039164993286132814, 0.018890687942504883, 0.01879449653625488, 0.018898944854736328, 0.01884364891052246, 0.01887948799133301, 0.018958335876464845, 0.018923519134521484, 0.01886412811279297, 0.018898944854736328, 0.01880575942993164, 0.018783231735229493, 0.01892665672302246, 0.018853824615478517, 0.018969600677490234, 0.018869247436523438, 0.01883647918701172, 0.018824192047119142, 0.018800640106201173, 0.018921472549438476, 0.01887539291381836, 0.01885593605041504, 0.01979903984069824, 0.019344383239746094, 0.018884607315063476, 0.018948095321655273, 0.018898944854736328, 0.018990079879760743, 0.018898944854736328, 0.018888704299926756, 0.018954240798950195, 0.018979839324951172, 0.018965503692626954, 0.0189040641784668, 0.018892799377441406, 0.019195903778076173, 0.018902015686035157, 0.019319807052612305, 0.018910207748413087, 0.018923519134521484, 0.01885593605041504, 0.018907136917114258, 0.018942975997924806, 0.018910207748413087, 0.018942975997924806, 0.01893791961669922, 0.018949056625366213, 0.018901023864746094, 0.018958303451538087, 0.01904332733154297, 0.01888768005371094, 0.018919424057006837, 0.018865152359008788, 0.018877439498901367, 0.018975744247436522, 0.018919424057006837, 0.018908159255981445, 0.018915327072143554, 0.018958335876464845, 0.018891775131225585, 0.018947071075439453, 0.01886310386657715, 0.018923519134521484, 0.03909836959838867, 0.01883340835571289, 0.01886207962036133, 0.018886655807495118, 0.018896896362304686, 0.018965503692626954, 0.018927616119384767, 0.018890752792358398, 0.01892767906188965, 0.018843584060668946, 0.0188590087890625, 0.018939903259277344, 0.01887129592895508, 0.0188272647857666, 0.018886655807495118, 0.018790399551391602, 0.01882624053955078, 0.018883583068847656, 0.018746368408203123, 0.018634752273559572, 0.018725887298583984, 0.018893823623657227, 0.01884876823425293, 0.018918399810791017, 0.01884160041809082, 0.018906112670898437, 0.019374080657958984, 0.018900991439819336, 0.018880512237548826, 0.018881536483764647, 0.01883545684814453, 0.018950143814086915, 0.018902015686035157, 0.018894847869873048, 0.01902796745300293, 0.01883750343322754, 0.018899967193603515, 0.018920448303222655, 0.018886655807495118, 0.01900339126586914, 0.01883340835571289, 0.018869247436523438, 0.01883135986328125, 0.018894847869873048, 0.019130367279052735, 0.018873344421386717, 0.01880678367614746, 0.018938880920410156, 0.01884364891052246, 0.018963455200195312, 0.01887027168273926, 0.01884160041809082, 0.01883955192565918, 0.018953216552734374, 0.018898944854736328, 0.01883955192565918, 0.01887948799133301, 0.018961408615112304, 0.01882624053955078, 0.01883033561706543, 0.018896896362304686, 0.01885593605041504, 0.018902015686035157, 0.03915161514282227, 0.01865318489074707, 0.018689023971557618, 0.01885491180419922, 0.018960384368896483, 0.0188590087890625, 0.018880512237548826, 0.018817024230957033, 0.018921472549438476, 0.018939903259277344, 0.01881804847717285, 0.018884607315063476, 0.018950143814086915, 0.01882931137084961, 0.018920448303222655, 0.019128320693969726, 0.019002368927001953, 0.018886655807495118, 0.018991104125976564, 0.018869247436523438, 0.018889728546142577, 0.018968576431274413, 0.018873344421386717, 0.019005439758300782, 0.018898944854736328, 0.018921472549438476, 0.018980863571166993, 0.01903718376159668, 0.018922496795654296, 0.018926591873168946, 0.018938880920410156, 0.018906112670898437, 0.018961408615112304, 0.018905088424682616, 0.018906112670898437, 0.018897920608520507, 0.018939903259277344, 0.019024927139282225, 0.019011552810668946, 0.01882009506225586, 0.018836511611938476, 0.01891734313964844, 0.018923519134521484, 0.01879756736755371, 0.01886720085144043, 0.018876415252685547, 0.01884262466430664, 0.01884876823425293, 0.01887846374511719, 0.01891846466064453, 0.018816959381103514, 0.01885081672668457, 0.018899967193603515, 0.018883583068847656, 0.018899967193603515, 0.01886617660522461, 0.01900032043457031, 0.018948095321655273, 0.018963455200195312, 0.018927616119384767, 0.018929664611816405, 0.01886617660522461, 0.01902284812927246, 0.03923046493530274, 0.018872352600097658, 0.01886409568786621, 0.01903718376159668, 0.01922764778137207, 0.018937856674194335, 0.018920448303222655, 0.01883852767944336, 0.019001344680786132, 0.018916351318359375, 0.018905120849609373, 0.019001312255859375, 0.01892870330810547, 0.01889683151245117, 0.018906112670898437, 0.018914304733276367, 0.01901568031311035, 0.018893823623657227, 0.018902015686035157, 0.019145727157592773, 0.018893823623657227, 0.01909350395202637, 0.018868223190307617, 0.018891775131225585, 0.01898700714111328, 0.01881497573852539, 0.018958335876464845, 0.018963455200195312, 0.018975744247436522, 0.018890752792358398, 0.018893823623657227, 0.018919424057006837, 0.018957311630249024, 0.019728384017944335, 0.019495935440063478, 0.018959360122680666, 0.020290559768676757, 0.019354623794555666, 0.01901568031311035, 0.018905088424682616, 0.019127296447753905, 0.01899929618835449, 0.018929664611816405, 0.018883583068847656, 0.018922496795654296, 0.018942975997924806, 0.018946048736572265, 0.018877439498901367, 0.018990079879760743, 0.019139583587646485, 0.019086336135864256, 0.01900032043457031, 0.01881497573852539, 0.018922496795654296, 0.018952192306518553, 0.018913280487060546, 0.018954240798950195, 0.019172351837158205, 0.018990079879760743, 0.018905088424682616, 0.018851839065551757, 0.01888768005371094, 0.018949119567871094, 0.039215103149414066, 0.018902015686035157, 0.018935808181762694, 0.018942975997924806, 0.019001344680786132, 0.018923519134521484, 0.018911231994628908, 0.01884569549560547, 0.018873344421386717, 0.019013631820678712, 0.018909183502197266, 0.018928640365600585, 0.01881395149230957, 0.018860095977783202, 0.018867136001586914, 0.01883852767944336, 0.018968576431274413, 0.018880512237548826, 0.01886720085144043, 0.018780160903930664, 0.018912256240844725, 0.01883852767944336, 0.018911231994628908, 0.01883545684814453, 0.018886655807495118, 0.01882111930847168, 0.018874368667602538, 0.01887129592895508, 0.019190784454345702, 0.018885631561279297, 0.018800640106201173, 0.018883583068847656, 0.01892252731323242, 0.018823135375976564, 0.018967552185058592, 0.01906073570251465, 0.018993152618408202, 0.0188538875579834, 0.018909183502197266, 0.01883545684814453, 0.01884979248046875, 0.018897920608520507, 0.018911264419555665, 0.018855903625488283, 0.018925567626953126, 0.019170303344726563, 0.019722240447998047, 0.019086336135864256, 0.018884607315063476, 0.018925567626953126, 0.01897881507873535, 0.018908159255981445, 0.018927616119384767, 0.019109888076782225, 0.018885631561279297, 0.018911231994628908, 0.018950143814086915, 0.018877439498901367, 0.01901875114440918, 0.01898700714111328, 0.018869247436523438, 0.018932767868041992, 0.01894294357299805, 0.03943119812011719, 0.018851808547973633, 0.018909183502197266, 0.018942975997924806, 0.018916351318359375, 0.018962432861328125, 0.018976831436157228, 0.01894598388671875, 0.018832384109497072, 0.018888704299926756, 0.018896896362304686, 0.018880512237548826, 0.018926591873168946, 0.01884876823425293, 0.018923519134521484, 0.018911231994628908, 0.01904640007019043, 0.018892799377441406, 0.018910207748413087, 0.01885593605041504, 0.01890614318847656, 0.01891427230834961, 0.018912256240844725, 0.0188723201751709, 0.018856960296630858, 0.018876415252685547, 0.018922496795654296, 0.01887129592895508, 0.018919424057006837, 0.018888704299926756, 0.01884262466430664, 0.019005439758300782, 0.018939903259277344, 0.01884876823425293, 0.01900032043457031, 0.0191016960144043, 0.018941951751708985, 0.018934783935546876, 0.018918399810791017, 0.018851839065551757, 0.018885631561279297, 0.018888704299926756, 0.018896896362304686, 0.018881536483764647, 0.018954240798950195, 0.019108863830566408, 0.020132863998413086, 0.019084287643432618, 0.018916351318359375, 0.018970624923706055, 0.018896896362304686, 0.018957311630249024, 0.0188538875579834, 0.018889728546142577, 0.01890822410583496, 0.01892140769958496, 0.018942975997924806, 0.018969600677490234, 0.018876415252685547, 0.018920448303222655, 0.018932735443115235, 0.01881292724609375, 0.019187711715698243]",tokens/s,51.760828532143066,,,main,False,False,,,,,,,,,,,,,,,,,,,,,,, bfloat16-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-1.8B,,cuda,0,42,,,,,True,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,217063f5c507ed7cc255df7e1f64c4333a0b4dfe,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,,MB,1272.225792,5222.432768,0.0,4575.985664,4408.404992,s,10,0.22872108650207515,0.02287210865020752,0.0014333987066370093,0.02240225601196289,0.02294665241241455,0.025058829784393305,0.02674857168197632,"[0.027171007156372072, 0.022398687362670897, 0.02240876770019531, 0.022405824661254882, 0.022477279663085936, 0.02237504005432129, 0.022351551055908202, 0.022351680755615236, 0.02241417694091797, 0.0223670711517334]",tokens/s,11192.671559719847,kWh,2.662624772246536e-07,1.458880744185326e-07,1.223439417189218e-06,1.6355899688324041e-06,tokens/kWh,156518445.86864904,MB,1272.225792,5222.432768,0.0,4575.985664,4408.407552,s,10,12.555456298828124,1.2555456298828125,0.0031298792264682768,1.2557788696289063,1.2598449340820312,1.2602314880371093,1.2605407312011718,"[1.2606180419921875, 1.256319091796875, 1.25747705078125, 1.252046875, 1.251937744140625, 1.2525050048828126, 1.2520966796875, 1.2552386474609376, 1.259759033203125, 1.2574581298828125]",tokens/s,50.177387822918206,kWh,1.4821127998578297e-05,8.121728151220826e-06,3.728181360921318e-05,6.02246697590123e-05,tokens/kWh,1046082.9466079784,,s,629,12.73492682266234,0.020246306554312164,0.002720546001744483,0.019894271850585937,0.02012180404663086,0.02045071334838867,0.04268367858886719,"[0.021436416625976562, 0.020221952438354493, 0.020122623443603514, 0.0196945915222168, 0.01990656089782715, 0.019689472198486328, 0.01974176025390625, 0.019659711837768556, 0.019768320083618163, 0.01964851188659668, 0.019677183151245118, 0.01965465545654297, 0.020074495315551756, 0.019970048904418947, 0.019922943115234376, 0.01995065689086914, 0.019897279739379884, 0.020199424743652345, 0.020116479873657226, 0.019900415420532228, 0.019878911972045898, 0.019958784103393554, 0.020130815505981444, 0.019900415420532228, 0.01988096046447754, 0.01982259178161621, 0.019911680221557617, 0.020585472106933594, 0.020221952438354493, 0.02038374328613281, 0.01990553665161133, 0.019741695404052736, 0.01969152069091797, 0.01989836883544922, 0.020676607131958007, 0.019954687118530275, 0.019932159423828123, 0.019792896270751953, 0.02030284881591797, 0.01974995231628418, 0.019694528579711913, 0.019721216201782226, 0.019755008697509766, 0.01981439971923828, 0.019718143463134767, 0.019741695404052736, 0.01966796875, 0.019731456756591798, 0.01970278358459473, 0.019698688507080078, 0.02006630325317383, 0.020017215728759766, 0.02019731140136719, 0.020510719299316405, 0.020756479263305663, 0.02023219108581543, 0.019986431121826173, 0.019962879180908204, 0.019936256408691407, 0.019900415420532228, 0.019918848037719726, 0.019933183670043944, 0.042788864135742184, 0.019915775299072267, 0.01992915153503418, 0.019855295181274414, 0.019941375732421874, 0.020148223876953125, 0.019959808349609375, 0.0198922233581543, 0.019710975646972655, 0.019826688766479493, 0.019912704467773438, 0.019907583236694337, 0.01987379264831543, 0.01983897590637207, 0.019817535400390624, 0.01973446464538574, 0.01987583923339844, 0.01986764717102051, 0.019899391174316407, 0.01985843276977539, 0.019997695922851562, 0.020105215072631837, 0.02041753578186035, 0.020494335174560546, 0.020567039489746093, 0.020505599975585938, 0.020007936477661133, 0.019976192474365235, 0.019938304901123048, 0.020377599716186523, 0.019900415420532228, 0.01986457633972168, 0.019900415420532228, 0.019876863479614256, 0.019920896530151368, 0.01985536003112793, 0.01992192077636719, 0.019954687118530275, 0.019796991348266603, 0.02030284881591797, 0.019876863479614256, 0.020017152786254884, 0.019895296096801757, 0.019973119735717772, 0.019989503860473632, 0.019919872283935547, 0.01997209548950195, 0.01968230438232422, 0.01966694450378418, 0.0196823673248291, 0.019783615112304687, 0.01985536003112793, 0.019937280654907227, 0.019907583236694337, 0.019935232162475586, 0.019955711364746095, 0.020006912231445313, 0.019777536392211914, 0.019738624572753907, 0.019817472457885742, 0.019705856323242187, 0.01968339157104492, 0.01973855972290039, 0.042807296752929686, 0.01998028755187988, 0.020099071502685546, 0.019969024658203126, 0.019975168228149414, 0.02002022361755371, 0.019973119735717772, 0.0200898551940918, 0.01993011283874512, 0.01988198471069336, 0.019966976165771484, 0.019957759857177734, 0.019938304901123048, 0.019933183670043944, 0.019922943115234376, 0.020139007568359374, 0.020372480392456056, 0.021173248291015623, 0.020789247512817383, 0.02006937599182129, 0.019727359771728514, 0.019710975646972655, 0.019771392822265626, 0.02001919937133789, 0.019931135177612306, 0.019938304901123048, 0.019664896011352538, 0.019899391174316407, 0.019941375732421874, 0.019919872283935547, 0.019935232162475586, 0.019944448471069336, 0.01987379264831543, 0.019775487899780272, 0.019768320083618163, 0.019909631729125975, 0.02000486373901367, 0.019948543548583983, 0.019927040100097656, 0.020007936477661133, 0.019909631729125975, 0.019991552352905274, 0.019895296096801757, 0.019851264953613282, 0.019885055541992186, 0.019776512145996093, 0.019742719650268553, 0.019765247344970704, 0.01987379264831543, 0.019968000411987305, 0.019973119735717772, 0.019932159423828123, 0.019903488159179687, 0.019991552352905274, 0.019900415420532228, 0.019776512145996093, 0.019786752700805665, 0.02009600067138672, 0.020002815246582033, 0.019920896530151368, 0.01972428894042969, 0.019693567276000978, 0.019941375732421874, 0.042744831085205076, 0.019855392456054687, 0.01982972717285156, 0.019851264953613282, 0.01982464027404785, 0.019796991348266603, 0.019631103515625, 0.0196495361328125, 0.01960550308227539, 0.01972428894042969, 0.01995884895324707, 0.019676095962524415, 0.019728384017944335, 0.019941375732421874, 0.01984000015258789, 0.01983590316772461, 0.0198287353515625, 0.019719167709350584, 0.01984000015258789, 0.019667999267578125, 0.019641311645507812, 0.019765247344970704, 0.01981644821166992, 0.019842048645019532, 0.019809280395507813, 0.019909631729125975, 0.01982771110534668, 0.02001919937133789, 0.019885055541992186, 0.019920896530151368, 0.019891199111938478, 0.019896320343017578, 0.019861503601074217, 0.01988812828063965, 0.0198922233581543, 0.019877887725830077, 0.020006912231445313, 0.019948543548583983, 0.019857440948486328, 0.02017791938781738, 0.020028383255004882, 0.019896320343017578, 0.019910655975341796, 0.019739648818969727, 0.019917823791503905, 0.019923967361450197, 0.01989017677307129, 0.01986867141723633, 0.019870719909667968, 0.019944448471069336, 0.02012057685852051, 0.02002124786376953, 0.02003046417236328, 0.02007040023803711, 0.01997209548950195, 0.019931135177612306, 0.0198287353515625, 0.01981439971923828, 0.019748863220214845, 0.01988915252685547, 0.01984000015258789, 0.02000486373901367, 0.019944511413574218, 0.042783679962158205, 0.01982975959777832, 0.01966182327270508, 0.019784704208374023, 0.01969049644470215, 0.01964646339416504, 0.019734527587890623, 0.01966694450378418, 0.019729408264160156, 0.019719167709350584, 0.019736576080322265, 0.01989017677307129, 0.019909631729125975, 0.019955711364746095, 0.019862527847290038, 0.020023296356201172, 0.019853311538696287, 0.019934207916259765, 0.01967411231994629, 0.01965056037902832, 0.019720191955566405, 0.01963724708557129, 0.020051967620849608, 0.01998847961425781, 0.01993011283874512, 0.019899391174316407, 0.019862527847290038, 0.019917823791503905, 0.019922943115234376, 0.020214784622192384, 0.02003558349609375, 0.02002739143371582, 0.01987379264831543, 0.019945472717285157, 0.01984409523010254, 0.019869695663452147, 0.019647487640380858, 0.019773439407348634, 0.019929088592529298, 0.020242431640625, 0.02000588798522949, 0.019851264953613282, 0.0198922233581543, 0.01985843276977539, 0.019854335784912108, 0.019993600845336915, 0.01984102439880371, 0.019907583236694337, 0.019824703216552733, 0.019876800537109374, 0.019919872283935547, 0.01983590316772461, 0.01984409523010254, 0.019882015228271484, 0.019843040466308595, 0.019714048385620117, 0.020011007308959963, 0.019883007049560548, 0.01988403129577637, 0.019932159423828123, 0.019922943115234376, 0.01988096046447754, 0.019911680221557617, 0.042714111328125, 0.019883007049560548, 0.019887104034423828, 0.01966080093383789, 0.01969049644470215, 0.019696640014648437, 0.019870719909667968, 0.01983283233642578, 0.02002739143371582, 0.020229120254516602, 0.01985843276977539, 0.01986662483215332, 0.019963903427124022, 0.019851264953613282, 0.01983180809020996, 0.019951616287231445, 0.02002739143371582, 0.019932159423828123, 0.019957759857177734, 0.019914751052856446, 0.019899391174316407, 0.01991372871398926, 0.019903488159179687, 0.019664896011352538, 0.019706880569458008, 0.019734527587890623, 0.019813375473022463, 0.01971615982055664, 0.019678144454956054, 0.019671104431152345, 0.0196658878326416, 0.019643360137939454, 0.01981644821166992, 0.01998028755187988, 0.019983360290527344, 0.01983897590637207, 0.02006425666809082, 0.019810304641723633, 0.01989017677307129, 0.019936256408691407, 0.01990457534790039, 0.01994028854370117, 0.020025344848632814, 0.019895296096801757, 0.020462591171264647, 0.019970048904418947, 0.01987379264831543, 0.01987583923339844, 0.019700735092163087, 0.019793920516967774, 0.019942399978637695, 0.019938304901123048, 0.019968000411987305, 0.019898399353027344, 0.01987171173095703, 0.01984511947631836, 0.01986764717102051, 0.019672063827514647, 0.019599359512329103, 0.019895296096801757, 0.019887104034423828, 0.020206592559814454, 0.019900415420532228, 0.042635265350341796, 0.020101119995117187, 0.02004377555847168, 0.01980620765686035, 0.01988403129577637, 0.019935232162475586, 0.019870719909667968, 0.01985843276977539, 0.019870719909667968, 0.019680255889892577, 0.019722240447998047, 0.019719200134277345, 0.019775455474853515, 0.01980723190307617, 0.020254720687866212, 0.019887104034423828, 0.01987174415588379, 0.02005504035949707, 0.019728384017944335, 0.01997209548950195, 0.020008960723876954, 0.019974143981933593, 0.019994623184204103, 0.019934207916259765, 0.01988403129577637, 0.019851264953613282, 0.019741727828979493, 0.01969865608215332, 0.019680255889892577, 0.019801088333129883, 0.0197325439453125, 0.019651519775390626, 0.01965158462524414, 0.019834880828857423, 0.01984819221496582, 0.019753984451293945, 0.02002022361755371, 0.01988812828063965, 0.02043391990661621, 0.02008684730529785, 0.019865535736083986, 0.019862527847290038, 0.01981439971923828, 0.019825664520263672, 0.019842048645019532, 0.019924991607666014, 0.02005504035949707, 0.019878911972045898, 0.019886079788208007, 0.019962911605834962, 0.019795936584472658, 0.019719167709350584, 0.019767295837402343, 0.019915775299072267, 0.0198922233581543, 0.019825664520263672, 0.019894271850585937, 0.019869695663452147, 0.01986662483215332, 0.01982156753540039, 0.019727359771728514, 0.019598335266113282, 0.01992192077636719, 0.042665985107421874, 0.019796991348266603, 0.019671104431152345, 0.019626943588256836, 0.019834880828857423, 0.019802112579345704, 0.02008780860900879, 0.02000486373901367, 0.01986774444580078, 0.0198624324798584, 0.01988812828063965, 0.01970790481567383, 0.019796991348266603, 0.01963724708557129, 0.019760128021240234, 0.019642400741577148, 0.01968227195739746, 0.01964851188659668, 0.019864608764648437, 0.019857376098632813, 0.019926015853881835, 0.0198973445892334, 0.019885055541992186, 0.01987276840209961, 0.019908607482910155, 0.019929088592529298, 0.019773439407348634, 0.01967820739746094, 0.01965977668762207, 0.019765247344970704, 0.019628032684326172, 0.020123647689819335, 0.019994623184204103, 0.019861503601074217, 0.019883039474487305, 0.019828704833984374, 0.019762176513671875, 0.01966592025756836, 0.01962188720703125, 0.02007142448425293, 0.019853311538696287, 0.01984511947631836, 0.019918848037719726, 0.019951616287231445, 0.01983692741394043, 0.01985638427734375, 0.020115455627441405, 0.01990553665161133, 0.020060159683227538, 0.019943424224853516, 0.019910655975341796, 0.019917823791503905, 0.01987276840209961, 0.019965951919555663, 0.022185983657836913, 0.020792320251464845, 0.02006937599182129, 0.020121599197387697, 0.02001919937133789, 0.020135936737060548, 0.020017215728759766, 0.01994745635986328, 0.019939327239990236, 0.04274995040893555, 0.019887104034423828, 0.019883007049560548, 0.019943424224853516, 0.019900415420532228, 0.01988921546936035, 0.019882944107055663, 0.01985945510864258, 0.019933183670043944, 0.019917823791503905, 0.019885055541992186, 0.01984614372253418, 0.019711999893188475, 0.019596288681030274, 0.01962700843811035, 0.01962188720703125, 0.01962700843811035, 0.019775487899780272, 0.01965772819519043, 0.01984409523010254, 0.019903488159179687, 0.01990656089782715, 0.019951616287231445, 0.019878911972045898, 0.019944448471069336, 0.02001919937133789, 0.020098047256469728, 0.019896320343017578, 0.02007244873046875, 0.019896320343017578, 0.019926015853881835, 0.020002815246582033, 0.019974143981933593, 0.019900415420532228, 0.019894271850585937, 0.01979091262817383, 0.019626943588256836, 0.019718175888061525, 0.01968124771118164, 0.019720191955566405, 0.019933183670043944, 0.01992192077636719, 0.02021887969970703, 0.020384767532348632, 0.021629951477050782, 0.02067865562438965, 0.02047283172607422, 0.02043903923034668, 0.01996905517578125, 0.019977184295654298, 0.019965951919555663, 0.019932159423828123, 0.019985408782958985, 0.020526079177856444, 0.020519935607910156, 0.02045849609375, 0.019933183670043944, 0.01993011283874512, 0.02024345588684082, 0.020413440704345705, 0.019909631729125975, 0.019993600845336915, 0.019927040100097656, 0.04269055938720703, 0.019899391174316407, 0.019826688766479493, 0.0196997127532959, 0.019809280395507813, 0.01988403129577637, 0.019918848037719726, 0.01987993621826172, 0.01985228729248047, 0.019811328887939454, 0.019748863220214845, 0.021755903244018555, 0.02129408073425293, 0.02049228858947754, 0.019876863479614256, 0.019935232162475586, 0.01982259178161621, 0.02031718444824219, 0.02123980712890625, 0.020098047256469728, 0.019899391174316407, 0.019940351486206053, 0.019998720169067383, 0.019607551574707033, 0.01965056037902832, 0.01963827133178711, 0.019721216201782226, 0.019899391174316407, 0.020147199630737304, 0.019833856582641602, 0.019966976165771484, 0.019886079788208007, 0.01991372871398926, 0.0198973445892334, 0.019903488159179687, 0.019885152816772462, 0.019882911682128905, 0.019907583236694337, 0.019917823791503905, 0.019818496704101563, 0.01987174415588379, 0.019655679702758787, 0.019612672805786133, 0.01959321594238281, 0.01969254493713379, 0.019885055541992186, 0.019826688766479493, 0.019812351226806642, 0.01990656089782715, 0.02000588798522949, 0.01987174415588379, 0.019868703842163087, 0.01990345573425293, 0.019948543548583983, 0.019848224639892578, 0.0200980167388916, 0.019912704467773438, 0.01985843276977539, 0.01984819221496582, 0.01986662483215332, 0.01984000015258789, 0.02001919937133789, 0.019944448471069336]",tokens/s,49.39172472358988,,,main,False,False,,,,,,,,,,,,,,,,,,,,,,, bfloat16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-70m,EleutherAI/pythia-70m,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.42.1,,0.31.0,,,,1.20.0,,,,0.11.1,,,,MB,1261.473792,885.522432,0.0,239.075328,215.592448,s,33,0.17760159921646118,0.005381866642923065,0.00021256305866220254,0.005331840038299561,0.005506867408752441,0.005538207817077637,0.0061647781372070306,"[0.006458911895751953, 0.00553718376159668, 0.005409152030944824, 0.005300479888916016, 0.005301663875579834, 0.005507808208465576, 0.005503104209899902, 0.005539743900299072, 0.005370463848114014, 0.005302879810333252, 0.005262335777282715, 0.005385791778564453, 0.00529420804977417, 0.005373023986816407, 0.005361023902893066, 0.0052997121810913084, 0.0052657279968261715, 0.005485407829284668, 0.005422688007354736, 0.005368063926696778, 0.005369887828826904, 0.005331840038299561, 0.005489312171936035, 0.005390624046325684, 0.005196896076202393, 0.005220831871032715, 0.0052566399574279785, 0.005258016109466553, 0.005291840076446533, 0.005233024120330811, 0.005234464168548584, 0.005262879848480224, 0.005315968036651611]",tokens/s,47567.13924463913,kWh,6.52325902842943e-08,3.5744362460391916e-08,1.2993813695858626e-07,2.3091508970327246e-07,tokens/kWh,1108632616.12293,MB,1261.473792,885.522432,0.0,239.075328,215.595008,s,33,9.956242248535156,0.3017043105616714,0.004753360648444936,0.3008162841796875,0.3083882080078125,0.3100695129394531,0.3143244152832031,"[0.3156562805175781, 0.3091197204589844, 0.3029146728515625, 0.3015790100097656, 0.30037689208984375, 0.3048340148925781, 0.3076331787109375, 0.31149420166015623, 0.3031180419921875, 0.2998338623046875, 0.300911865234375, 0.2993684997558594, 0.30265963745117186, 0.3028968505859375, 0.3008162841796875, 0.29663265991210935, 0.30330133056640624, 0.2971231994628906, 0.3002906188964844, 0.30857696533203127, 0.3035245361328125, 0.3047144775390625, 0.306418701171875, 0.2971261291503906, 0.296899658203125, 0.29722442626953127, 0.297948974609375, 0.29650949096679685, 0.29779470825195314, 0.2966572265625, 0.29601364135742186, 0.2977251281738281, 0.29854736328125]",tokens/s,208.8137218945109,kWh,3.4801169149282775e-06,1.906940933858737e-06,6.023130510982667e-06,1.1410188359769686e-05,tokens/kWh,5521381.243987777,,s,2078,10.122535941600766,0.004871287748604812,0.0007229612673888442,0.004738048076629638,0.004935679912567138,0.004990259289741516,0.010457559404373169,"[0.005657599925994873, 0.005623807907104492, 0.00557260799407959, 0.005538815975189209, 0.00557260799407959, 0.005535744190216065, 0.005500927925109863, 0.00597811222076416, 0.005108736038208008, 0.005064703941345215, 0.005000192165374756, 0.005198847770690918, 0.004978687763214112, 0.004997119903564453, 0.004960256099700928, 0.0049203200340270994, 0.004969471931457519, 0.00499510383605957, 0.0049192638397216795, 0.004971519947052002, 0.004939775943756103, 0.004951039791107178, 0.0050135040283203125, 0.0049162240028381345, 0.0048158721923828125, 0.004757503986358643, 0.00485478401184082, 0.004915200233459473, 0.005027872085571289, 0.004943903923034668, 0.004935616016387939, 0.004935679912567138, 0.004937727928161621, 0.004931583881378174, 0.004873216152191162, 0.004753407955169678, 0.004767744064331054, 0.004745215892791748, 0.00475648021697998, 0.004764671802520752, 0.0047554559707641605, 0.004806655883789063, 0.004715519905090332, 0.004744192123413086, 0.004754432201385498, 0.004754432201385498, 0.00473199987411499, 0.004908959865570068, 0.004918272018432617, 0.005008384227752686, 0.004905983924865722, 0.004891647815704346, 0.00501964807510376, 0.004903935909271241, 0.004988927841186524, 0.004915200233459473, 0.004898816108703613, 0.004899839878082276, 0.004894752025604248, 0.0049387521743774416, 0.004889567852020263, 0.004937727928161621, 0.011019264221191406, 0.004941823959350586, 0.004907008171081543, 0.004868095874786377, 0.004747263908386231, 0.004809728145599365, 0.004836351871490479, 0.004865024089813232, 0.004959231853485108, 0.004898816108703613, 0.004965375900268554, 0.004923391819000244, 0.004850751876831054, 0.004759488105773926, 0.004731904029846192, 0.004738048076629638, 0.0047513599395751956, 0.0047513599395751956, 0.004744192123413086, 0.004746240139007568, 0.004731904029846192, 0.004878335952758789, 0.004947968006134033, 0.005246975898742676, 0.004908031940460205, 0.004935679912567138, 0.004941823959350586, 0.0048865280151367185, 0.004967423915863037, 0.00491215991973877, 0.004923359870910644, 0.004967423915863037, 0.004908031940460205, 0.004994048118591309, 0.004894720077514648, 0.0049500160217285155, 0.004894720077514648, 0.004893695831298828, 0.004958208084106445, 0.004931583881378174, 0.004958208084106445, 0.004902912139892578, 0.00491212797164917, 0.0051066880226135255, 0.0050421757698059086, 0.005015552043914795, 0.00495411205291748, 0.004965375900268554, 0.00502784013748169, 0.004943871974945068, 0.005045248031616211, 0.004988927841186524, 0.0049459199905395506, 0.004941823959350586, 0.004931583881378174, 0.005007359981536865, 0.004957183837890625, 0.004787199974060059, 0.004774911880493164, 0.004748288154602051, 0.004762656211853027, 0.004761568069458008, 0.004731904029846192, 0.01061683177947998, 0.004759552001953125, 0.0047288317680358885, 0.0047329277992248535, 0.004757503986358643, 0.004729856014251709, 0.0048189439773559575, 0.0047216639518737795, 0.004723711967468262, 0.004753407955169678, 0.0047216639518737795, 0.004753407955169678, 0.0047964158058166504, 0.004729856014251709, 0.004839424133300781, 0.004770815849304199, 0.004846591949462891, 0.004729856014251709, 0.004746240139007568, 0.004760575771331787, 0.004723775863647461, 0.00472979211807251, 0.004922368049621582, 0.004894720077514648, 0.0049060478210449215, 0.004876224040985108, 0.004910079956054687, 0.004888576030731201, 0.004767744064331054, 0.004750336170196533, 0.0047288317680358885, 0.004741119861602783, 0.004741119861602783, 0.004749311923980713, 0.004776959896087647, 0.0047288317680358885, 0.004746240139007568, 0.004736000061035156, 0.004736000061035156, 0.004772863864898682, 0.004733952045440673, 0.0047216639518737795, 0.0048895998001098635, 0.004944896221160889, 0.004929535865783692, 0.004911104202270508, 0.004892672061920166, 0.00490502405166626, 0.004876224040985108, 0.0049203200340270994, 0.004877312183380127, 0.004907008171081543, 0.004903935909271241, 0.0047554559707641605, 0.004914175987243652, 0.004958208084106445, 0.004884479999542236, 0.004803584098815918, 0.004767744064331054, 0.0047513599395751956, 0.004748288154602051, 0.004748288154602051, 0.004758528232574463, 0.010498047828674317, 0.004727807998657227, 0.004729856014251709, 0.004753439903259277, 0.004736991882324218, 0.004777984142303467, 0.004738048076629638, 0.004709375858306885, 0.00474729585647583, 0.0047513279914855955, 0.004741119861602783, 0.004754432201385498, 0.004747263908386231, 0.004743167877197266, 0.004742144107818603, 0.004740096092224121, 0.004777984142303467, 0.004769792079925537, 0.0048220157623291016, 0.004729856014251709, 0.004743167877197266, 0.004832255840301514, 0.00496127986907959, 0.0049162240028381345, 0.004914175987243652, 0.004899839878082276, 0.0049387521743774416, 0.004829184055328369, 0.004922368049621582, 0.004918272018432617, 0.004946944236755371, 0.004923391819000244, 0.004794367790222168, 0.004772863864898682, 0.004717567920684815, 0.004757503986358643, 0.004765696048736572, 0.004746240139007568, 0.004775936126708984, 0.004733952045440673, 0.0047554559707641605, 0.004746240139007568, 0.004737023830413818, 0.004767744064331054, 0.004768767833709717, 0.004733952045440673, 0.004758528232574463, 0.004731904029846192, 0.004734975814819336, 0.004750400066375732, 0.004734911918640137, 0.004780096054077148, 0.004734911918640137, 0.00475648021697998, 0.004757503986358643, 0.004726784229278564, 0.004833280086517334, 0.004744192123413086, 0.004744192123413086, 0.004758528232574463, 0.004750336170196533, 0.004795392036437988, 0.004759552001953125, 0.010460160255432128, 0.004724736213684082, 0.004781055927276612, 0.004726784229278564, 0.004733952045440673, 0.00475648021697998, 0.004727807998657227, 0.0047851519584655765, 0.004726784229278564, 0.004724800109863281, 0.004790207862854004, 0.00475648021697998, 0.004777984142303467, 0.0047216639518737795, 0.0047185921669006346, 0.004750336170196533, 0.004729856014251709, 0.004744192123413086, 0.004748288154602051, 0.004712448120117187, 0.004825088024139404, 0.004736000061035156, 0.004717567920684815, 0.004779007911682129, 0.004731904029846192, 0.004740096092224121, 0.004758528232574463, 0.004726784229278564, 0.004777984142303467, 0.004742144107818603, 0.0047719039916992185, 0.0047738242149353025, 0.004745215892791748, 0.004772863864898682, 0.004761600017547608, 0.004747263908386231, 0.004736000061035156, 0.004740096092224121, 0.004773888111114502, 0.005253119945526123, 0.004904960155487061, 0.004891647815704346, 0.00487014389038086, 0.004935679912567138, 0.0047851519584655765, 0.004716544151306152, 0.004684800148010254, 0.004679679870605469, 0.004724736213684082, 0.004685823917388916, 0.0046694397926330565, 0.004677631855010986, 0.004685823917388916, 0.0047329277992248535, 0.0047513599395751956, 0.004727807998657227, 0.00475648021697998, 0.004766719818115234, 0.004760575771331787, 0.004743167877197266, 0.00471449613571167, 0.004757503986358643, 0.0048455681800842285, 0.010941439628601075, 0.004881408214569092, 0.004910079956054687, 0.004890624046325683, 0.004911104202270508, 0.004903935909271241, 0.0049162240028381345, 0.004847616195678711, 0.004750336170196533, 0.004760575771331787, 0.004743167877197266, 0.004740096092224121, 0.004784128189086914, 0.004771840095520019, 0.0047626237869262695, 0.004745215892791748, 0.0048865280151367185, 0.0049203519821166995, 0.005751776218414307, 0.004941823959350586, 0.004888576030731201, 0.0047185921669006346, 0.0047851519584655765, 0.004680704116821289, 0.004720640182495117, 0.004750336170196533, 0.004758528232574463, 0.004794367790222168, 0.004744192123413086, 0.00481279993057251, 0.004964352130889893, 0.004955135822296143, 0.004982816219329834, 0.004890592098236084, 0.004829184055328369, 0.004771840095520019, 0.004723711967468262, 0.004843520164489746, 0.004907008171081543, 0.004834303855895996, 0.004726784229278564, 0.004738048076629638, 0.004779007911682129, 0.0047226881980895995, 0.004737023830413818, 0.004723711967468262, 0.004655104160308838, 0.0047523841857910155, 0.004686848163604736, 0.004719615936279297, 0.004830207824707031, 0.004810815811157227, 0.004732863903045654, 0.004707327842712402, 0.004676608085632325, 0.004737023830413818, 0.004791296005249023, 0.0048158721923828125, 0.004900864124298096, 0.005286911964416504, 0.004924416065216064, 0.004866079807281494, 0.004943903923034668, 0.010836928367614747, 0.004881408214569092, 0.004866047859191895, 0.0050032639503479, 0.004934656143188477, 0.0049387521743774416, 0.004808767795562744, 0.004734911918640137, 0.004748288154602051, 0.004715519905090332, 0.00475654411315918, 0.004726719856262207, 0.004744224071502686, 0.004772831916809082, 0.004736000061035156, 0.004745215892791748, 0.004749311923980713, 0.004726784229278564, 0.0047554559707641605, 0.004723711967468262, 0.004777984142303467, 0.004927487850189209, 0.0048855037689208985, 0.004951039791107178, 0.004888576030731201, 0.004968448162078858, 0.004881408214569092, 0.004892672061920166, 0.004947968006134033, 0.0048793601989746095, 0.004966400146484375, 0.004883488178253174, 0.0049049282073974606, 0.004952064037322998, 0.004895743846893311, 0.00521830415725708, 0.004943871974945068, 0.004899839878082276, 0.004976640224456787, 0.004863999843597412, 0.004959231853485108, 0.004816895961761475, 0.004811776161193848, 0.004882431983947754, 0.004828159809112549, 0.00486195182800293, 0.004900864124298096, 0.004928512096405029, 0.00495411205291748, 0.004880383968353271, 0.0049530878067016604, 0.004890624046325683, 0.0049172801971435545, 0.004949984073638916, 0.004893695831298828, 0.004944960117340088, 0.004908991813659668, 0.004887551784515381, 0.004964352130889893, 0.004899839878082276, 0.004956160068511963, 0.0048752641677856446, 0.004894720077514648, 0.010869759559631348, 0.004925439834594727, 0.004901887893676758, 0.004923391819000244, 0.004911104202270508, 0.004914175987243652, 0.004986879825592041, 0.004890624046325683, 0.00501043176651001, 0.004890624046325683, 0.004896768093109131, 0.0049489917755126955, 0.004873216152191162, 0.004952064037322998, 0.004880383968353271, 0.0048855037689208985, 0.004859903812408447, 0.004741119861602783, 0.004876287937164306, 0.00486300802230835, 0.004883423805236816, 0.0048855037689208985, 0.004891647815704346, 0.004952064037322998, 0.004928512096405029, 0.004878335952758789, 0.004968448162078858, 0.004893695831298828, 0.004937727928161621, 0.004902912139892578, 0.00486195182800293, 0.004941823959350586, 0.00487014389038086, 0.005005311965942383, 0.004859903812408447, 0.004895743846893311, 0.004924416065216064, 0.004867072105407715, 0.005441535949707031, 0.005022719860076904, 0.004976640224456787, 0.005322751998901368, 0.005819392204284668, 0.005039103984832764, 0.0049459199905395506, 0.004897791862487793, 0.004947968006134033, 0.0049500160217285155, 0.004902912139892578, 0.005134335994720459, 0.004911104202270508, 0.0048855037689208985, 0.005132287979125977, 0.004976672172546387, 0.004926464080810547, 0.0049008321762084965, 0.004880383968353271, 0.004921343803405762, 0.004893695831298828, 0.004739071846008301, 0.004712448120117187, 0.004720640182495117, 0.0047226881980895995, 0.010514431953430176, 0.00470527982711792, 0.004717567920684815, 0.00471449613571167, 0.004747263908386231, 0.004708352088928222, 0.004798463821411133, 0.004898816108703613, 0.004884479999542236, 0.0049275197982788085, 0.004886496067047119, 0.004908031940460205, 0.0049428482055664065, 0.004903935909271241, 0.0047636480331420894, 0.004733952045440673, 0.0047226881980895995, 0.004743167877197266, 0.004709375858306885, 0.004724736213684082, 0.004738048076629638, 0.004723711967468262, 0.004721727848052979, 0.0047103362083435055, 0.004729856014251709, 0.0047226881980895995, 0.0047185921669006346, 0.0047329277992248535, 0.00511078405380249, 0.005001215934753418, 0.004899839878082276, 0.004885536193847656, 0.0048875842094421385, 0.004921279907226563, 0.004781055927276612, 0.0047820801734924315, 0.0047257599830627444, 0.004723711967468262, 0.004730879783630371, 0.004760575771331787, 0.004764671802520752, 0.004720640182495117, 0.004729856014251709, 0.004947968006134033, 0.005404672145843506, 0.005214208126068115, 0.004891647815704346, 0.004947999954223633, 0.004888544082641602, 0.00487116813659668, 0.004884479999542236, 0.004869120121002197, 0.0048261117935180665, 0.004731904029846192, 0.004710464000701904, 0.004737984180450439, 0.004704256057739258, 0.004702208042144776, 0.004651008129119873, 0.004666368007659912, 0.004651008129119873, 0.004676608085632325, 0.00466534423828125, 0.010437631607055664, 0.004730879783630371, 0.004719615936279297, 0.004740096092224121, 0.004724736213684082, 0.004696063995361328, 0.0047226881980895995, 0.004681727886199951, 0.004754432201385498, 0.004880383968353271, 0.004775968074798584, 0.004718560218811035, 0.004743167877197266, 0.0047257599830627444, 0.004738048076629638, 0.0047216639518737795, 0.004709375858306885, 0.004723743915557861, 0.004702176094055176, 0.004683775901794434, 0.004696063995361328, 0.004670464038848877, 0.00466431999206543, 0.004670464038848877, 0.004680704116821289, 0.004683775901794434, 0.004775936126708984, 0.004703231811523437, 0.00471449613571167, 0.0047329277992248535, 0.004717567920684815, 0.0047288317680358885, 0.004730879783630371, 0.004727807998657227, 0.0047216639518737795, 0.0047257599830627444, 0.004708352088928222, 0.004746304035186767, 0.004717504024505615, 0.0047226881980895995, 0.0047185921669006346, 0.00470630407333374, 0.004757503986358643, 0.004766719818115234, 0.004890624046325683, 0.004884479999542236, 0.004827136039733886, 0.004746240139007568, 0.004717567920684815, 0.004711423873901367, 0.004724736213684082, 0.004712448120117187, 0.0049827837944030765, 0.004893695831298828, 0.004905983924865722, 0.004910079956054687, 0.004899839878082276, 0.0048455681800842285, 0.004742144107818603, 0.004873216152191162, 0.004899839878082276, 0.004739071846008301, 0.004771840095520019, 0.010399744033813477, 0.00466431999206543, 0.004731904029846192, 0.004729856014251709, 0.004727807998657227, 0.004733952045440673, 0.004715519905090332, 0.0047523841857910155, 0.004719615936279297, 0.004744192123413086, 0.004738048076629638, 0.004746240139007568, 0.00471449613571167, 0.0048865280151367185, 0.0048793601989746095, 0.0049203200340270994, 0.004900864124298096, 0.004908031940460205, 0.004794367790222168, 0.004733952045440673, 0.004733952045440673, 0.004761600017547608, 0.004743167877197266, 0.004715583801269531, 0.004723648071289063, 0.004742144107818603, 0.004733952045440673, 0.004729856014251709, 0.004736000061035156, 0.004723711967468262, 0.004740096092224121, 0.004750336170196533, 0.0047288317680358885, 0.004747263908386231, 0.004750336170196533, 0.004724736213684082, 0.004719615936279297, 0.004727807998657227, 0.0047185921669006346, 0.004745215892791748, 0.0046929922103881834, 0.004703231811523437, 0.0047185921669006346, 0.004742144107818603, 0.004744224071502686, 0.004711391925811767, 0.004702208042144776, 0.004736000061035156, 0.004897791862487793, 0.004978687763214112, 0.0048056321144104, 0.004728864192962647, 0.004722655773162842, 0.004727807998657227, 0.00471347188949585, 0.005137407779693603, 0.004897791862487793, 0.0049500160217285155, 0.0050022401809692385, 0.004836351871490479, 0.004742144107818603, 0.0047329277992248535, 0.004733952045440673, 0.010522624015808106, 0.004744192123413086, 0.004723711967468262, 0.004724736213684082, 0.00511078405380249, 0.004803584098815918, 0.004723711967468262, 0.004685823917388916, 0.004668416023254395, 0.004685823917388916, 0.004679679870605469, 0.00471347188949585, 0.004673535823822021, 0.004707327842712402, 0.004801536083221435, 0.004914175987243652, 0.004711423873901367, 0.004666368007659912, 0.0046888961791992185, 0.004676608085632325, 0.00470527982711792, 0.004668416023254395, 0.004701183795928955, 0.004676608085632325, 0.0046991357803344725, 0.0046592001914978025, 0.004689919948577881, 0.004687871932983399, 0.004666463851928711, 0.004686751842498779, 0.004687871932983399, 0.00467251205444336, 0.004744192123413086, 0.004683775901794434, 0.004674560070037842, 0.004680704116821289, 0.004811776161193848, 0.004963327884674072, 0.004897791862487793, 0.004914175987243652, 0.004894720077514648, 0.004726784229278564, 0.004759552001953125, 0.004758528232574463, 0.0047626237869262695, 0.004731904029846192, 0.004727807998657227, 0.0047554559707641605, 0.004743167877197266, 0.0047185921669006346, 0.004737023830413818, 0.004707327842712402, 0.004803616046905518, 0.004791264057159424, 0.004724800109863281, 0.004734911918640137, 0.004737023830413818, 0.004837376117706299, 0.0048261117935180665, 0.004795392036437988, 0.004779007911682129, 0.0047257599830627444, 0.004790272235870361, 0.010459136009216309, 0.004710400104522705, 0.004716544151306152, 0.0047288317680358885, 0.004702208042144776, 0.004761600017547608, 0.004717567920684815, 0.004702208042144776, 0.004712448120117187, 0.004738048076629638, 0.004739071846008301, 0.004715519905090332, 0.004717567920684815, 0.00471347188949585, 0.0048364481925964355, 0.004906911849975586, 0.00491212797164917, 0.0049500160217285155, 0.004859903812408447, 0.004754496097564697, 0.004736959934234619, 0.004736000061035156, 0.004733952045440673, 0.0047226881980895995, 0.004939775943756103, 0.005127168178558349, 0.004844543933868409, 0.004726784229278564, 0.004766719818115234, 0.004746240139007568, 0.004799488067626953, 0.0050135040283203125, 0.004931583881378174, 0.0049500160217285155, 0.004900864124298096, 0.004992000102996826, 0.004895743846893311, 0.0048865280151367185, 0.004740096092224121, 0.004968448162078858, 0.004898816108703613, 0.004873280048370361, 0.004921279907226563, 0.004872191905975342, 0.004895808219909668, 0.004861887931823731, 0.0047513599395751956, 0.004724736213684082, 0.004725791931152344, 0.004729824066162109, 0.0047329277992248535, 0.00475551986694336, 0.0047502717971801755, 0.004744192123413086, 0.004717567920684815, 0.004730879783630371, 0.004726784229278564, 0.00475648021697998, 0.004717567920684815, 0.004716544151306152, 0.004711423873901367, 0.004736000061035156, 0.0047226881980895995, 0.010531840324401855, 0.004824063777923584, 0.005174272060394287, 0.004899839878082276, 0.004911104202270508, 0.004883456230163574, 0.004872255802154541, 0.004823999881744385, 0.004720640182495117, 0.00474732780456543, 0.0047441282272338865, 0.004734975814819336, 0.004747263908386231, 0.004779007911682129, 0.0047288317680358885, 0.0047226881980895995, 0.0047216639518737795, 0.004747263908386231, 0.004717567920684815, 0.0047329277992248535, 0.004711423873901367, 0.004749311923980713, 0.004733952045440673, 0.004808703899383545, 0.004736000061035156, 0.004727807998657227, 0.004733952045440673, 0.004703231811523437, 0.004739071846008301, 0.004720640182495117, 0.004709375858306885, 0.0047257599830627444, 0.0047257599830627444, 0.004722752094268799, 0.004736959934234619, 0.00471347188949585, 0.004701183795928955, 0.004726784229278564, 0.0048056321144104, 0.004734975814819336, 0.004723711967468262, 0.004775936126708984, 0.004746240139007568, 0.004749311923980713, 0.004709375858306885, 0.004753407955169678, 0.005934080123901367, 0.004956160068511963, 0.004847616195678711, 0.0047636480331420894, 0.0047381119728088375, 0.004789184093475342, 0.0049090561866760255, 0.004831232070922851, 0.004872191905975342, 0.004898816108703613, 0.004935679912567138, 0.004898816108703613, 0.004893695831298828, 0.00488044786453247, 0.004753344058990479, 0.004768767833709717, 0.0047513599395751956, 0.010688511848449708, 0.0048793601989746095, 0.004749311923980713, 0.00465715217590332, 0.004708352088928222, 0.0050503678321838375, 0.004892672061920166, 0.00491315221786499, 0.004810751914978028, 0.004702208042144776, 0.004729856014251709, 0.004666431903839112, 0.004705215930938721, 0.004685823917388916, 0.00466534423828125, 0.004689919948577881, 0.004712448120117187, 0.004660223960876465, 0.004727807998657227, 0.004681727886199951, 0.004704256057739258, 0.004978687763214112, 0.004765696048736572, 0.004779007911682129, 0.004739071846008301, 0.004754432201385498, 0.004741119861602783, 0.004682752132415771, 0.004694015979766846, 0.004701183795928955, 0.004679679870605469, 0.004670464038848877, 0.004689919948577881, 0.004787199974060059, 0.004848639965057373, 0.004850687980651855, 0.004731904029846192, 0.004746240139007568, 0.004758592128753662, 0.004777919769287109, 0.004737023830413818, 0.004737023830413818, 0.004733952045440673, 0.004774911880493164, 0.004764671802520752, 0.00475648021697998, 0.004749311923980713, 0.004734975814819336, 0.004768767833709717, 0.004745215892791748, 0.004759552001953125, 0.0048895998001098635, 0.004884479999542236, 0.004947968006134033, 0.0048793601989746095, 0.004897791862487793, 0.004977663993835449, 0.004829184055328369, 0.004835328102111816, 0.004700160026550293, 0.00467251205444336, 0.004702208042144776, 0.00467251205444336, 0.01041817569732666, 0.004671487808227539, 0.004690944194793701, 0.004674560070037842, 0.004827136039733886, 0.004694015979766846, 0.00467251205444336, 0.004683775901794434, 0.004682752132415771, 0.004675583839416504, 0.00469708776473999, 0.004746240139007568, 0.004824063777923584, 0.004679679870605469, 0.004748288154602051, 0.004695040225982666, 0.0046888961791992185, 0.0046991357803344725, 0.004680704116821289, 0.004679679870605469, 0.00469708776473999, 0.004671487808227539, 0.004716544151306152, 0.004670464038848877, 0.004670464038848877, 0.004703231811523437, 0.004686848163604736, 0.004667391777038574, 0.004770815849304199, 0.004679679870605469, 0.004745215892791748, 0.0047226881980895995, 0.004739071846008301, 0.004733952045440673, 0.0047216639518737795, 0.004745215892791748, 0.0047288317680358885, 0.004723711967468262, 0.004709375858306885, 0.004700160026550293, 0.004682784080505371, 0.004676576137542725, 0.004749311923980713, 0.004724736213684082, 0.004678656101226806, 0.00469708776473999, 0.00469708776473999, 0.004680704116821289, 0.004703231811523437, 0.004677631855010986, 0.004668416023254395, 0.004776959896087647, 0.004676608085632325, 0.004684800148010254, 0.004687871932983399, 0.004663296222686767, 0.0047185921669006346, 0.00468179178237915, 0.004667327880859375, 0.004695040225982666, 0.004684800148010254, 0.004687871932983399, 0.004686912059783935, 0.010416064262390137, 0.004710464000701904, 0.004719552040100098, 0.004660223960876465, 0.0046694397926330565, 0.004680704116821289, 0.00467251205444336, 0.004671487808227539, 0.004702208042144776, 0.0046705279350280765, 0.004786143779754639, 0.0048475837707519535, 0.004947968006134033, 0.004931583881378174, 0.004899839878082276, 0.004910079956054687, 0.004929535865783692, 0.004923391819000244, 0.0048455681800842285, 0.004739071846008301, 0.0047861762046813965, 0.00469708776473999, 0.004719615936279297, 0.0047636480331420894, 0.004724736213684082, 0.004744192123413086, 0.004726784229278564, 0.004719615936279297, 0.0047636480331420894, 0.00480460786819458, 0.004925439834594727, 0.004892672061920166, 0.00487116813659668, 0.004843520164489746, 0.004733952045440673, 0.004761600017547608, 0.004737023830413818, 0.004737023830413818, 0.0047820801734924315, 0.004698112010955811, 0.004716544151306152, 0.004702208042144776, 0.004681727886199951, 0.004736032009124756, 0.004700128078460693, 0.004860928058624267, 0.004863999843597412, 0.004748320102691651, 0.004767712116241455, 0.005153791904449463, 0.0055101442337036136, 0.005029888153076172, 0.004966464042663574, 0.004941760063171386, 0.004923391819000244, 0.005024767875671386, 0.004895743846893311, 0.00481385612487793, 0.004782048225402832, 0.004744192123413086, 0.0047636480331420894, 0.004768767833709717, 0.0047216639518737795, 0.010638336181640624, 0.004768767833709717, 0.004715519905090332, 0.00471347188949585, 0.004733952045440673, 0.0047309441566467285, 0.004753344058990479, 0.004686848163604736, 0.004652031898498535, 0.004676608085632325, 0.004681727886199951, 0.00481382417678833, 0.00469923210144043, 0.004673439979553223, 0.004694015979766846, 0.004666368007659912, 0.004682752132415771, 0.004675583839416504, 0.004692031860351563, 0.004800447940826416, 0.004673535823822021, 0.004715519905090332, 0.004702208042144776, 0.004673535823822021, 0.004702208042144776, 0.004670464038848877, 0.00466534423828125, 0.0046929922103881834, 0.004690944194793701, 0.0046694397926330565, 0.004931583881378174, 0.004784128189086914, 0.004775936126708984, 0.004682752132415771, 0.004675583839416504, 0.0046991357803344725, 0.004666368007659912, 0.004682752132415771, 0.00466534423828125, 0.004684800148010254, 0.004759552001953125, 0.004694015979766846, 0.004700160026550293, 0.0046663999557495115, 0.004678624153137207, 0.00469708776473999, 0.00466534423828125, 0.004695040225982666, 0.004685823917388916, 0.004681727886199951, 0.004702208042144776, 0.004667391777038574, 0.0047523841857910155, 0.004884479999542236, 0.004837376117706299, 0.004811776161193848, 0.004673535823822021, 0.004689919948577881, 0.0046888961791992185, 0.004670464038848877, 0.004707327842712402, 0.004698112010955811, 0.004687871932983399, 0.010593279838562012, 0.004729856014251709, 0.004839424133300781, 0.004784128189086914, 0.004789247989654541, 0.00487116813659668, 0.004911104202270508, 0.004917247772216797, 0.004834303855895996, 0.004758528232574463, 0.004746240139007568, 0.004768767833709717, 0.0047257599830627444, 0.0047523841857910155, 0.004738048076629638, 0.004726784229278564, 0.004724736213684082, 0.004744192123413086, 0.004742144107818603, 0.004731967926025391, 0.004727744102478028, 0.00470630407333374, 0.0048230400085449215, 0.004738048076629638, 0.004774911880493164, 0.004712480068206787, 0.004730847835540772, 0.004758528232574463, 0.0047288317680358885, 0.004769792079925537, 0.004788224220275879, 0.004734975814819336, 0.0048158721923828125, 0.0047216639518737795, 0.004726784229278564, 0.004734975814819336, 0.0047329277992248535, 0.004742144107818603, 0.004744192123413086, 0.0047216639518737795, 0.004773888111114502, 0.004744192123413086, 0.004780032157897949, 0.004790272235870361, 0.004726784229278564, 0.0047820801734924315, 0.004715519905090332, 0.004741119861602783, 0.004717567920684815, 0.004739071846008301, 0.004759552001953125, 0.004716544151306152, 0.004729856014251709, 0.00471347188949585, 0.004740096092224121, 0.004753407955169678, 0.004760575771331787, 0.004727807998657227, 0.004742144107818603, 0.004832255840301514, 0.004867072105407715, 0.004784128189086914, 0.004738048076629638, 0.010537983894348145, 0.0047933440208435055, 0.005203968048095703, 0.005061632156372071, 0.005014527797698975, 0.004895743846893311, 0.004816895961761475, 0.004743167877197266, 0.005340159893035889, 0.005588992118835449, 0.0049387521743774416, 0.004936704158782959, 0.004965375900268554, 0.004914175987243652, 0.00491315221786499, 0.004892672061920166, 0.004887551784515381, 0.0049459199905395506, 0.004891647815704346, 0.004902912139892578, 0.004910079956054687, 0.0049162240028381345, 0.004835328102111816, 0.004829184055328369, 0.0048496642112731934, 0.004889632225036621, 0.004927455902099609, 0.00487116813659668, 0.004874303817749024, 0.0048782720565795894, 0.004801536083221435, 0.0047513599395751956, 0.004738048076629638, 0.004969471931457519, 0.004736000061035156, 0.005091328144073487, 0.004921343803405762, 0.004839424133300781, 0.004893695831298828, 0.004730879783630371, 0.004759552001953125, 0.004744192123413086, 0.0047216639518737795, 0.004730879783630371, 0.004772863864898682, 0.005049344062805176, 0.004874239921569825, 0.004902912139892578, 0.0049162240028381345, 0.005070847988128662, 0.004873216152191162, 0.004881408214569092, 0.004892672061920166, 0.005047296047210693, 0.0049489917755126955, 0.004895743846893311, 0.004902912139892578, 0.004876287937164306, 0.0047554559707641605, 0.0047288317680358885, 0.004730879783630371, 0.004736000061035156, 0.004720640182495117, 0.01053286361694336, 0.004983808040618896, 0.004923391819000244, 0.004882431983947754, 0.004843520164489746, 0.004891647815704346, 0.00467251205444336, 0.0047513599395751956, 0.004704256057739258, 0.004673535823822021, 0.004729856014251709, 0.004715519905090332, 0.004726784229278564, 0.0047288317680358885, 0.004741119861602783, 0.004680704116821289, 0.004663296222686767, 0.004736000061035156, 0.004668416023254395, 0.004704256057739258, 0.004689919948577881, 0.0048261117935180665, 0.0049530878067016604, 0.004836351871490479, 0.0048793601989746095, 0.0048752641677856446, 0.004936704158782959, 0.004918272018432617, 0.004902912139892578, 0.004788224220275879, 0.004670464038848877, 0.00471347188949585, 0.004809728145599365, 0.0048865280151367185, 0.004927487850189209, 0.004927487850189209, 0.004905983924865722, 0.004893695831298828, 0.004929535865783692, 0.004893695831298828, 0.0048855037689208985, 0.004901887893676758, 0.004892672061920166, 0.004937727928161621, 0.004899839878082276, 0.004884479999542236, 0.0048865280151367185, 0.0049489917755126955, 0.0047861762046813965, 0.0047216639518737795, 0.004728864192962647, 0.0047103681564331056, 0.004727807998657227, 0.004734975814819336, 0.0047216639518737795, 0.004907008171081543, 0.004986879825592041, 0.004836351871490479, 0.0047216639518737795, 0.004776959896087647, 0.004733952045440673, 0.004743167877197266, 0.004739071846008301, 0.010884096145629883, 0.004883456230163574, 0.004895743846893311, 0.004863999843597412, 0.004898816108703613, 0.004876287937164306, 0.004929535865783692, 0.004895743846893311, 0.0048793601989746095, 0.004911104202270508, 0.004911104202270508, 0.004941823959350586, 0.004868095874786377, 0.004716544151306152, 0.0047329277992248535, 0.004724736213684082, 0.004846591949462891, 0.004730879783630371, 0.004726784229278564, 0.0048855037689208985, 0.004924416065216064, 0.004894720077514648, 0.0049530878067016604, 0.0048855037689208985, 0.0049090561866760255, 0.004923391819000244, 0.004915200233459473, 0.004893695831298828, 0.004901887893676758, 0.004918272018432617, 0.004820991992950439, 0.004740096092224121, 0.004754432201385498, 0.004731904029846192, 0.004736000061035156, 0.004733952045440673, 0.004738048076629638, 0.004803584098815918, 0.004740096092224121, 0.00471347188949585, 0.004741119861602783, 0.004841472148895264, 0.004887551784515381, 0.004902912139892578, 0.0049459199905395506, 0.0049489917755126955, 0.004869120121002197, 0.004933631896972656, 0.0048855037689208985, 0.004888576030731201, 0.004824063777923584, 0.004731904029846192, 0.0047185921669006346, 0.0047216639518737795, 0.004738048076629638, 0.004719615936279297, 0.00471449613571167, 0.004740096092224121, 0.0047226881980895995, 0.0047226881980895995, 0.004744192123413086, 0.0047216639518737795, 0.004767744064331054, 0.010810367584228516, 0.00487116813659668, 0.004917247772216797, 0.004883456230163574, 0.004910079956054687, 0.0047933440208435055, 0.004733952045440673, 0.004859903812408447, 0.004881408214569092, 0.004750336170196533, 0.004907008171081543, 0.004750336170196533, 0.0047185921669006346, 0.0047185921669006346, 0.0047513599395751956, 0.004723775863647461, 0.004730815887451172, 0.004736000061035156, 0.004754432201385498, 0.004801536083221435, 0.0048220157623291016, 0.005091328144073487, 0.0057794561386108395, 0.0062044157981872555, 0.005454847812652588, 0.005227519989013672, 0.004931583881378174, 0.004896768093109131, 0.0049090561866760255, 0.004944896221160889, 0.004895743846893311, 0.0049090561866760255, 0.004892672061920166, 0.004946944236755371, 0.004881408214569092, 0.004891647815704346, 0.00491212797164917, 0.004890624046325683, 0.004951039791107178, 0.004902976036071777, 0.004759488105773926, 0.004730879783630371, 0.004738048076629638, 0.004748288154602051, 0.004674560070037842, 0.004669536113739014, 0.0047010879516601566, 0.004683775901794434, 0.004690944194793701, 0.004708352088928222, 0.0046919679641723635, 0.00467251205444336, 0.0047288317680358885, 0.004739071846008301, 0.0047513599395751956, 0.004745215892791748, 0.004739071846008301, 0.004737023830413818, 0.004736000061035156, 0.004730879783630371, 0.004741119861602783, 0.0047647361755371095, 0.004717504024505615, 0.010569727897644043, 0.004720640182495117, 0.004717599868774414, 0.004740064144134522, 0.004724736213684082, 0.004731904029846192, 0.004712448120117187, 0.0047226881980895995, 0.004731904029846192, 0.004740096092224121, 0.0047226881980895995, 0.004740096092224121, 0.004715519905090332, 0.004731904029846192, 0.004736000061035156, 0.0047185921669006346, 0.004743167877197266, 0.0047216639518737795, 0.00471347188949585, 0.004727807998657227, 0.004719615936279297, 0.0047933440208435055, 0.0047636480331420894, 0.004676608085632325, 0.004679679870605469, 0.0046694397926330565, 0.004670464038848877, 0.004731904029846192, 0.0047226881980895995, 0.004743167877197266, 0.004742144107818603, 0.004734975814819336, 0.004719615936279297, 0.004748288154602051, 0.004733952045440673, 0.004745215892791748, 0.004690944194793701, 0.00466431999206543, 0.004696063995361328, 0.0046929922103881834, 0.004678656101226806, 0.0046694397926330565, 0.00467251205444336, 0.004673535823822021, 0.004704256057739258, 0.004684800148010254, 0.004681727886199951, 0.004673535823822021, 0.004709375858306885, 0.004673535823822021, 0.004773888111114502, 0.004671487808227539, 0.004663296222686767, 0.004673535823822021, 0.0046929922103881834, 0.004684800148010254, 0.004707327842712402, 0.004663296222686767, 0.004740096092224121, 0.0047288317680358885, 0.004682752132415771, 0.004680768013000488, 0.004665279865264892, 0.010347519874572754, 0.004683775901794434, 0.004701183795928955, 0.004724736213684082, 0.004734975814819336, 0.004736000061035156, 0.004736000061035156, 0.004729856014251709, 0.0047185921669006346, 0.004726784229278564, 0.0047288317680358885, 0.004747263908386231, 0.0047216639518737795, 0.004758528232574463, 0.004717567920684815, 0.0047226881980895995, 0.004715519905090332, 0.004719615936279297, 0.004703231811523437, 0.004670464038848877, 0.00466534423828125, 0.004679679870605469, 0.004679711818695068, 0.004670432090759277, 0.0046991357803344725, 0.004683775901794434, 0.004696063995361328, 0.004679679870605469, 0.00470630407333374, 0.004690976142883301, 0.004702176094055176, 0.0046888961791992185, 0.00468998384475708, 0.004689856052398682, 0.004694015979766846, 0.004680704116821289, 0.004676608085632325, 0.004675583839416504, 0.00467251205444336, 0.004779007911682129, 0.0047226881980895995, 0.00471449613571167, 0.004737023830413818, 0.004716544151306152, 0.004724736213684082, 0.004771840095520019, 0.004716544151306152, 0.004729856014251709, 0.00471347188949585, 0.0047226881980895995, 0.0047288317680358885, 0.00467251205444336, 0.00466534423828125, 0.004729856014251709, 0.004674560070037842, 0.004679679870605469, 0.0046991357803344725, 0.004680704116821289, 0.0046991357803344725, 0.004690944194793701, 0.004671487808227539, 0.004689919948577881, 0.004726784229278564, 0.010350591659545898, 0.004685855865478515, 0.00467964792251587, 0.004666368007659912, 0.0046592001914978025, 0.004681727886199951, 0.004671487808227539, 0.004661248207092285, 0.004675583839416504, 0.0046694397926330565, 0.004712448120117187, 0.004689919948577881, 0.004667424201965332, 0.0046581439971923825, 0.004670464038848877, 0.004685823917388916, 0.004677631855010986, 0.004701183795928955, 0.004678688049316406, 0.004673503875732422, 0.0046929922103881834, 0.004660223960876465, 0.0046991357803344725, 0.004701183795928955, 0.0047257599830627444, 0.004729856014251709, 0.00471449613571167, 0.00470630407333374, 0.004775936126708984, 0.0047257599830627444, 0.004709375858306885, 0.0047226881980895995, 0.004720640182495117, 0.00474729585647583, 0.0047175359725952146, 0.004731904029846192, 0.004737023830413818, 0.0047350401878356934, 0.004715456008911133, 0.004744224071502686, 0.004740064144134522, 0.004724736213684082, 0.004730879783630371, 0.004727807998657227, 0.004746240139007568, 0.004753407955169678, 0.004712448120117187, 0.004717567920684815, 0.004733952045440673, 0.004719615936279297, 0.004730879783630371, 0.004724736213684082, 0.004738048076629638, 0.004743167877197266, 0.004754432201385498, 0.004745215892791748, 0.004729856014251709, 0.004710400104522705, 0.0047288317680358885, 0.0047257599830627444, 0.004711423873901367, 0.004776959896087647, 0.004715519905090332, 0.010396672248840331, 0.004747263908386231, 0.0047185921669006346, 0.004731904029846192, 0.0047185921669006346, 0.00475648021697998, 0.004753407955169678, 0.004727807998657227, 0.004711423873901367, 0.0047185921669006346, 0.004734975814819336, 0.004710400104522705, 0.004731904029846192, 0.004727807998657227, 0.004720640182495117, 0.004734975814819336, 0.004747263908386231, 0.0047381119728088375, 0.00472160005569458, 0.0047216639518737795, 0.004739071846008301, 0.004780032157897949, 0.004726784229278564, 0.004738048076629638, 0.004717567920684815, 0.004726784229278564, 0.004749311923980713, 0.004740096092224121, 0.0047523841857910155, 0.0047288317680358885, 0.004742144107818603, 0.004729856014251709, 0.004731904029846192, 0.004734975814819336, 0.0047626237869262695, 0.004726784229278564, 0.004719615936279297, 0.004723711967468262, 0.004739071846008301, 0.004765696048736572, 0.004737023830413818, 0.0048056321144104, 0.0046991357803344725, 0.004685823917388916, 0.004684800148010254, 0.0046919679641723635, 0.004679679870605469, 0.004674560070037842, 0.004673535823822021, 0.004661248207092285, 0.004702208042144776, 0.0047185921669006346, 0.004729856014251709, 0.004719615936279297, 0.004733952045440673, 0.0047329277992248535, 0.0047288317680358885, 0.004671487808227539, 0.0046622719764709475, 0.004747263908386231, 0.004680736064910889, 0.00467248010635376, 0.004689919948577881, 0.010457088470458984, 0.0047288317680358885, 0.004736000061035156, 0.004720640182495117, 0.004720640182495117, 0.00481382417678833, 0.004753407955169678, 0.004727839946746826, 0.004730847835540772, 0.004717567920684815, 0.004668416023254395, 0.004685823917388916, 0.00466534423828125, 0.004695040225982666, 0.004687871932983399, 0.004690944194793701, 0.004679679870605469, 0.004676608085632325, 0.004671487808227539, 0.004678656101226806, 0.0046929922103881834, 0.0046581759452819825, 0.0046592001914978025, 0.004680704116821289, 0.004667391777038574, 0.004701183795928955, 0.004681727886199951, 0.004720640182495117, 0.004737023830413818, 0.004744192123413086, 0.0047524161338806156, 0.0047216320037841795, 0.004720672130584717, 0.004727776050567627, 0.0047626237869262695, 0.004712448120117187, 0.004749343872070312, 0.004727776050567627, 0.004727807998657227, 0.004700160026550293, 0.004677631855010986, 0.004686848163604736, 0.0046694397926330565, 0.004660223960876465, 0.00466534423828125, 0.004674560070037842, 0.004724736213684082, 0.00469708776473999, 0.004655104160308838, 0.0046888961791992185, 0.004671487808227539, 0.004675583839416504, 0.004675583839416504, 0.004704256057739258, 0.004667391777038574, 0.0046622719764709475, 0.004660223960876465, 0.0046888961791992185, 0.004698112010955811, 0.004698112010955811, 0.004676608085632325, 0.004675583839416504, 0.004686848163604736, 0.010459136009216309, 0.004709375858306885, 0.004726816177368164, 0.004743135929107666, 0.004724736213684082, 0.004742144107818603, 0.0047185921669006346, 0.004704256057739258, 0.004727807998657227, 0.004766719818115234, 0.004744192123413086, 0.0047185921669006346, 0.00471347188949585, 0.004749311923980713, 0.004734975814819336, 0.004712448120117187, 0.004730879783630371, 0.004716544151306152, 0.004739071846008301, 0.00475648021697998, 0.0047185921669006346, 0.004742144107818603, 0.004729856014251709, 0.0047329277992248535, 0.0047185921669006346, 0.004794367790222168, 0.004720640182495117, 0.004727807998657227, 0.004724736213684082, 0.004723711967468262, 0.004729856014251709, 0.004719615936279297, 0.004739071846008301, 0.004724736213684082, 0.004731904029846192, 0.004723711967468262, 0.004737023830413818, 0.004726784229278564, 0.004750336170196533, 0.0047216639518737795, 0.0047216639518737795, 0.004729856014251709, 0.004727807998657227, 0.004736000061035156, 0.00470630407333374, 0.004708352088928222, 0.004716544151306152, 0.004760575771331787, 0.004703231811523437, 0.004734975814819336, 0.004736000061035156, 0.004680704116821289, 0.004678656101226806, 0.004660223960876465, 0.004651008129119873, 0.004717567920684815, 0.004677631855010986, 0.004677631855010986, 0.004675583839416504, 0.004684800148010254, 0.004702208042144776, 0.004689919948577881, 0.004668416023254395, 0.01031987190246582, 0.004695040225982666, 0.00464899206161499, 0.004662240028381347, 0.004663296222686767, 0.004661248207092285, 0.004651008129119873, 0.004679679870605469, 0.004682752132415771, 0.0046581759452819825, 0.004651008129119873, 0.0046592001914978025, 0.004652031898498535, 0.00469708776473999, 0.004883456230163574, 0.004989952087402344, 0.004891647815704346, 0.004900864124298096, 0.004921343803405762, 0.004764671802520752, 0.0047257599830627444, 0.004719615936279297, 0.004716544151306152, 0.004726784229278564, 0.004719615936279297, 0.004704256057739258, 0.004731904029846192, 0.004717567920684815, 0.004749311923980713, 0.004681727886199951, 0.00467251205444336, 0.0046888961791992185, 0.004670464038848877, 0.00466534423828125, 0.004698112010955811, 0.004671520233154297, 0.004726751804351807, 0.004666368007659912, 0.004682752132415771, 0.004668416023254395, 0.0046929922103881834, 0.004667424201965332, 0.00467251205444336, 0.00467964792251587, 0.004683775901794434, 0.004655104160308838, 0.004685823917388916, 0.004661248207092285, 0.004678656101226806, 0.004698112010955811, 0.004680704116821289, 0.00466534423828125, 0.00469708776473999, 0.004670464038848877, 0.004660223960876465, 0.0046694397926330565, 0.004645887851715088, 0.004655104160308838, 0.004687871932983399, 0.0046561279296875, 0.00467251205444336, 0.00471347188949585, 0.0047329277992248535, 0.010312704086303711, 0.004653056144714355, 0.004675583839416504, 0.0046694397926330565, 0.004690944194793701, 0.0046622719764709475, 0.004667391777038574, 0.00465718412399292, 0.004688864231109619, 0.004676608085632325, 0.004677631855010986, 0.004684800148010254, 0.004671487808227539, 0.004674560070037842, 0.004683775901794434, 0.004670464038848877, 0.004680704116821289, 0.004684800148010254, 0.00466431999206543, 0.004670464038848877, 0.004676608085632325, 0.0046581759452819825, 0.004694015979766846, 0.00466534423828125, 0.00467251205444336, 0.00471347188949585, 0.0046694397926330565, 0.00466537618637085, 0.0046919360160827634, 0.004671487808227539, 0.004684832096099854, 0.004683743953704834, 0.004663392066955567, 0.004667295932769776, 0.004667391777038574, 0.00465715217590332, 0.004660223960876465, 0.004682752132415771, 0.004687871932983399, 0.004695040225982666, 0.0046694397926330565, 0.004671487808227539, 0.004660223960876465, 0.004677631855010986, 0.004671487808227539, 0.004679711818695068, 0.0046560959815979, 0.0046592001914978025, 0.004660223960876465, 0.004728896141052246, 0.004678656101226806, 0.004703167915344238, 0.004671487808227539, 0.004661248207092285, 0.004675583839416504, 0.004674560070037842, 0.00466534423828125, 0.004680704116821289, 0.005478432178497314, 0.004813792228698731, 0.0047288317680358885, 0.0047216639518737795, 0.004737023830413818, 0.010397695541381835, 0.004719615936279297, 0.004750336170196533, 0.004735007762908936, 0.004764639854431153, 0.004668416023254395, 0.0046991357803344725, 0.0046919679641723635, 0.004687871932983399, 0.004678656101226806, 0.004680704116821289, 0.004708352088928222, 0.004741119861602783, 0.0047626237869262695, 0.004771840095520019, 0.004734975814819336, 0.004738048076629638, 0.0047257599830627444, 0.004741119861602783, 0.004744192123413086, 0.004730879783630371, 0.004719615936279297, 0.004738048076629638, 0.004727807998657227, 0.004753407955169678, 0.004745215892791748, 0.004716544151306152, 0.0047216639518737795, 0.004729856014251709, 0.004776959896087647, 0.004723711967468262, 0.0047226881980895995, 0.004724736213684082, 0.0047288317680358885, 0.004734975814819336, 0.004759552001953125, 0.0047257599830627444, 0.004738048076629638, 0.0048558077812194825, 0.0047523841857910155, 0.004744192123413086, 0.0047216639518737795, 0.004670464038848877, 0.004679679870605469, 0.004681727886199951, 0.004675583839416504, 0.00467251205444336, 0.00467251205444336, 0.0046622719764709475, 0.00467251205444336, 0.004674560070037842, 0.004670464038848877, 0.00470527982711792, 0.004731904029846192, 0.004716544151306152, 0.004739071846008301, 0.004701183795928955, 0.004694015979766846, 0.004663296222686767, 0.00470630407333374, 0.004720640182495117, 0.004734975814819336, 0.004716544151306152, 0.010479616165161134, 0.004733952045440673, 0.004715583801269531, 0.004764607906341553, 0.004708352088928222, 0.004719615936279297, 0.004717567920684815, 0.004724736213684082, 0.004712448120117187, 0.004738048076629638, 0.00471347188949585, 0.004715519905090332, 0.004731904029846192, 0.0047523841857910155, 0.004746240139007568, 0.004710400104522705, 0.004716544151306152, 0.004761600017547608, 0.004726784229278564, 0.004717567920684815, 0.0047288317680358885, 0.0047216639518737795, 0.00470630407333374, 0.004708352088928222, 0.0047216639518737795, 0.004744192123413086, 0.004716544151306152, 0.004710400104522705, 0.004739071846008301, 0.004727807998657227, 0.004716544151306152, 0.004743167877197266, 0.004719615936279297, 0.004724736213684082, 0.004737023830413818, 0.004723711967468262, 0.004750336170196533, 0.004799488067626953, 0.004724736213684082, 0.004759552001953125, 0.004721727848052979, 0.00472160005569458, 0.0047226881980895995, 0.004727807998657227, 0.004900864124298096, 0.004761600017547608, 0.004717567920684815, 0.004743167877197266, 0.004720640182495117, 0.0047329277992248535, 0.004723711967468262, 0.004712512016296386, 0.004728767871856689, 0.004726784229278564, 0.004709375858306885, 0.004727807998657227, 0.004753407955169678, 0.004715519905090332, 0.004739071846008301, 0.0047923197746276855, 0.004716544151306152, 0.004741119861602783, 0.0047216639518737795]",tokens/s,205.28452672220206,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,google/gemma-2b,google/gemma-2b,cuda,0,42,,,True,,,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.223-212.873.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.0,,,,1.21.4,,,,0.12.0,,,gemma,MB,1540.071424,6797.39392,0.0,6150.946816,5419.87328,s,10,0.3088504314422607,0.03088504314422607,4.952747365149912e-05,0.03087827205657959,0.030958012199401856,0.0309631817817688,0.030967317447662353,"[0.03096835136413574, 0.030920127868652343, 0.0308973445892334, 0.030860927581787108, 0.030956863403320312, 0.030827072143554686, 0.030853376388549805, 0.030891584396362304, 0.030864959716796876, 0.030809823989868163]",tokens/s,8288.80176092158,kWh,3.6415854097007713e-07,1.9951573763366047e-07,1.881762189169245e-06,2.445436467772982e-06,tokens/kWh,104684788.73758471,MB,1592.795136,6797.39392,0.0,6150.946816,5419.87584,s,10,12.881293212890625,1.2881293212890625,0.017658481033884246,1.2900848388671875,1.3068117919921876,1.3107702758789062,1.3139370629882812,"[1.30593212890625, 1.314728759765625, 1.3011173095703126, 1.3026669921875, 1.3007484130859375, 1.2719969482421876, 1.2794212646484375, 1.2726861572265624, 1.2656256103515624, 1.26636962890625]",tokens/s,48.90813287050586,kWh,1.5118707867710757e-05,8.283590300433077e-06,5.1232085857430134e-05,7.463438402557397e-05,tokens/kWh,844114.9588427317,,s,630,12.879658994674662,0.020443903166150287,0.0004387009483782593,0.02025164794921875,0.020876287460327148,0.020972851848602293,0.021762406692504892,"[0.02006118392944336, 0.02003865623474121, 0.0200581111907959, 0.02003865623474121, 0.020031488418579102, 0.02004787254333496, 0.020008960723876954, 0.02047590446472168, 0.020583423614501953, 0.020750335693359375, 0.02087424087524414, 0.02084249687194824, 0.02078620719909668, 0.021024736404418945, 0.020892671585083008, 0.0218470401763916, 0.022180864334106445, 0.021176319122314453, 0.020888576507568358, 0.020969472885131835, 0.020801536560058592, 0.020918272018432618, 0.02080460739135742, 0.020080703735351563, 0.02018707275390625, 0.020808704376220705, 0.02002227210998535, 0.02004377555847168, 0.020090879440307616, 0.020466751098632812, 0.020625343322753908, 0.020754432678222655, 0.020847616195678712, 0.021104639053344726, 0.021019647598266602, 0.020975616455078124, 0.020759552001953126, 0.020730880737304686, 0.021111808776855468, 0.02081279945373535, 0.020757503509521484, 0.02089574432373047, 0.02090188789367676, 0.020777984619140624, 0.020736000061035157, 0.0214968318939209, 0.020876287460327148, 0.02083737564086914, 0.020749311447143554, 0.020738048553466795, 0.02083328056335449, 0.020783103942871094, 0.020800512313842775, 0.020739072799682616, 0.020943872451782225, 0.020737024307250978, 0.02083430480957031, 0.020805631637573242, 0.020768768310546876, 0.020770816802978515, 0.020766719818115235, 0.020900863647460938, 0.020768768310546876, 0.023553024291992186, 0.021456895828247072, 0.020786176681518553, 0.020746240615844725, 0.020816896438598635, 0.020926464080810548, 0.020765695571899414, 0.020748287200927733, 0.0207554874420166, 0.021201887130737306, 0.020936704635620116, 0.020750335693359375, 0.020726783752441406, 0.020806655883789063, 0.020925439834594727, 0.020780031204223632, 0.022024192810058595, 0.02082815933227539, 0.020809728622436522, 0.021320703506469727, 0.0208721923828125, 0.020784128189086915, 0.02082815933227539, 0.020858879089355468, 0.020724735260009765, 0.020723712921142577, 0.020780031204223632, 0.02069811248779297, 0.020818975448608397, 0.020742111206054688, 0.020798463821411133, 0.020808704376220705, 0.02088960075378418, 0.020915199279785156, 0.02082099151611328, 0.020739072799682616, 0.02064076805114746, 0.020772863388061523, 0.020722688674926756, 0.020736000061035157, 0.020742143630981445, 0.020768768310546876, 0.020724735260009765, 0.020738048553466795, 0.020676607131958007, 0.020931583404541015, 0.020730880737304686, 0.020813823699951172, 0.020706304550170897, 0.020762624740600585, 0.020748287200927733, 0.020814847946166993, 0.020754432678222655, 0.021013504028320314, 0.020956159591674805, 0.020797439575195312, 0.020632575988769532, 0.020692991256713866, 0.020477951049804686, 0.020571136474609376, 0.020726783752441406, 0.020732927322387695, 0.020710399627685547, 0.020853759765625, 0.02081177520751953, 0.020719615936279297, 0.020977664947509765, 0.02083737564086914, 0.020748287200927733, 0.020793344497680662, 0.02087424087524414, 0.020682752609252928, 0.02085785675048828, 0.020720640182495118, 0.020767744064331056, 0.020978687286376953, 0.020771839141845702, 0.020774911880493165, 0.020740095138549804, 0.020814847946166993, 0.02071660804748535, 0.020794303894042968, 0.020694015502929687, 0.02071347236633301, 0.020746271133422853, 0.020882400512695312, 0.020749311447143554, 0.020699136734008788, 0.020784128189086915, 0.020736000061035157, 0.020794368743896483, 0.020798463821411133, 0.019919872283935547, 0.019974143981933593, 0.02001203155517578, 0.020136959075927736, 0.02003455924987793, 0.020051967620849608, 0.02003046417236328, 0.020305919647216796, 0.02091929626464844, 0.02084556770324707, 0.020618240356445314, 0.02048307228088379, 0.02082815933227539, 0.020961280822753905, 0.020762624740600585, 0.020743167877197266, 0.02087833595275879, 0.02107187271118164, 0.020793344497680662, 0.02081996726989746, 0.020823040008544923, 0.020791296005249024, 0.020051967620849608, 0.02008166313171387, 0.02006118392944336, 0.02003455924987793, 0.02068992042541504, 0.020703231811523438, 0.02086911964416504, 0.020791296005249024, 0.02083020782470703, 0.020732927322387695, 0.020733951568603515, 0.020734975814819336, 0.020075519561767577, 0.02006937599182129, 0.020098079681396486, 0.020055007934570313, 0.020730880737304686, 0.020784128189086915, 0.02088960075378418, 0.02007040023803711, 0.02023628807067871, 0.02099404716491699, 0.020725759506225586, 0.020789247512817383, 0.02074015998840332, 0.020814783096313478, 0.020764671325683593, 0.020763647079467772, 0.02069811248779297, 0.020813823699951172, 0.020776960372924806, 0.020547584533691408, 0.02083430480957031, 0.02084351921081543, 0.020922367095947265, 0.02092748832702637, 0.020822015762329102, 0.020780031204223632, 0.020770816802978515, 0.020779008865356444, 0.020750335693359375, 0.021555200576782226, 0.021287935256958008, 0.020959232330322267, 0.02079641532897949, 0.020890623092651366, 0.020822015762329102, 0.020822015762329102, 0.021053440093994142, 0.020775936126708985, 0.02084556770324707, 0.020802560806274413, 0.020883455276489257, 0.020059135437011717, 0.020059135437011717, 0.020065280914306642, 0.020017152786254884, 0.020074495315551756, 0.020076543807983398, 0.02004275131225586, 0.02007142448425293, 0.020175872802734376, 0.02042163276672363, 0.020789247512817383, 0.02084147262573242, 0.02086604881286621, 0.02088960075378418, 0.022373376846313478, 0.020805631637573242, 0.02084659194946289, 0.02083020782470703, 0.02082508850097656, 0.02081177520751953, 0.02066534423828125, 0.020534271240234374, 0.020205568313598633, 0.02011238479614258, 0.020063232421875, 0.02063667106628418, 0.020800512313842775, 0.020789247512817383, 0.020788223266601562, 0.020737024307250978, 0.02003046417236328, 0.02004684829711914, 0.020009983062744142, 0.02004582405090332, 0.02043801689147949, 0.020880384445190428, 0.020771839141845702, 0.020758527755737305, 0.020572160720825194, 0.02088960075378418, 0.020770847320556642, 0.020789215087890625, 0.020733951568603515, 0.02087424087524414, 0.02079539108276367, 0.02008780860900879, 0.020007936477661133, 0.020048896789550782, 0.02065920066833496, 0.020788223266601562, 0.020752384185791017, 0.020752384185791017, 0.020823040008544923, 0.02067967987060547, 0.020033536911010744, 0.02002227210998535, 0.020025344848632814, 0.020411392211914063, 0.020771839141845702, 0.02071347236633301, 0.020763647079467772, 0.020751359939575196, 0.020781055450439453, 0.020701183319091796, 0.020741119384765624, 0.020738048553466795, 0.0210882568359375, 0.02067865562438965, 0.020716543197631835, 0.02265190315246582, 0.02102681541442871, 0.020818944931030273, 0.020761600494384767, 0.020754432678222655, 0.020824064254760744, 0.020816896438598635, 0.020754432678222655, 0.020768768310546876, 0.020772863388061523, 0.020810752868652343, 0.020730880737304686, 0.020719615936279297, 0.020774911880493165, 0.020742143630981445, 0.020773887634277344, 0.020817920684814452, 0.02090598487854004, 0.020764671325683593, 0.020801536560058592, 0.020780031204223632, 0.020920320510864256, 0.0208353271484375, 0.020765695571899414, 0.020684799194335936, 0.020709375381469726, 0.020749311447143554, 0.02006630325317383, 0.0200581111907959, 0.02003455924987793, 0.020059135437011717, 0.019997695922851562, 0.02010316848754883, 0.020016128540039063, 0.0200581111907959, 0.02008883285522461, 0.02005299186706543, 0.02011136054992676, 0.020175872802734376, 0.02012774467468262, 0.02019327926635742, 0.020124671936035156, 0.02008780860900879, 0.020033536911010744, 0.020106239318847655, 0.02006220817565918, 0.020132863998413086, 0.020025344848632814, 0.020074495315551756, 0.02005606460571289, 0.020102144241333008, 0.02004684829711914, 0.020075519561767577, 0.02000588798522949, 0.020135936737060548, 0.0200898551940918, 0.019969024658203126, 0.019997695922851562, 0.020024320602416993, 0.020009983062744142, 0.02004684829711914, 0.020082687377929686, 0.019974143981933593, 0.019998720169067383, 0.020057088851928712, 0.02003046417236328, 0.020033536911010744, 0.0200263671875, 0.020098047256469728, 0.020060159683227538, 0.02003660774230957, 0.02004377555847168, 0.020067327499389647, 0.020001792907714845, 0.020105215072631837, 0.020059135437011717, 0.019993600845336915, 0.02005299186706543, 0.02002841567993164, 0.02004275131225586, 0.020065280914306642, 0.019834880828857423, 0.019977216720581056, 0.020745216369628908, 0.020142080307006836, 0.02009702491760254, 0.020063232421875, 0.0210513916015625, 0.02082611274719238, 0.02070425605773926, 0.020855808258056642, 0.02086297607421875, 0.021140480041503908, 0.020824064254760744, 0.020769792556762694, 0.020706304550170897, 0.020802560806274413, 0.02082611274719238, 0.020718591690063477, 0.020771839141845702, 0.020749311447143554, 0.020876287460327148, 0.021005311965942384, 0.0200898551940918, 0.0200130558013916, 0.01999564743041992, 0.02005401611328125, 0.02002943992614746, 0.020039680480957032, 0.02005401611328125, 0.019991552352905274, 0.02005504035949707, 0.02007142448425293, 0.020090879440307616, 0.021045248031616212, 0.020766719818115235, 0.020787200927734374, 0.020792320251464845, 0.02001817512512207, 0.019997695922851562, 0.020015104293823242, 0.019978239059448243, 0.019970048904418947, 0.019976192474365235, 0.020171775817871093, 0.02002227210998535, 0.02002841567993164, 0.020016128540039063, 0.02008883285522461, 0.02023423957824707, 0.02002841567993164, 0.02003558349609375, 0.020051967620849608, 0.02006118392944336, 0.020068351745605468, 0.019951616287231445, 0.02008780860900879, 0.020001792907714845, 0.020050943374633787, 0.019997695922851562, 0.02006118392944336, 0.02001408004760742, 0.019907583236694337, 0.02002739143371582, 0.0200447998046875, 0.020040704727172853, 0.020039680480957032, 0.020007936477661133, 0.019990528106689453, 0.020023296356201172, 0.02001203155517578, 0.02004684829711914, 0.02003558349609375, 0.020007936477661133, 0.019984384536743165, 0.020157440185546875, 0.020050943374633787, 0.02002841567993164, 0.020057088851928712, 0.02026700782775879, 0.02008166313171387, 0.020068351745605468, 0.020059135437011717, 0.02001203155517578, 0.02004991912841797, 0.020139007568359374, 0.020002815246582033, 0.02004275131225586, 0.02007859230041504, 0.020090879440307616, 0.0200581111907959, 0.02004787254333496, 0.020023296356201172, 0.02005606460571289, 0.02004377555847168, 0.02003660774230957, 0.020098047256469728, 0.02003455924987793, 0.020009983062744142, 0.019994623184204103, 0.02002022361755371, 0.02004172706604004, 0.020024320602416993, 0.02005504035949707, 0.02006425666809082, 0.0200898551940918, 0.02008166313171387, 0.02002943992614746, 0.02005299186706543, 0.0200581111907959, 0.0200447998046875, 0.020168703079223634, 0.019911680221557617, 0.020427776336669923, 0.021914623260498048, 0.020823040008544923, 0.02085068893432617, 0.02082713508605957, 0.020745216369628908, 0.020716543197631835, 0.02064691162109375, 0.020747264862060546, 0.020757503509521484, 0.020864000320434572, 0.020799488067626954, 0.020157440185546875, 0.020074495315551756, 0.01999667167663574, 0.020051967620849608, 0.020068351745605468, 0.020090879440307616, 0.02011136054992676, 0.020113407135009767, 0.02005401611328125, 0.020084735870361328, 0.02005299186706543, 0.020346879959106445, 0.020126720428466797, 0.020024320602416993, 0.02007859230041504, 0.02008166313171387, 0.020101119995117187, 0.020126720428466797, 0.020122623443603514, 0.020098047256469728, 0.02012774467468262, 0.020175872802734376, 0.020147199630737304, 0.02012876892089844, 0.02004582405090332, 0.02007756805419922, 0.020082687377929686, 0.02007859230041504, 0.02006118392944336, 0.020093952178955078, 0.02008883285522461, 0.020093952178955078, 0.02005504035949707, 0.020083711624145507, 0.020150272369384766, 0.019961856842041017, 0.020065280914306642, 0.02004787254333496, 0.02005299186706543, 0.020068351745605468, 0.02001817512512207, 0.020025344848632814, 0.020048896789550782, 0.02007347106933594, 0.020039680480957032, 0.02000383949279785, 0.02004684829711914, 0.020050943374633787, 0.02000486373901367, 0.02003660774230957, 0.020093952178955078, 0.020074495315551756, 0.020002815246582033, 0.02023321533203125, 0.020191232681274415, 0.0200263671875, 0.02003455924987793, 0.020137983322143553, 0.020126720428466797, 0.020178943634033202, 0.0200949764251709, 0.020188159942626953, 0.020075519561767577, 0.0200447998046875, 0.020074495315551756, 0.02000588798522949, 0.02003046417236328, 0.020024320602416993, 0.020040704727172853, 0.02005299186706543, 0.020118528366088868, 0.02001817512512207, 0.020076543807983398, 0.020175872802734376, 0.02002227210998535, 0.02002841567993164, 0.020025344848632814, 0.019987455368041994, 0.020231168746948244, 0.02001919937133789, 0.02001919937133789, 0.02008883285522461, 0.020162559509277343, 0.020160512924194338, 0.020083711624145507, 0.020092927932739257, 0.01987993621826172, 0.02046566390991211, 0.02079641532897949, 0.020742143630981445, 0.021000192642211913, 0.02049945640563965, 0.019937280654907227, 0.02004582405090332, 0.020076543807983398, 0.020060159683227538, 0.020090879440307616, 0.02003763198852539, 0.020031488418579102, 0.02004991912841797, 0.020032512664794923, 0.019992639541625976, 0.020117439270019532, 0.020091903686523437, 0.020106239318847655, 0.02002841567993164, 0.02000486373901367, 0.020001792907714845, 0.02002739143371582, 0.01998847961425781, 0.020039680480957032, 0.0200949764251709, 0.020048896789550782, 0.02002841567993164, 0.020107263565063475, 0.02006937599182129, 0.02001919937133789, 0.02004787254333496, 0.020039680480957032, 0.0200447998046875, 0.01999977684020996, 0.019978208541870116, 0.020006912231445313, 0.019984384536743165, 0.020020256042480467, 0.019993568420410158]",tokens/s,48.91433851319235,,,,,,MB,934.129664,6768.033792,0.0,6138.363904,6060.931072,s,1,6.85633740234375,6.85633740234375,0.0,6.85633740234375,6.85633740234375,6.85633740234375,6.85633740234375,[6.85633740234375],,kWh,1.689535661811053e-06,8.890438926758776e-07,1.801390330007091e-06,4.379969884494022e-06, bfloat16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,s,s,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/s/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6694736c-7dc7f9183d347ac031da4ff3;9fc07c5c-19d5-4eb5-922d-d33cf3318a42) Repository Not Found for url: https://huggingface.co/s/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: s is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-40b,tiiuae/falcon-40b,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ self.load_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 250, in load_model_with_no_weights self.pretrained_model = self.automodel_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 559, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3704, in from_pretrained config = cls._autoset_attn_implementation( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1490, in _autoset_attn_implementation config = cls._check_and_enable_sdpa( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1656, in _check_and_enable_sdpa raise ValueError( ValueError: FalconForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,google/gemma-7b,google/gemma-7b,cuda,0,42,,,True,,,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.223-212.873.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.0,,,,1.21.4,,,,0.12.0,,,gemma,MB,1513.10336,19300.614144,0.0,18654.16704,17596.665856,s,10,0.9371416015625,0.09371416015625,0.00010236322870737306,0.09371141052246093,0.09384496765136718,0.09388114013671875,0.093910078125,"[0.09383692932128906, 0.09370829010009765, 0.09361990356445313, 0.0936607666015625, 0.09371453094482422, 0.09375843048095703, 0.09391731262207031, 0.09359276580810547, 0.0935785903930664, 0.09375408172607422]",tokens/s,2731.710977008919,kWh,1.1076672646936931e-06,6.069461037712454e-07,6.201113475944055e-06,7.915726844408994e-06,tokens/kWh,32340681.409543198,MB,1566.527488,19302.711296,0.0,18656.264192,17596.668416,s,10,27.48278198242188,2.748278198242188,0.0012452002655436959,2.747768310546875,2.7498784423828124,2.750382824707031,2.7507863305664064,"[2.747681396484375, 2.74757421875, 2.749451416015625, 2.747508056640625, 2.746569580078125, 2.74780712890625, 2.7477919921875, 2.74774462890625, 2.749766357421875, 2.75088720703125]",tokens/s,22.923443500114036,kWh,3.24635648151678e-05,1.7791422537480437e-05,0.0001385764745682555,0.00018883146192090372,tokens/kWh,333630.8439236093,,s,630,27.480954860687266,0.04362056327093215,0.00011093597528972763,0.043606014251708985,0.04370431900024414,0.043769993400573734,0.04413892467498779,"[0.043758590698242186, 0.04355276870727539, 0.043509761810302736, 0.043478015899658204, 0.04350873565673828, 0.04350873565673828, 0.04349747085571289, 0.04348825454711914, 0.04349747085571289, 0.04353023910522461, 0.04355891036987305, 0.04352819061279297, 0.043545600891113284, 0.04354355239868164, 0.043514881134033206, 0.04353843307495117, 0.04362444686889649, 0.04350668716430664, 0.04354969787597656, 0.04357017517089844, 0.04358041763305664, 0.04355891036987305, 0.04355379104614258, 0.04358246231079101, 0.04355891036987305, 0.043581439971923826, 0.04375449752807617, 0.04402175903320313, 0.04367871856689453, 0.0436756477355957, 0.04358860778808594, 0.04360908889770508, 0.04362137603759766, 0.04359065628051758, 0.04357734298706055, 0.0435599365234375, 0.04360806274414063, 0.04365824127197265, 0.043619327545166016, 0.04363673782348633, 0.043638782501220705, 0.04360192108154297, 0.04369919967651367, 0.043635711669921876, 0.04372582244873047, 0.04363161468505859, 0.043602943420410156, 0.043630592346191405, 0.043646976470947264, 0.04363673782348633, 0.043663360595703124, 0.04369715118408203, 0.04369203186035156, 0.04363468933105469, 0.043632640838623046, 0.04363161468505859, 0.043673599243164066, 0.04365926361083984, 0.04367871856689453, 0.043663360595703124, 0.043663360595703124, 0.04363980865478516, 0.04370022583007813, 0.043719680786132815, 0.043512832641601565, 0.04360704040527344, 0.043461631774902344, 0.04348723220825195, 0.043581439971923826, 0.043491329193115234, 0.04350156784057617, 0.04348825454711914, 0.04363980865478516, 0.043545600891113284, 0.043496448516845705, 0.04352511978149414, 0.04357529449462891, 0.04359372711181641, 0.04353126525878906, 0.04352819061279297, 0.04353945541381836, 0.04351180648803711, 0.04352102279663086, 0.04355686569213867, 0.043643905639648435, 0.04355481719970703, 0.043679744720458984, 0.04362649536132813, 0.0435865592956543, 0.04356710433959961, 0.04355891036987305, 0.04353433609008789, 0.04356915283203125, 0.04356710433959961, 0.04355686569213867, 0.04401049423217773, 0.04368588638305664, 0.04358041763305664, 0.04359167861938477, 0.04358553695678711, 0.04362035369873047, 0.04355686569213867, 0.04357427215576172, 0.04356403350830078, 0.04362547302246094, 0.043597824096679685, 0.04359167861938477, 0.043650047302246094, 0.04376883316040039, 0.04369919967651367, 0.043676673889160154, 0.043632640838623046, 0.04363468933105469, 0.04371353530883789, 0.04368588638305664, 0.04360192108154297, 0.04369612884521484, 0.0437841911315918, 0.04373299026489258, 0.043670528411865236, 0.043670528411865236, 0.043657215118408206, 0.04372480010986328, 0.04365107345581055, 0.04366745758056641, 0.04362956619262695, 0.04371558380126953, 0.04415180969238281, 0.04358758544921875, 0.043509761810302736, 0.04352819061279297, 0.04351692962646484, 0.043507713317871094, 0.04357734298706055, 0.04354252624511719, 0.04350156784057617, 0.04353023910522461, 0.04354355239868164, 0.04363776016235352, 0.0436121597290039, 0.04356095886230469, 0.04351692962646484, 0.04353228759765625, 0.04357734298706055, 0.04357324981689453, 0.043515903472900394, 0.04356710433959961, 0.04353638458251953, 0.043545600891113284, 0.0435230712890625, 0.04355276870727539, 0.04354252624511719, 0.043581439971923826, 0.04366643142700195, 0.044027904510498046, 0.04363468933105469, 0.043600894927978515, 0.04355897521972656, 0.043587520599365236, 0.04355686569213867, 0.04359270477294922, 0.04357734298706055, 0.043786239624023435, 0.043595775604248044, 0.04361523056030273, 0.0435968017578125, 0.0436121597290039, 0.04366643142700195, 0.04360806274414063, 0.04359372711181641, 0.04360704040527344, 0.043584510803222655, 0.04359884643554687, 0.04361011123657227, 0.044499969482421874, 0.0437657585144043, 0.043668479919433595, 0.043630592346191405, 0.04413747024536133, 0.04369203186035156, 0.04368384170532227, 0.04363673782348633, 0.04367462539672851, 0.04375142288208008, 0.04364492797851562, 0.04367871856689453, 0.04367871856689453, 0.04361011123657227, 0.04364595031738281, 0.04369612884521484, 0.04348825454711914, 0.043496448516845705, 0.043476993560791016, 0.043493377685546876, 0.04352102279663086, 0.04354150390625, 0.043568126678466795, 0.043545600891113284, 0.04355379104614258, 0.04356300735473633, 0.043515903472900394, 0.04352614212036133, 0.04357324981689453, 0.04354355239868164, 0.043535358428955076, 0.04352000045776367, 0.043611137390136716, 0.04352511978149414, 0.043507713317871094, 0.04354662322998047, 0.043561985015869144, 0.04356915283203125, 0.04354355239868164, 0.043581439971923826, 0.04359167861938477, 0.043581439971923826, 0.043600894927978515, 0.04356300735473633, 0.043579391479492184, 0.0436756477355957, 0.043657215118408206, 0.043681793212890625, 0.043641857147216793, 0.04359987258911133, 0.043545600891113284, 0.04365926361083984, 0.043600894927978515, 0.0435968017578125, 0.04357529449462891, 0.04365824127197265, 0.043619327545166016, 0.04363980865478516, 0.043663360595703124, 0.043614208221435545, 0.04367769622802734, 0.04361523056030273, 0.043660289764404295, 0.043686912536621096, 0.04366131210327148, 0.04361318588256836, 0.043638782501220705, 0.04362137603759766, 0.04379852676391602, 0.043670528411865236, 0.04369305419921875, 0.04369100952148437, 0.043652095794677735, 0.04365926361083984, 0.04369203186035156, 0.043937793731689455, 0.04374220657348633, 0.04365414428710938, 0.04381388854980469, 0.043545600891113284, 0.04348416137695312, 0.043491329193115234, 0.0434442253112793, 0.04349235153198242, 0.04349030303955078, 0.04351078414916992, 0.043529216766357424, 0.04354457473754883, 0.04354969787597656, 0.043568126678466795, 0.04349542236328125, 0.04354252624511719, 0.04349030303955078, 0.043515903472900394, 0.04356710433959961, 0.04355481719970703, 0.04350668716430664, 0.04353843307495117, 0.04357222366333008, 0.043566078186035154, 0.043491329193115234, 0.04359270477294922, 0.04356300735473633, 0.04369100952148437, 0.04356915283203125, 0.04354764938354492, 0.04354048156738281, 0.04359372711181641, 0.043568126678466795, 0.04354150390625, 0.04355276870727539, 0.0436121597290039, 0.04359987258911133, 0.04358553695678711, 0.043686912536621096, 0.04360704040527344, 0.04356915283203125, 0.043602943420410156, 0.04356300735473633, 0.04361830520629883, 0.04358348846435547, 0.04364799880981445, 0.04368998336791992, 0.043652095794677735, 0.04359065628051758, 0.0436234245300293, 0.043632640838623046, 0.04368281555175781, 0.043660289764404295, 0.043619327545166016, 0.043665409088134766, 0.043908096313476565, 0.04363776016235352, 0.04362649536132813, 0.043652095794677735, 0.04371046447753906, 0.043660289764404295, 0.043650047302246094, 0.043655166625976564, 0.043663360595703124, 0.043670528411865236, 0.043770942687988285, 0.043639743804931644, 0.04351078414916992, 0.04348416137695312, 0.04348723220825195, 0.043529216766357424, 0.043524097442626954, 0.043504638671875, 0.04353638458251953, 0.043504638671875, 0.04353228759765625, 0.04352102279663086, 0.04353638458251953, 0.043494400024414064, 0.04352204895019531, 0.04356300735473633, 0.04355788803100586, 0.043617279052734374, 0.04354252624511719, 0.043535358428955076, 0.04353843307495117, 0.04351692962646484, 0.04357222366333008, 0.043617279052734374, 0.043796478271484376, 0.04357734298706055, 0.043579391479492184, 0.04358348846435547, 0.043597824096679685, 0.04357734298706055, 0.04357632064819336, 0.04355788803100586, 0.04357120132446289, 0.04355379104614258, 0.04361625671386719, 0.04357324981689453, 0.043878398895263675, 0.04368588638305664, 0.04362547302246094, 0.043584510803222655, 0.04364799880981445, 0.04364191818237305, 0.043606975555419925, 0.04358041763305664, 0.0436121597290039, 0.04393881607055664, 0.04369305419921875, 0.04372275161743164, 0.04365932846069336, 0.0436376953125, 0.04365619277954102, 0.043632640838623046, 0.043687934875488284, 0.043649024963378906, 0.04365926361083984, 0.04362137603759766, 0.04363980865478516, 0.043627521514892575, 0.043640830993652346, 0.04362649536132813, 0.04369100952148437, 0.043671550750732424, 0.04398080062866211, 0.04371148681640625, 0.04348723220825195, 0.04346879959106445, 0.04356710433959961, 0.04353433609008789, 0.04366233444213867, 0.043548671722412106, 0.04352511978149414, 0.043529216766357424, 0.04355583953857422, 0.04353747177124023, 0.04349228668212891, 0.04356403350830078, 0.04363980865478516, 0.043566078186035154, 0.043524097442626954, 0.043529216766357424, 0.04350668716430664, 0.043594753265380856, 0.04353023910522461, 0.04352000045776367, 0.04355686569213867, 0.0435230712890625, 0.043578369140625, 0.04355686569213867, 0.0435968017578125, 0.043643905639648435, 0.043585601806640624, 0.04358342361450195, 0.043606014251708985, 0.04366233444213867, 0.04374016189575195, 0.043600894927978515, 0.04361014556884766, 0.04358243179321289, 0.043600894927978515, 0.04361318588256836, 0.043581439971923826, 0.043581439971923826, 0.043589630126953126, 0.04360192108154297, 0.04377804946899414, 0.04366233444213867, 0.04366233444213867, 0.043617279052734374, 0.043649024963378906, 0.043617279052734374, 0.043650047302246094, 0.04371353530883789, 0.04368896102905274, 0.04364287948608398, 0.043772926330566404, 0.04370431900024414, 0.04377395248413086, 0.043650047302246094, 0.04374220657348633, 0.04370431900024414, 0.04367769622802734, 0.04364799880981445, 0.04365107345581055, 0.043665409088134766, 0.043686912536621096, 0.043660289764404295, 0.043756542205810545, 0.043561985015869144, 0.04346777725219726, 0.04344934463500977, 0.04346777725219726, 0.04357632064819336, 0.04350054550170898, 0.04361011123657227, 0.04355788803100586, 0.04353945541381836, 0.043504638671875, 0.043499519348144534, 0.043514881134033206, 0.04355276870727539, 0.04354252624511719, 0.04353638458251953, 0.043561985015869144, 0.043535358428955076, 0.043606014251708985, 0.043561985015869144, 0.04359372711181641, 0.04353843307495117, 0.043602943420410156, 0.04359987258911133, 0.043551742553710936, 0.04360704040527344, 0.04357427215576172, 0.04356095886230469, 0.04360396957397461, 0.04378112030029297, 0.043643905639648435, 0.04359065628051758, 0.04359372711181641, 0.043617279052734374, 0.0436121597290039, 0.04358758544921875, 0.04363161468505859, 0.043650047302246094, 0.04357734298706055, 0.044300289154052735, 0.04362345504760742, 0.043643871307373044, 0.04360192108154297, 0.04360396957397461, 0.0435865592956543, 0.04361830520629883, 0.04361318588256836, 0.04363980865478516, 0.043632640838623046, 0.043655166625976564, 0.043671550750732424, 0.04365926361083984, 0.04366643142700195, 0.04363161468505859, 0.043679744720458984, 0.04362649536132813, 0.04362547302246094, 0.04365311813354492, 0.043663360595703124, 0.04365311813354492, 0.04375244903564453, 0.043701248168945314, 0.04363673782348633, 0.04373708724975586, 0.04365107345581055, 0.043638782501220705, 0.043529216766357424, 0.04348825454711914, 0.043597824096679685, 0.043529216766357424, 0.04349542236328125, 0.04351180648803711, 0.04363776016235352, 0.04361830520629883, 0.043747329711914064, 0.04371660614013672, 0.04353740692138672, 0.043551742553710936, 0.0435599365234375, 0.04354662322998047, 0.04357529449462891, 0.04402073669433594, 0.04356710433959961, 0.04357222366333008, 0.04356300735473633, 0.04358246231079101, 0.043579391479492184, 0.04357222366333008, 0.043568126678466795, 0.04357324981689453, 0.04355481719970703, 0.04356403350830078, 0.043594753265380856, 0.04360192108154297, 0.04367769622802734, 0.04380672073364258, 0.04361011123657227, 0.043701248168945314, 0.04360192108154297, 0.043627521514892575, 0.04359987258911133, 0.043617279052734374, 0.043595775604248044, 0.043655166625976564, 0.04363980865478516, 0.0441794548034668, 0.043627521514892575, 0.043652095794677735, 0.04364287948608398, 0.04364287948608398, 0.04364287948608398, 0.04365926361083984, 0.04369715118408203, 0.043606014251708985, 0.04360908889770508, 0.04369715118408203, 0.043665409088134766, 0.043655166625976564, 0.043640830993652346, 0.0436756477355957, 0.04365926361083984, 0.044216320037841796, 0.043663360595703124, 0.04369305419921875, 0.04366643142700195, 0.043673599243164066, 0.043943935394287106, 0.043600894927978515, 0.04353433609008789, 0.043504638671875, 0.04351795196533203, 0.04355788803100586, 0.04419174575805664, 0.043561985015869144, 0.043589630126953126, 0.043746368408203125, 0.04351071929931641, 0.04355379104614258, 0.043597824096679685, 0.04363161468505859, 0.04402687835693359, 0.043954177856445314, 0.043594753265380856, 0.04352614212036133, 0.04353740692138672, 0.04350668716430664, 0.04355788803100586, 0.043652095794677735, 0.0436234245300293, 0.04355379104614258, 0.04409139251708984, 0.04371046447753906, 0.04358553695678711, 0.043622432708740236, 0.04358448028564453, 0.04364799880981445, 0.04375244903564453, 0.043614208221435545, 0.04361830520629883, 0.043622432708740236, 0.043607009887695315, 0.043617279052734374, 0.0436234245300293, 0.04365311813354492, 0.043635711669921876, 0.043600894927978515, 0.0435968017578125, 0.0436049919128418, 0.043606014251708985, 0.04362444686889649, 0.04359065628051758, 0.043652095794677735, 0.04362137603759766, 0.04370943832397461, 0.04413951873779297, 0.043650047302246094, 0.04366438293457031, 0.043665409088134766, 0.043684864044189455, 0.04370636749267578, 0.043638782501220705, 0.04366233444213867, 0.04362246322631836, 0.0436479377746582, 0.04365414428710938, 0.043681793212890625, 0.043673599243164066, 0.04372172927856445, 0.04372480010986328]",tokens/s,22.924967607338978,,,,,,MB,935.391232,19279.642624,0.0,18649.972736,18648.241152,s,1,6.99146630859375,6.99146630859375,0.0,6.99146630859375,6.99146630859375,6.99146630859375,6.99146630859375,[6.99146630859375],,kWh,2.3003205277771284e-06,1.2445283130222494e-06,4.447781336000545e-06,7.992630176799922e-06, bfloat16-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-125m,facebook/opt-125m,cuda,0,42,,,True,,,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",opt,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,databricks/dbrx-base,databricks/dbrx-base,cuda,0,42,,,True,,,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",dbrx,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-350m,facebook/opt-350m,cuda,0,42,,,True,,,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",opt,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,-,-,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 106, in _inner_fn validate_repo_id(arg_value) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 160, in validate_repo_id raise HFValidationError( huggingface_hub.errors.HFValidationError: Repo id must use alphanumeric chars or '-', '_', '.', '--' and '..' are forbidden, '-' and '.' cannot start or end the name, max length is 96: '-'. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 466, in cached_file raise EnvironmentError( OSError: Incorrect path_or_model_id: '-'. Please provide either the path to a local folder or the repo_id of a model on the Hub. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,stabilityai/stablelm-2-12b,stabilityai/stablelm-2-12b,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ self.load_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 250, in load_model_with_no_weights self.pretrained_model = self.automodel_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3710, in from_pretrained model = cls(config, *model_args, **model_kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/stablelm/modeling_stablelm.py"", line 1188, in __init__ self.model = StableLmModel(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/stablelm/modeling_stablelm.py"", line 966, in __init__ [StableLmDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/stablelm/modeling_stablelm.py"", line 966, in [StableLmDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] File ""/usr/local/lib/python3.10/dist-packages/transformers/models/stablelm/modeling_stablelm.py"", line 748, in __init__ self.mlp = StableLmMLP(config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/stablelm/modeling_stablelm.py"", line 206, in __init__ self.down_proj = nn.Linear(self.intermediate_size, self.hidden_size, bias=False) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/linear.py"", line 98, in __init__ self.weight = Parameter(torch.empty((out_features, in_features), **factory_kwargs)) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_device.py"", line 78, in __torch_function__ return func(*args, **kwargs) torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 136.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,v,v,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/v/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 402, in cached_file resolved_file = hf_hub_download( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir _raise_on_head_call_error(head_call_error, force_download, local_files_only) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn return fn(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata r = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66932575-6daa35b55d7a026b0328da48;a4556409-6abc-4dd5-bbd7-afb45c3fbb20) Repository Not Found for url: https://huggingface.co/v/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. The above exception was the direct cause of the following exception: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 40, in __init__ super().__init__(config) File ""/workspace/optimum_benchmark/backends/base.py"", line 65, in __init__ self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config return AutoConfig.from_pretrained(model, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 965, in from_pretrained config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 425, in cached_file raise EnvironmentError( OSError: v is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,openai-community/gpt2,openai-community/gpt2,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.42.3,,0.31.0,,,,1.20.0,,,,0.11.1,,,,MB,1187.037184,1032.323072,0.0,385.875968,319.013888,s,24,0.16614979171752928,0.006922907988230388,0.00025689639029946853,0.006908672094345093,0.007092284774780274,0.007169988632202148,0.007770443029403686,"[0.00794601583480835, 0.007182655811309814, 0.00672051191329956, 0.006576992034912109, 0.006629983901977539, 0.0069253120422363285, 0.006882175922393798, 0.007098207950592041, 0.006853184223175049, 0.006936384201049805, 0.006966239929199218, 0.006958655834197998, 0.0069008002281188965, 0.006776383876800537, 0.007078464031219482, 0.006978975772857666, 0.0070256319046020504, 0.006923903942108154, 0.0067259521484375, 0.006828351974487305, 0.006757760047912597, 0.006720704078674316, 0.006840000152587891, 0.006916543960571289]",tokens/s,36978.68012043851,kWh,7.999625488838834e-08,4.383453264821003e-08,2.0520478313607772e-07,3.290355706726761e-07,tokens/kWh,778031382.6758513,MB,1187.037184,1032.323072,0.0,385.875968,319.016448,s,24,10.069377044677735,0.4195573768615722,0.011799049685952266,0.42221612548828125,0.42381627502441405,0.42443595428466796,0.45633053283691405,"[0.46583551025390624, 0.4108739013671875, 0.4038103637695313, 0.4043648681640625, 0.40876107788085936, 0.42334307861328124, 0.42300326538085936, 0.42265847778320315, 0.42234780883789064, 0.4216253967285156, 0.42231155395507813, 0.4224419250488281, 0.423015869140625, 0.4163055419921875, 0.4232668151855469, 0.4225952758789063, 0.4240190734863281, 0.41185775756835935, 0.424509521484375, 0.4156019592285156, 0.40542184448242186, 0.40987591552734376, 0.4194095458984375, 0.42212069702148436]",tokens/s,150.15824646264312,kWh,4.913428861606461e-06,2.692341967668101e-06,8.987059515113179e-06,1.6592830344387743e-05,tokens/kWh,3796820.596150357,,s,1511,10.220302339553845,0.006763932719757667,0.000904736665580847,0.006684671878814697,0.006779903888702392,0.0073190400600433355,0.013552898883819588,"[0.007762944221496582, 0.007612415790557861, 0.007504896163940429, 0.007604288101196289, 0.007594944000244141, 0.007506944179534912, 0.007333888053894043, 0.0074065918922424315, 0.007269375801086426, 0.007230463981628418, 0.007327744007110596, 0.007361536026000977, 0.007455743789672851, 0.007585792064666748, 0.007387135982513428, 0.0073359360694885255, 0.007442431926727295, 0.007425024032592774, 0.007442431926727295, 0.007409664154052734, 0.007387135982513428, 0.007373824119567871, 0.007395328044891358, 0.007492703914642334, 0.007273375988006592, 0.007266304016113281, 0.007322624206542969, 0.00738918399810791, 0.007346176147460938, 0.0074291200637817386, 0.007414783954620361, 0.007488512039184571, 0.007408639907836914, 0.007367680072784424, 0.007332863807678222, 0.007392255783081054, 0.00738099193572998, 0.007820288181304931, 0.007282688140869141, 0.007279615879058838, 0.007362559795379638, 0.007420928001403809, 0.007305215835571289, 0.007282688140869141, 0.007341055870056152, 0.007400447845458984, 0.007321599960327148, 0.007444479942321777, 0.0075345921516418455, 0.007476352214813233, 0.007643008232116699, 0.007334911823272705, 0.007385087966918945, 0.007267327785491944, 0.007355391979217529, 0.007333888053894043, 0.007236608028411865, 0.00709222412109375, 0.007198719978332519, 0.007374847888946533, 0.007044095993041993, 0.006820864200592041, 0.014736384391784667, 0.007174208164215088, 0.007349184036254883, 0.00724070405960083, 0.0073400321006774905, 0.007422976016998291, 0.007511040210723877, 0.006703104019165039, 0.006416384220123291, 0.006462463855743408, 0.00642460823059082, 0.006494175910949707, 0.006435840129852295, 0.0064430079460144046, 0.006437888145446777, 0.006408192157745361, 0.006425600051879882, 0.006415359973907471, 0.006375423908233643, 0.00638976001739502, 0.006394879817962646, 0.006709311962127686, 0.006417344093322754, 0.0063805441856384275, 0.006434815883636475, 0.006433792114257812, 0.00638156795501709, 0.006423552036285401, 0.006379519939422608, 0.006429696083068847, 0.006412288188934326, 0.0063907837867736815, 0.006404096126556396, 0.006408192157745361, 0.006394879817962646, 0.006449215888977051, 0.006369215965270996, 0.006415359973907471, 0.006372352123260498, 0.006442048072814941, 0.006399936199188232, 0.006471680164337158, 0.006392831802368164, 0.006387712001800537, 0.0063907837867736815, 0.0064102401733398436, 0.006367231845855713, 0.00643071985244751, 0.0063539199829101565, 0.006390880107879639, 0.0063927359580993656, 0.006372352123260498, 0.0064245758056640625, 0.0063907837867736815, 0.006367231845855713, 0.0063836159706115725, 0.006415359973907471, 0.006390880107879639, 0.006386591911315918, 0.00638976001739502, 0.006386688232421875, 0.006369279861450195, 0.0064174079895019534, 0.013435903549194337, 0.0063919677734375, 0.006342495918273926, 0.006405119895935059, 0.0064471039772033695, 0.006396959781646729, 0.00638972806930542, 0.006401023864746094, 0.006397952079772949, 0.006360064029693604, 0.006384640216827392, 0.006406144142150879, 0.006393856048583985, 0.006435840129852295, 0.0063805441856384275, 0.0064133119583129885, 0.006397952079772949, 0.006421504020690918, 0.006388735771179199, 0.006451200008392334, 0.006419456005096436, 0.0064143362045288085, 0.006372352123260498, 0.006397952079772949, 0.006394879817962646, 0.006385663986206055, 0.006389920234680176, 0.006368095874786377, 0.006375423908233643, 0.006408192157745361, 0.006396992206573486, 0.006410208225250244, 0.006375391960144043, 0.006492159843444824, 0.006405119895935059, 0.006455296039581298, 0.006401023864746094, 0.006385663986206055, 0.0064133119583129885, 0.006386688232421875, 0.006375423908233643, 0.006388735771179199, 0.006375423908233643, 0.006402048110961914, 0.006379519939422608, 0.0064011521339416505, 0.006477695941925049, 0.0064102401733398436, 0.0064737281799316405, 0.006461440086364746, 0.006361120223999024, 0.006451168060302734, 0.006397952079772949, 0.006403071880340576, 0.006392831802368164, 0.006362112045288086, 0.006423552036285401, 0.0064143362045288085, 0.006455296039581298, 0.00643891191482544, 0.006382656097412109, 0.006494143962860108, 0.006363135814666748, 0.013321215629577637, 0.0063805441856384275, 0.00643993616104126, 0.006362112045288086, 0.006406144142150879, 0.006504447937011719, 0.00638156795501709, 0.006416512012481689, 0.006410111904144287, 0.0064133119583129885, 0.006405119895935059, 0.0063907837867736815, 0.006441984176635742, 0.0064440321922302245, 0.006440959930419922, 0.006486112117767334, 0.006431647777557373, 0.006376512050628662, 0.006428607940673828, 0.006421504020690918, 0.006404128074645996, 0.00640719985961914, 0.006407104015350342, 0.006402048110961914, 0.006405119895935059, 0.006463488101959228, 0.006377471923828125, 0.006416384220123291, 0.006434815883636475, 0.006402048110961914, 0.006428671836853027, 0.006388735771179199, 0.0064031357765197755, 0.006416319847106933, 0.00643071985244751, 0.006426623821258545, 0.006426623821258545, 0.006549503803253174, 0.006434815883636475, 0.006396927833557129, 0.006452223777770996, 0.0064204797744750975, 0.006354944229125976, 0.006436863899230957, 0.006394879817962646, 0.006423552036285401, 0.006437888145446777, 0.0063805441856384275, 0.006404096126556396, 0.0064143362045288085, 0.006364160060882569, 0.006402048110961914, 0.00638976001739502, 0.00642252779006958, 0.006367231845855713, 0.00643174409866333, 0.006402048110961914, 0.00636518383026123, 0.006474751949310303, 0.006378496170043945, 0.0063805441856384275, 0.006385663986206055, 0.006360064029693604, 0.013294591903686523, 0.006421504020690918, 0.006387712001800537, 0.006364160060882569, 0.00638976001739502, 0.006377503871917724, 0.006418399810791016, 0.006400000095367431, 0.006386688232421875, 0.006388735771179199, 0.006418432235717773, 0.006393856048583985, 0.006416384220123291, 0.0064245758056640625, 0.006394879817962646, 0.0064133119583129885, 0.00643891191482544, 0.006469632148742676, 0.006351871967315674, 0.006429696083068847, 0.006412288188934326, 0.006395904064178467, 0.00637337589263916, 0.00638156795501709, 0.006450175762176514, 0.006406144142150879, 0.006446176052093506, 0.006417312145233154, 0.006355967998504639, 0.006417535781860352, 0.0063855361938476565, 0.006394879817962646, 0.0064143362045288085, 0.006378496170043945, 0.006441023826599121, 0.006414271831512451, 0.006489088058471679, 0.006407167911529541, 0.0063836159706115725, 0.00643993616104126, 0.006376448154449463, 0.006405119895935059, 0.006391808032989502, 0.006376448154449463, 0.006391808032989502, 0.006398975849151611, 0.006576159954071045, 0.006783967971801758, 0.006676479816436768, 0.006692863941192627, 0.006669312000274658, 0.006687744140625, 0.006654975891113281, 0.006745151996612549, 0.0066631040573120115, 0.006737055778503418, 0.006729568004608154, 0.006699007987976074, 0.006635519981384277, 0.006723584175109864, 0.006740992069244385, 0.0067041277885437015, 0.006750304222106934, 0.014007200241088867, 0.006707263946533203, 0.0067265920639038084, 0.00677785587310791, 0.006719488143920899, 0.006718463897705078, 0.00669593620300293, 0.0066447358131408694, 0.00675328016281128, 0.006749184131622315, 0.006687744140625, 0.006704192161560058, 0.006676447868347168, 0.006746079921722412, 0.006717440128326416, 0.006740992069244385, 0.0067686400413513184, 0.006670335769653321, 0.006717440128326416, 0.006654975891113281, 0.006677504062652588, 0.006744063854217529, 0.0066909117698669435, 0.006728608131408692, 0.006762495994567871, 0.0066694397926330565, 0.006675327777862549, 0.006684671878814697, 0.006681600093841553, 0.0067348480224609375, 0.00673689603805542, 0.007169023990631103, 0.0067348480224609375, 0.0066979842185974124, 0.00667955207824707, 0.0067276802062988285, 0.006708223819732666, 0.006676479816436768, 0.006750207901000976, 0.006691840171813965, 0.006684671878814697, 0.006685696125030518, 0.0066826238632202144, 0.006707200050354004, 0.006648831844329834, 0.006687744140625, 0.006719583988189697, 0.0066835517883300785, 0.006681600093841553, 0.006713344097137451, 0.006652927875518798, 0.006702079772949219, 0.006788095951080322, 0.006694911956787109, 0.006724607944488525, 0.0066979842185974124, 0.006687744140625, 0.006656000137329102, 0.006788095951080322, 0.006681600093841553, 0.006699007987976074, 0.0067420158386230465, 0.006681600093841553, 0.013900799751281738, 0.006754303932189941, 0.006715392112731934, 0.006729728221893311, 0.006715392112731934, 0.006648831844329834, 0.0066744318008422855, 0.006690815925598144, 0.006670335769653321, 0.0067123198509216305, 0.0066979842185974124, 0.006723584175109864, 0.006722559928894043, 0.006811647891998291, 0.006799359798431396, 0.006700032234191895, 0.006677504062652588, 0.006670335769653321, 0.006672383785247803, 0.006763552188873291, 0.006719456195831299, 0.006684671878814697, 0.006687744140625, 0.006767615795135498, 0.006666240215301514, 0.0066641921997070315, 0.006715392112731934, 0.006685696125030518, 0.006676479816436768, 0.006704160213470459, 0.006677472114562989, 0.006697055816650391, 0.006732704162597657, 0.006716415882110595, 0.006730751991271973, 0.006714367866516113, 0.006725632190704346, 0.006774784088134766, 0.006783999919891357, 0.00672870397567749, 0.0067041277885437015, 0.006731776237487793, 0.006708223819732666, 0.006677504062652588, 0.006761568069458008, 0.006799263954162598, 0.006700032234191895, 0.006721536159515381, 0.006719488143920899, 0.0066713600158691405, 0.006699007987976074, 0.006700064182281494, 0.006659039974212647, 0.006721536159515381, 0.006694911956787109, 0.0066641921997070315, 0.006692895889282227, 0.006664159774780273, 0.006701087951660157, 0.006671328067779541, 0.006708223819732666, 0.006726655960083008, 0.006716415882110595, 0.0141844482421875, 0.006669312000274658, 0.006723584175109864, 0.006713344097137451, 0.006689792156219483, 0.006710271835327148, 0.006761472225189209, 0.006687744140625, 0.006721536159515381, 0.006722559928894043, 0.006692863941192627, 0.00669593620300293, 0.006684671878814697, 0.006696959972381592, 0.006820864200592041, 0.00672870397567749, 0.006687776088714599, 0.006632415771484375, 0.006710271835327148, 0.006645823955535889, 0.006780863761901856, 0.006699071884155273, 0.006681536197662354, 0.006677504062652588, 0.0067041277885437015, 0.0066641921997070315, 0.006684671878814697, 0.006710271835327148, 0.006643712043762207, 0.006691840171813965, 0.0066979842185974124, 0.006639616012573242, 0.006667263984680176, 0.006696959972381592, 0.0067348480224609375, 0.006726655960083008, 0.006708223819732666, 0.0066735677719116215, 0.006653791904449463, 0.006739967823028564, 0.006717472076416015, 0.006688767910003662, 0.006701024055480957, 0.0067686400413513184, 0.006651904106140137, 0.006715392112731934, 0.00676966381072998, 0.006686719894409179, 0.006752255916595459, 0.006751232147216797, 0.006668288230895996, 0.006713344097137451, 0.006743040084838867, 0.006684671878814697, 0.006676479816436768, 0.0066713600158691405, 0.00667852783203125, 0.006761472225189209, 0.006773759841918945, 0.006666240215301514, 0.006717504024505615, 0.006680511951446534, 0.006665215969085693, 0.013929471969604493, 0.006689792156219483, 0.006700032234191895, 0.006699007987976074, 0.0066713600158691405, 0.006651904106140137, 0.006711296081542969, 0.006721536159515381, 0.006691840171813965, 0.0066938881874084475, 0.006732800006866455, 0.006694911956787109, 0.006726655960083008, 0.006738944053649902, 0.006661248207092285, 0.0067357439994812016, 0.0068178558349609375, 0.00668665599822998, 0.0067010560035705566, 0.006723584175109864, 0.006690815925598144, 0.006684671878814697, 0.006709248065948486, 0.006687744140625, 0.0067010560035705566, 0.006677504062652588, 0.006683648109436035, 0.006681600093841553, 0.006689792156219483, 0.006785024166107178, 0.006645760059356689, 0.006710271835327148, 0.006701119899749756, 0.006652863979339599, 0.006716415882110595, 0.0066979842185974124, 0.006636608123779297, 0.006700992107391357, 0.006709248065948486, 0.006657023906707763, 0.006759424209594727, 0.006717440128326416, 0.006676511764526367, 0.006853600025177002, 0.0066979842185974124, 0.006718463897705078, 0.0066938881874084475, 0.006688767910003662, 0.006677504062652588, 0.006667263984680176, 0.006691872119903564, 0.006672351837158203, 0.006681600093841553, 0.006681600093841553, 0.00666323184967041, 0.006675392150878906, 0.006703104019165039, 0.006648831844329834, 0.0066724162101745605, 0.006739935874938965, 0.006661119937896728, 0.006683648109436035, 0.006730879783630371, 0.013947775840759277, 0.006686848163604736, 0.006686592102050781, 0.006650879859924317, 0.006683648109436035, 0.006694911956787109, 0.0066375679969787596, 0.006696959972381592, 0.00674508810043335, 0.0067123198509216305, 0.006726655960083008, 0.006710271835327148, 0.006674623966217041, 0.006691648006439209, 0.006702079772949219, 0.006668288230895996, 0.0066713600158691405, 0.006670335769653321, 0.006626304149627686, 0.006653952121734619, 0.006626304149627686, 0.006692863941192627, 0.006651904106140137, 0.006779903888702392, 0.006732800006866455, 0.006685696125030518, 0.00669593620300293, 0.006669407844543457, 0.006682528018951416, 0.006718463897705078, 0.006675456047058105, 0.006657023906707763, 0.0067338237762451176, 0.0066416640281677245, 0.006668288230895996, 0.006687808036804199, 0.006668223857879639, 0.006681600093841553, 0.006662144184112549, 0.006686719894409179, 0.006680607795715332, 0.006693855762481689, 0.0066488637924194335, 0.0066877121925354, 0.006689792156219483, 0.006680575847625733, 0.006670335769653321, 0.006765567779541016, 0.006659071922302246, 0.006647903919219971, 0.006700960159301758, 0.0068351998329162595, 0.006684671878814697, 0.006675456047058105, 0.006717440128326416, 0.006665215969085693, 0.0067010560035705566, 0.006707200050354004, 0.006677599906921386, 0.006672287940979004, 0.006702079772949219, 0.006656000137329102, 0.006722559928894043, 0.013931520462036133, 0.006628352165222168, 0.006713344097137451, 0.006667263984680176, 0.006685696125030518, 0.006684671878814697, 0.006660096168518067, 0.006743040084838867, 0.006759424209594727, 0.006677504062652588, 0.0066795840263366695, 0.006682591915130615, 0.006649856090545654, 0.006677504062652588, 0.006685696125030518, 0.006699007987976074, 0.006676479816436768, 0.006668288230895996, 0.006650879859924317, 0.0067717118263244626, 0.006688767910003662, 0.006732800006866455, 0.006690815925598144, 0.00672870397567749, 0.006732800006866455, 0.006676479816436768, 0.006723584175109864, 0.006692863941192627, 0.006663167953491211, 0.006715487957000732, 0.006706079959869385, 0.006685696125030518, 0.006732800006866455, 0.006711296081542969, 0.006669312000274658, 0.006707200050354004, 0.006700032234191895, 0.006661119937896728, 0.006781951904296875, 0.006739007949829102, 0.006682559967041015, 0.006724607944488525, 0.006724607944488525, 0.006717440128326416, 0.006716415882110595, 0.006706175804138184, 0.006661119937896728, 0.0066979842185974124, 0.006747136116027832, 0.00672051191329956, 0.006669312000274658, 0.0067051520347595215, 0.0066641921997070315, 0.006706175804138184, 0.0067420158386230465, 0.0067041277885437015, 0.006687744140625, 0.006706175804138184, 0.006675456047058105, 0.006672383785247803, 0.006714367866516113, 0.0067338237762451176, 0.006692863941192627, 0.014052351951599122, 0.006694911956787109, 0.006692863941192627, 0.006696959972381592, 0.006649856090545654, 0.0068853759765625, 0.006732800006866455, 0.006765696048736572, 0.006716288089752197, 0.006732800006866455, 0.006707200050354004, 0.006683648109436035, 0.006694911956787109, 0.006776832103729248, 0.006661119937896728, 0.0067358717918396, 0.006792191982269287, 0.006668288230895996, 0.00669593620300293, 0.006722559928894043, 0.006662144184112549, 0.006714367866516113, 0.006703104019165039, 0.006690815925598144, 0.006713344097137451, 0.006913023948669434, 0.006785024166107178, 0.0066344962120056155, 0.006584320068359375, 0.006555647850036621, 0.006665215969085693, 0.006758399963378906, 0.00667955207824707, 0.006690815925598144, 0.006691840171813965, 0.006647808074951172, 0.006664256095886231, 0.006678463935852051, 0.0066406397819519045, 0.00667852783203125, 0.006691840171813965, 0.006744063854217529, 0.006685696125030518, 0.006737919807434082, 0.00675328016281128, 0.006653952121734619, 0.006710271835327148, 0.006692863941192627, 0.0066109437942504885, 0.006656000137329102, 0.006663167953491211, 0.006759583950042725, 0.006719456195831299, 0.006682496070861816, 0.006651904106140137, 0.006706175804138184, 0.006711423873901367, 0.006628223896026611, 0.006709343910217285, 0.0066620478630065915, 0.006690815925598144, 0.006714367866516113, 0.006690815925598144, 0.013918208122253419, 0.006684671878814697, 0.006651936054229736, 0.00669488000869751, 0.006721536159515381, 0.006688767910003662, 0.006646783828735352, 0.006677504062652588, 0.006688767910003662, 0.006685696125030518, 0.00672051191329956, 0.006690815925598144, 0.006683648109436035, 0.006721536159515381, 0.006638591766357422, 0.006710271835327148, 0.0067041277885437015, 0.006668288230895996, 0.006675551891326905, 0.006716320037841797, 0.006658048152923584, 0.00672156810760498, 0.006681568145751953, 0.006874112129211426, 0.006703104019165039, 0.006715424060821533, 0.006768608093261718, 0.006725632190704346, 0.00669593620300293, 0.0067123198509216305, 0.006732800006866455, 0.006859776020050049, 0.006793216228485107, 0.006746111869812011, 0.006726655960083008, 0.006719488143920899, 0.0067010560035705566, 0.00669593620300293, 0.006700032234191895, 0.006729728221893311, 0.006643712043762207, 0.006670368194580078, 0.0067040958404541015, 0.006684671878814697, 0.006714367866516113, 0.006822912216186523, 0.006681600093841553, 0.006684671878814697, 0.0067123198509216305, 0.006686719894409179, 0.0067348480224609375, 0.006756351947784424, 0.006686719894409179, 0.00673689603805542, 0.006727744102478028, 0.006672319889068604, 0.006719488143920899, 0.006743040084838867, 0.0066744318008422855, 0.00672051191329956, 0.0067338237762451176, 0.006684671878814697, 0.006699071884155273, 0.01348908805847168, 0.006379519939422608, 0.006418432235717773, 0.0063836159706115725, 0.006387712001800537, 0.006404096126556396, 0.0064204797744750975, 0.0064349122047424314, 0.00639785623550415, 0.00643174409866333, 0.006388735771179199, 0.006393983840942383, 0.00649510383605957, 0.006391808032989502, 0.006402048110961914, 0.006396927833557129, 0.006409215927124024, 0.0064174079895019534, 0.0063569917678833006, 0.006481919765472412, 0.006418432235717773, 0.006594560146331787, 0.00669593620300293, 0.0067348480224609375, 0.00673689603805542, 0.0067348480224609375, 0.006726655960083008, 0.006675456047058105, 0.006730751991271973, 0.006729728221893311, 0.006688767910003662, 0.006659071922302246, 0.006681600093841553, 0.006675456047058105, 0.0067358717918396, 0.006737919807434082, 0.006651904106140137, 0.0067123198509216305, 0.006821887969970703, 0.0066826238632202144, 0.0066724162101745605, 0.006758368015289307, 0.006688767910003662, 0.006714367866516113, 0.006686783790588379, 0.006676415920257569, 0.006684800148010254, 0.006758272171020508, 0.006636544227600098, 0.006682688236236572, 0.006684607982635498, 0.0067041277885437015, 0.00667852783203125, 0.006729728221893311, 0.006638591766357422, 0.006699007987976074, 0.006699007987976074, 0.006661119937896728, 0.006714367866516113, 0.006691840171813965, 0.006723584175109864, 0.006669312000274658, 0.006749184131622315, 0.014187520027160644, 0.006675456047058105, 0.006710271835327148, 0.006692863941192627, 0.006724607944488525, 0.006743040084838867, 0.006747136116027832, 0.006714367866516113, 0.006726655960083008, 0.0067041277885437015, 0.006681600093841553, 0.00719155216217041, 0.0068351998329162595, 0.006699007987976074, 0.006715392112731934, 0.0067358717918396, 0.0066744318008422855, 0.006648831844329834, 0.006685696125030518, 0.006673408031463623, 0.006702079772949219, 0.006715392112731934, 0.006694911956787109, 0.006682752132415771, 0.006732672214508057, 0.0066826238632202144, 0.006845439910888672, 0.006770688056945801, 0.0066406397819519045, 0.006691840171813965, 0.006719488143920899, 0.00667852783203125, 0.006643712043762207, 0.006710368156433106, 0.006732704162597657, 0.006652927875518798, 0.006635519981384277, 0.006700096130371094, 0.006661056041717529, 0.006708223819732666, 0.006662144184112549, 0.006654975891113281, 0.006691840171813965, 0.006697152137756348, 0.006677311897277832, 0.006704224109649658, 0.006729631900787353, 0.006658048152923584, 0.0066979842185974124, 0.00672051191329956, 0.0067010560035705566, 0.006829055786132812, 0.006713344097137451, 0.0067358717918396, 0.0066938881874084475, 0.006714367866516113, 0.006684671878814697, 0.006666240215301514, 0.006713344097137451, 0.006743040084838867, 0.006729728221893311, 0.006714399814605713, 0.006646751880645752, 0.01406873607635498, 0.006676479816436768, 0.0066713600158691405, 0.006714367866516113, 0.00667852783203125, 0.006646783828735352, 0.006706175804138184, 0.006762495994567871, 0.00667955207824707, 0.0067051520347595215, 0.006746111869812011, 0.006713344097137451, 0.006686719894409179, 0.006730751991271973, 0.0067010560035705566, 0.006725632190704346, 0.0067348480224609375, 0.00666534423828125, 0.006645631790161133, 0.006719488143920899, 0.006662144184112549, 0.006694911956787109, 0.006847487926483154, 0.006717440128326416, 0.006690815925598144, 0.006714367866516113, 0.006774784088134766, 0.006702079772949219, 0.006723584175109864, 0.006766687870025634, 0.006688672065734863, 0.0066979842185974124, 0.006714367866516113, 0.006767615795135498, 0.006692863941192627, 0.006688928127288819, 0.006677343845367432, 0.006686816215515137, 0.006686624050140381, 0.006662144184112549, 0.0067338237762451176, 0.006701087951660157, 0.006613984107971192, 0.006811647891998291, 0.00672051191329956, 0.0067010560035705566, 0.006690815925598144, 0.0067010560035705566, 0.006650879859924317, 0.006663167953491211, 0.006680575847625733, 0.006652927875518798, 0.006670335769653321, 0.006721536159515381, 0.006677504062652588, 0.006688767910003662, 0.006732800006866455, 0.006660096168518067, 0.006717567920684815, 0.00671014404296875, 0.006773759841918945, 0.006647808074951172, 0.00667955207824707, 0.014112768173217773, 0.006685696125030518, 0.006725632190704346, 0.006673408031463623, 0.00667955207824707, 0.00669593620300293, 0.006696000099182129, 0.006712255954742431, 0.006719488143920899, 0.006673408031463623, 0.00673689603805542, 0.006725632190704346, 0.006663167953491211, 0.006673535823822021, 0.0067029762268066406, 0.0066713600158691405, 0.006658048152923584, 0.006668288230895996, 0.006790143966674805, 0.0066744318008422855, 0.006672383785247803, 0.00669593620300293, 0.006648831844329834, 0.0067010560035705566, 0.006749311923980713, 0.006687615871429443, 0.006715392112731934, 0.006714367866516113, 0.006666240215301514, 0.006711296081542969, 0.006709248065948486, 0.006625279903411865, 0.006748159885406494, 0.006690815925598144, 0.006643712043762207, 0.006743040084838867, 0.0067686400413513184, 0.006677504062652588, 0.006657023906707763, 0.006691840171813965, 0.0066724162101745605, 0.006707168102264404, 0.0066938881874084475, 0.006723584175109864, 0.006694911956787109, 0.006730751991271973, 0.00704204797744751, 0.0068280320167541505, 0.006774784088134766, 0.007386112213134765, 0.006842368125915528, 0.006723584175109864, 0.006746111869812011, 0.006883359909057617, 0.006673376083374024, 0.006702079772949219, 0.0067338237762451176, 0.006690944194793701, 0.006689727783203125, 0.006673344135284424, 0.006700032234191895, 0.006710271835327148, 0.006791168212890625, 0.01396735954284668, 0.0067010560035705566, 0.006672383785247803, 0.0067348480224609375, 0.0066979842185974124, 0.006708223819732666, 0.006709311962127686, 0.006733759880065918, 0.00673689603805542, 0.006663167953491211, 0.006698016166687012, 0.006680543899536133, 0.006643712043762207, 0.006683680057525634, 0.006732768058776856, 0.006793216228485107, 0.006724607944488525, 0.006714367866516113, 0.006667263984680176, 0.0067010560035705566, 0.00672870397567749, 0.0065474557876586915, 0.006478847980499268, 0.006403168201446533, 0.006405024051666259, 0.006487040042877197, 0.006407167911529541, 0.006388735771179199, 0.006419456005096436, 0.006427648067474365, 0.006549503803253174, 0.00638259220123291, 0.006408192157745361, 0.006458367824554443, 0.0066375679969787596, 0.006731776237487793, 0.0067276802062988285, 0.006642687797546387, 0.006651904106140137, 0.006427775859832763, 0.006349696159362793, 0.006400000095367431, 0.006355967998504639, 0.006441984176635742, 0.0064133119583129885, 0.006390912055969238, 0.006406015872955322, 0.006377471923828125, 0.006385663986206055, 0.006400000095367431, 0.006351903915405273, 0.006428639888763427, 0.006418432235717773, 0.006396927833557129, 0.006504447937011719, 0.006367231845855713, 0.006398975849151611, 0.006345727920532227, 0.006448128223419189, 0.006415359973907471, 0.006406144142150879, 0.006386688232421875, 0.006375455856323242, 0.013490143775939942, 0.006718463897705078, 0.006758463859558105, 0.006817728042602539, 0.007229440212249756, 0.006803455829620361, 0.0067758078575134275, 0.006806528091430664, 0.006751232147216797, 0.00675328016281128, 0.006749184131622315, 0.00672870397567749, 0.006732800006866455, 0.00673689603805542, 0.006673408031463623, 0.0067358717918396, 0.00669593620300293, 0.0066713600158691405, 0.006686719894409179, 0.006726655960083008, 0.006657120227813721, 0.006680480003356934, 0.006862847805023193, 0.0067717118263244626, 0.006703167915344238, 0.006697919845581055, 0.006851583957672119, 0.006941696166992187, 0.006825984001159668, 0.006790143966674805, 0.006708223819732666, 0.006738944053649902, 0.006726655960083008, 0.006721536159515381, 0.006681600093841553, 0.0066826238632202144, 0.006710271835327148, 0.006668320178985596, 0.006714335918426514, 0.006750207901000976, 0.006654975891113281, 0.006722559928894043, 0.006718463897705078, 0.006770688056945801, 0.00673689603805542, 0.006748159885406494, 0.006687744140625, 0.006714367866516113, 0.006718463897705078, 0.006685696125030518, 0.006740992069244385, 0.006684671878814697, 0.006662144184112549, 0.006725632190704346, 0.006722559928894043, 0.006669312000274658, 0.006703104019165039, 0.006691840171813965, 0.006688767910003662, 0.006670335769653321, 0.006762495994567871, 0.006716415882110595, 0.006709311962127686, 0.013559871673583985, 0.0063630080223083495, 0.006401055812835693, 0.006364128112792969, 0.006378496170043945, 0.006428671836853027, 0.006405119895935059, 0.006403071880340576, 0.0064133119583129885, 0.006407167911529541, 0.0064338560104370114, 0.0063753600120544434, 0.006890495777130127, 0.006702144145965577, 0.006681536197662354, 0.0067123198509216305, 0.006740992069244385, 0.006849535942077637, 0.006908927917480469, 0.006751232147216797, 0.006648831844329834, 0.006726655960083008, 0.006703199863433838, 0.00665180778503418, 0.006710271835327148, 0.006716415882110595, 0.006687776088714599, 0.006649824142456054, 0.00669593620300293, 0.006662144184112549, 0.006691840171813965, 0.006699071884155273, 0.006657983779907227, 0.006730751991271973, 0.006718463897705078, 0.006675456047058105, 0.006790143966674805, 0.006713344097137451, 0.006675456047058105, 0.006669312000274658, 0.006694911956787109, 0.006726655960083008, 0.006645760059356689, 0.006724607944488525, 0.0067123198509216305, 0.006709248065948486, 0.006704256057739257, 0.006725503921508789, 0.0067573761940002445, 0.0067358717918396, 0.006694911956787109, 0.006346752166748047, 0.0063919677734375, 0.006368095874786377, 0.006377471923828125, 0.0063805441856384275, 0.006372352123260498, 0.0063836159706115725, 0.006368256092071533, 0.006396927833557129, 0.006459392070770263, 0.006370304107666015, 0.006397952079772949, 0.013467647552490235, 0.006360064029693604, 0.006369279861450195, 0.006406144142150879, 0.006436863899230957, 0.006387712001800537, 0.006416384220123291, 0.0063805441856384275, 0.0064204797744750975, 0.00642252779006958, 0.00638156795501709, 0.006411263942718506, 0.006409215927124024, 0.00642252779006958, 0.006401023864746094, 0.00658022403717041, 0.006428671836853027, 0.006408192157745361, 0.006416384220123291, 0.006427648067474365, 0.0064481601715087895, 0.0064440321922302245, 0.006395872116088867, 0.006405119895935059, 0.006472703933715821, 0.00637337589263916, 0.006561888217926025, 0.006395808219909668, 0.00638976001739502, 0.006441984176635742, 0.0064133119583129885, 0.006419456005096436, 0.00643174409866333, 0.006479872226715088, 0.006450175762176514, 0.006461440086364746, 0.006458367824554443, 0.0064143362045288085, 0.0064174079895019534, 0.006501408100128174, 0.00643174409866333, 0.006452191829681397, 0.006446080207824707, 0.006369279861450195, 0.006415359973907471, 0.006523903846740723, 0.006618112087249756, 0.0064133119583129885, 0.0063498239517211915, 0.006427648067474365, 0.006388735771179199, 0.006376448154449463, 0.006418560028076172, 0.006387648105621338, 0.006652863979339599, 0.006435840129852295, 0.006372352123260498, 0.006411263942718506, 0.006370304107666015, 0.006418432235717773, 0.006466559886932373, 0.006515711784362793, 0.006414400100708008, 0.013413311958312989, 0.00638156795501709, 0.006359039783477783, 0.0063610877990722655, 0.006385663986206055, 0.006392831802368164, 0.0063907837867736815, 0.006372352123260498, 0.006404096126556396, 0.006437888145446777, 0.006436863899230957, 0.0064849920272827145, 0.006358016014099121, 0.006408192157745361, 0.0064143362045288085, 0.0063805441856384275, 0.006455296039581298, 0.006346752166748047, 0.006367231845855713, 0.006415359973907471, 0.006332511901855469, 0.006388639926910401, 0.006375423908233643, 0.006386688232421875, 0.0064245758056640625, 0.00637337589263916, 0.006436863899230957, 0.006405119895935059, 0.006429696083068847, 0.0064430079460144046, 0.00637440013885498, 0.006403071880340576, 0.006553728103637696, 0.006393727779388428, 0.006428671836853027, 0.006375423908233643, 0.00638259220123291, 0.0063836159706115725, 0.006385663986206055, 0.00638976001739502, 0.00638156795501709, 0.006717440128326416, 0.0066713600158691405, 0.006684671878814697, 0.006789120197296142, 0.0067338237762451176, 0.0066713600158691405, 0.0067051520347595215, 0.006668447971343994, 0.006643551826477051, 0.006669312000274658, 0.006687839984893799, 0.006634399890899658, 0.006789120197296142, 0.006681600093841553, 0.006656000137329102, 0.0067010560035705566, 0.0066979842185974124, 0.006663167953491211, 0.0067358717918396, 0.006692863941192627, 0.006619232177734375, 0.0066877121925354, 0.013569055557250976, 0.006402976036071777, 0.006408192157745361, 0.00637337589263916, 0.006427648067474365, 0.006379519939422608, 0.0064143362045288085, 0.006432767868041992, 0.006418432235717773, 0.006388735771179199, 0.0063836159706115725, 0.006388735771179199, 0.006401023864746094, 0.006360064029693604, 0.00637440013885498, 0.0063907837867736815, 0.0067051839828491215, 0.007263199806213379, 0.006842368125915528, 0.007316480159759522, 0.007303167819976806, 0.006818816184997558, 0.006731776237487793, 0.006673408031463623, 0.006744063854217529, 0.006691840171813965, 0.006677504062652588, 0.00672976016998291, 0.00670307207107544, 0.006702079772949219, 0.0067041277885437015, 0.006675456047058105, 0.006670335769653321, 0.006703104019165039, 0.0066938881874084475, 0.006686719894409179, 0.0067051520347595215, 0.006708223819732666, 0.006708223819732666, 0.006668288230895996, 0.0066744318008422855, 0.006667263984680176, 0.006685696125030518, 0.006659071922302246, 0.006684671878814697, 0.0066641921997070315, 0.00669593620300293, 0.006651904106140137, 0.006668288230895996, 0.00672051191329956, 0.006702079772949219, 0.006689792156219483, 0.00667955207824707, 0.00667955207824707, 0.00666537618637085, 0.006656864166259766, 0.006689792156219483, 0.006815743923187256, 0.006721536159515381, 0.006702079772949219, 0.006723584175109864, 0.006739967823028564, 0.006696959972381592, 0.013950976371765136, 0.006729728221893311, 0.00667955207824707, 0.006680575847625733, 0.0067041277885437015, 0.006694911956787109, 0.006665215969085693, 0.006714367866516113, 0.0066744318008422855, 0.00675328016281128, 0.006763519763946534, 0.006645760059356689, 0.006665215969085693, 0.006731776237487793, 0.006690815925598144, 0.006683648109436035, 0.0067461438179016115, 0.006725599765777588, 0.0066796159744262696, 0.006673344135284424, 0.006725632190704346, 0.006654975891113281, 0.006694911956787109, 0.006683648109436035, 0.006662144184112549, 0.00669593620300293, 0.006683648109436035, 0.006648831844329834, 0.006684671878814697, 0.0067010560035705566, 0.006656000137329102, 0.006702079772949219, 0.0066938881874084475, 0.006720608234405518, 0.006694816112518311, 0.006715392112731934, 0.006842368125915528, 0.006729728221893311, 0.0067051520347595215, 0.006724639892578125, 0.006747104167938232, 0.006719488143920899, 0.006684671878814697, 0.0066641921997070315, 0.006685728073120117, 0.006712384223937988, 0.006687647819519043, 0.006707200050354004, 0.006636544227600098, 0.006686719894409179, 0.006706175804138184, 0.0066447358131408694, 0.006670335769653321, 0.006709248065948486, 0.006670368194580078, 0.006697951793670654, 0.006699007987976074, 0.0067041277885437015, 0.0066938881874084475, 0.0066826238632202144, 0.006658048152923584, 0.006691840171813965, 0.006648831844329834]",tokens/s,147.84298446360467,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Salesforce/codegen-6B-nl,Salesforce/codegen-6B-nl,cuda,0,42,,,True,,,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking context_stack.enter_context(energy_tracker.track()) File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context result = _cm_type.__enter__(cm) File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ return next(self.gen) File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track self.emission_tracker.start_task() File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",codegen,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-410m,EleutherAI/pythia-410m,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.42.1,,0.31.0,,,,1.20.0,,,,0.11.1,,,,MB,1263.976448,1711.80032,0.0,1065.353216,1018.332672,s,10,0.16609878349304202,0.0166098783493042,0.0005825325938120824,0.01635870361328125,0.017076227951049804,0.01764259376525879,0.018095686416625976,"[0.018208959579467773, 0.01631056022644043, 0.016950368881225586, 0.016470848083496095, 0.016225120544433595, 0.01622275161743164, 0.01677961540222168, 0.016333791732788087, 0.016383615493774414, 0.016213151931762697]",tokens/s,15412.515047753135,kWh,1.940086182180362e-07,1.0630672494639064e-07,5.477097977733979e-07,8.480251409378246e-07,tokens/kWh,301877842.580105,MB,1263.976448,1711.80032,0.0,1065.353216,1033.284608,s,10,10.226998229980468,1.0226998229980468,0.017756554434263713,1.01632958984375,1.0387885375976562,1.0531606872558594,1.0646584069824219,"[1.0675328369140624, 1.033716796875, 1.0355947265625, 1.0062955322265625, 1.0127122802734374, 1.0208916625976563, 1.0137723388671875, 1.0188868408203124, 1.0098925170898438, 1.0077026977539063]",tokens/s,61.60165337206704,kWh,1.1959291892294423e-05,6.5532007626287934e-06,2.3424531602226823e-05,4.1937024257150035e-05,tokens/kWh,1502252.511138981,,s,629,10.35675953960418,0.016465436469958964,0.0020292390118555937,0.01602355194091797,0.016716390228271486,0.017295769119262696,0.032694149627685555,"[0.01760767936706543, 0.017529855728149413, 0.016986112594604492, 0.01661644744873047, 0.017106943130493164, 0.017293312072753905, 0.017542144775390626, 0.017348608016967772, 0.017324031829833983, 0.017261568069458007, 0.017297407150268555, 0.0173885440826416, 0.017464319229125978, 0.01737113571166992, 0.01683967971801758, 0.016736255645751954, 0.0174653434753418, 0.01743052864074707, 0.017330175399780275, 0.01767628860473633, 0.01726361656188965, 0.017301504135131835, 0.017254432678222655, 0.0173024959564209, 0.017138687133789063, 0.017168384552001953, 0.01745305633544922, 0.017500160217285156, 0.017147903442382813, 0.016727039337158203, 0.01665843200683594, 0.016689151763916017, 0.01665433692932129, 0.016746496200561522, 0.016579584121704103, 0.01656729507446289, 0.016611328125, 0.016594944000244142, 0.016574464797973632, 0.01661644744873047, 0.016563199996948243, 0.01659187126159668, 0.01661644744873047, 0.01660927963256836, 0.016528383255004882, 0.016590848922729492, 0.01660211181640625, 0.01641472053527832, 0.016441343307495117, 0.0164454402923584, 0.016611328125, 0.016657407760620118, 0.016694271087646484, 0.01660006332397461, 0.016289791107177733, 0.01680281639099121, 0.016676864624023437, 0.016702463150024414, 0.016699392318725585, 0.01662259292602539, 0.01656937599182129, 0.016607200622558594, 0.03276595306396484, 0.01590988826751709, 0.015965184211730956, 0.015954943656921388, 0.015996928215026854, 0.016029695510864257, 0.01605734443664551, 0.01600307273864746, 0.01602355194091797, 0.015965184211730956, 0.01604915237426758, 0.015961088180541993, 0.015981568336486816, 0.016020479202270507, 0.015978495597839357, 0.016232448577880858, 0.016110591888427735, 0.016605215072631838, 0.016591840744018555, 0.016675840377807616, 0.016615423202514648, 0.01632972717285156, 0.016498687744140626, 0.016488447189331054, 0.01639219284057617, 0.01636147117614746, 0.01638809585571289, 0.016360448837280273, 0.016095232009887696, 0.01658470344543457, 0.01662566375732422, 0.016534528732299804, 0.016574464797973632, 0.01660927963256836, 0.016270336151123048, 0.016745471954345705, 0.016744447708129884, 0.01658060836791992, 0.016684032440185546, 0.016524288177490236, 0.01637990379333496, 0.016471040725708007, 0.01663283157348633, 0.01656729507446289, 0.016690176010131837, 0.016547840118408205, 0.01660416030883789, 0.016683008193969725, 0.016561151504516602, 0.01660316848754883, 0.016573408126831054, 0.016564224243164064, 0.01658060836791992, 0.0166430721282959, 0.016527360916137695, 0.016546815872192384, 0.01660825538635254, 0.016547840118408205, 0.016500736236572267, 0.01660211181640625, 0.016545791625976563, 0.01662054443359375, 0.01654374313354492, 0.03390259170532227, 0.01662873649597168, 0.016540672302246092, 0.01664614486694336, 0.016659456253051756, 0.016869375228881836, 0.01661440086364746, 0.01661440086364746, 0.01657548713684082, 0.016746496200561522, 0.016655359268188476, 0.01658163261413574, 0.016571392059326173, 0.016662527084350585, 0.01661644744873047, 0.01665433692932129, 0.016665599822998048, 0.01658572769165039, 0.016694271087646484, 0.016713727951049806, 0.016781312942504883, 0.01663283157348633, 0.01664204788208008, 0.016563199996948243, 0.016571392059326173, 0.01657548713684082, 0.016700416564941405, 0.016732160568237304, 0.01656934356689453, 0.0166246395111084, 0.01653555107116699, 0.016532480239868166, 0.016571392059326173, 0.016506879806518555, 0.01661440086364746, 0.016534528732299804, 0.01661952018737793, 0.01661849594116211, 0.01665023994445801, 0.016550912857055664, 0.01661644744873047, 0.01659395217895508, 0.016592863082885743, 0.01656729507446289, 0.015849472045898438, 0.01600921630859375, 0.015927295684814453, 0.015919103622436523, 0.015989760398864745, 0.01600102424621582, 0.01600102424621582, 0.01600921630859375, 0.01597439956665039, 0.015964159965515135, 0.015985664367675782, 0.01595084762573242, 0.015966208457946777, 0.015940608024597167, 0.01586995220184326, 0.015958016395568847, 0.015988736152648925, 0.01597644805908203, 0.016564224243164064, 0.032889854431152346, 0.015995903968811034, 0.015969280242919923, 0.015965184211730956, 0.01596723175048828, 0.01595084762573242, 0.015749119758605958, 0.01579417610168457, 0.01592627239227295, 0.015987711906433104, 0.015930368423461915, 0.016010240554809572, 0.01597542381286621, 0.016, 0.01598259162902832, 0.015969280242919923, 0.015965184211730956, 0.015924223899841307, 0.01599180793762207, 0.015910911560058593, 0.015987711906433104, 0.015905792236328126, 0.015930368423461915, 0.015951871871948242, 0.015880191802978515, 0.01600102424621582, 0.015941632270812987, 0.015922176361083985, 0.015987711906433104, 0.015968255996704102, 0.01603379249572754, 0.01596723175048828, 0.01597644805908203, 0.015987711906433104, 0.015956992149353026, 0.016004095077514647, 0.015888383865356445, 0.015977472305297852, 0.016045055389404296, 0.015963135719299318, 0.015932415962219237, 0.016046079635620117, 0.015949824333190917, 0.016035839080810545, 0.015949824333190917, 0.015980544090270995, 0.015948800086975096, 0.015968255996704102, 0.01593449592590332, 0.01594979190826416, 0.01598259162902832, 0.016676864624023437, 0.016257024765014647, 0.01597644805908203, 0.01588633632659912, 0.0159334716796875, 0.0159835844039917, 0.015900671958923338, 0.01581670379638672, 0.015912960052490235, 0.0158853120803833, 0.015933440208435058, 0.015905792236328126, 0.03259904098510742, 0.015889408111572266, 0.01593855953216553, 0.015904767990112305, 0.01595084762573242, 0.015944704055786133, 0.01586995220184326, 0.016160768508911134, 0.017145856857299805, 0.018129919052124025, 0.016694271087646484, 0.015969280242919923, 0.016030719757080078, 0.015985664367675782, 0.016044031143188475, 0.016044031143188475, 0.015908864021301268, 0.01597644805908203, 0.01598361587524414, 0.015981568336486816, 0.01598259162902832, 0.015989760398864745, 0.016003103256225587, 0.015947744369506835, 0.016829439163208008, 0.016663551330566406, 0.01613209533691406, 0.01577676773071289, 0.016021503448486327, 0.01601638412475586, 0.016038911819458008, 0.015929344177246094, 0.015953920364379884, 0.015952896118164063, 0.01601433563232422, 0.016093183517456054, 0.016250911712646483, 0.015987680435180663, 0.015954943656921388, 0.016184320449829103, 0.01601638412475586, 0.01595084762573242, 0.015960063934326172, 0.01600819206237793, 0.015949824333190917, 0.015948800086975096, 0.01591603183746338, 0.015980544090270995, 0.01598259162902832, 0.015985664367675782, 0.015977472305297852, 0.01593855953216553, 0.015948800086975096, 0.01597439956665039, 0.015932415962219237, 0.015970303535461427, 0.016074752807617186, 0.01590272045135498, 0.01593139171600342, 0.015929344177246094, 0.015945728302001954, 0.01594265556335449, 0.015920127868652344, 0.032573440551757815, 0.01592319965362549, 0.015912960052490235, 0.01592627239227295, 0.015946751594543458, 0.015921152114868165, 0.015965184211730956, 0.015910911560058593, 0.015851519584655763, 0.01574092769622803, 0.015783935546875, 0.015734784126281737, 0.015812607765197755, 0.015734784126281737, 0.01579212760925293, 0.01590681552886963, 0.015994879722595216, 0.015977472305297852, 0.016728063583374024, 0.016283647537231445, 0.015919103622436523, 0.015726592063903807, 0.015820799827575685, 0.015820799827575685, 0.016055295944213867, 0.015924223899841307, 0.016076799392700195, 0.01881395149230957, 0.018345983505249023, 0.016921600341796874, 0.016772096633911132, 0.01664204788208008, 0.01657142448425293, 0.01664508819580078, 0.01658572769165039, 0.016678911209106445, 0.01597542381286621, 0.01639833641052246, 0.016562175750732423, 0.01657651138305664, 0.016091136932373046, 0.016226303100585936, 0.016558080673217773, 0.016389120101928712, 0.016088064193725587, 0.015978495597839357, 0.015940608024597167, 0.016037919998168945, 0.01745097541809082, 0.016530431747436524, 0.015962112426757814, 0.01603276824951172, 0.015813632011413573, 0.015948800086975096, 0.01576140785217285, 0.015796223640441896, 0.01586995220184326, 0.016030719757080078, 0.015960063934326172, 0.01593855953216553, 0.015954943656921388, 0.01600819206237793, 0.016515071868896485, 0.033230846405029296, 0.015933440208435058, 0.015948800086975096, 0.015928319931030274, 0.015964159965515135, 0.015930368423461915, 0.01592627239227295, 0.016004095077514647, 0.016450559616088867, 0.01641372871398926, 0.0159518404006958, 0.01725132751464844, 0.016330751419067382, 0.01602560043334961, 0.016022527694702148, 0.01599180793762207, 0.01602355194091797, 0.0158853120803833, 0.01656729507446289, 0.016276479721069336, 0.01618124771118164, 0.015865856170654297, 0.015953920364379884, 0.015900671958923338, 0.015747072219848633, 0.01581670379638672, 0.016061439514160156, 0.01638502311706543, 0.01637580871582031, 0.016380928039550782, 0.016061439514160156, 0.015820799827575685, 0.015757311820983888, 0.015965184211730956, 0.015978495597839357, 0.015962112426757814, 0.015742976188659667, 0.01577779197692871, 0.015805439949035646, 0.016184320449829103, 0.015963135719299318, 0.01575014400482178, 0.016035839080810545, 0.01640243148803711, 0.01599180793762207, 0.015965184211730956, 0.015961088180541993, 0.01599078369140625, 0.015941632270812987, 0.01604812812805176, 0.01602560043334961, 0.01600716781616211, 0.015969280242919923, 0.016527360916137695, 0.017104896545410156, 0.016630783081054687, 0.01622528076171875, 0.016101375579833984, 0.016063488006591797, 0.016096256256103517, 0.01599897575378418, 0.01600819206237793, 0.015963135719299318, 0.03273113632202149, 0.015783935546875, 0.015962112426757814, 0.01580031967163086, 0.01619353675842285, 0.015873023986816406, 0.01618534469604492, 0.016078847885131836, 0.016069631576538086, 0.015839232444763183, 0.01619152069091797, 0.016012256622314452, 0.016052223205566405, 0.015961088180541993, 0.01590783977508545, 0.01581158447265625, 0.01590272045135498, 0.016950271606445313, 0.01622323226928711, 0.01584332847595215, 0.016097280502319337, 0.016053247451782226, 0.01590783977508545, 0.015989760398864745, 0.01587712001800537, 0.01880985641479492, 0.01678233528137207, 0.016267263412475585, 0.016047103881835938, 0.016043008804321288, 0.015925248146057128, 0.016110591888427735, 0.015900671958923338, 0.016876544952392578, 0.01662054443359375, 0.01744895935058594, 0.016228384017944335, 0.01651606369018555, 0.01617919921875, 0.01597542381286621, 0.016046079635620117, 0.01617919921875, 0.016124927520751953, 0.016679935455322266, 0.016137216567993166, 0.015896575927734375, 0.016012287139892577, 0.01590169620513916, 0.01600921630859375, 0.016183296203613282, 0.016611328125, 0.016176128387451173, 0.016191488265991212, 0.01600921630859375, 0.016082944869995116, 0.01599180793762207, 0.016047103881835938, 0.015945728302001954, 0.015977472305297852, 0.0160020809173584, 0.01608803176879883, 0.01599283218383789, 0.01597644805908203, 0.03278031921386719, 0.01597436809539795, 0.016074752807617186, 0.015968255996704102, 0.016107519149780272, 0.016207872390747072, 0.016061439514160156, 0.015945728302001954, 0.016012287139892577, 0.01595903968811035, 0.01601638412475586, 0.015972352027893065, 0.015996928215026854, 0.015980544090270995, 0.01600102424621582, 0.016022527694702148, 0.016036863327026366, 0.015986687660217287, 0.015952927589416504, 0.015979488372802733, 0.01601535987854004, 0.015933440208435058, 0.016046079635620117, 0.015949824333190917, 0.01601740837097168, 0.016528383255004882, 0.016220159530639648, 0.016, 0.016183296203613282, 0.016153600692749022, 0.01584332847595215, 0.015836159706115724, 0.01613926315307617, 0.016110591888427735, 0.01600819206237793, 0.016071680068969727, 0.016082944869995116, 0.01603993606567383, 0.016061439514160156, 0.01602764892578125, 0.016098304748535155, 0.016119808197021485, 0.016043008804321288, 0.01600614356994629, 0.015857664108276368, 0.015812607765197755, 0.015898624420166017, 0.015981568336486816, 0.015994879722595216, 0.01599897575378418, 0.016012287139892577, 0.015962112426757814, 0.016021503448486327, 0.016004095077514647, 0.016083967208862304, 0.01616998481750488, 0.01597439956665039, 0.016055295944213867, 0.016048160552978516, 0.015952863693237306, 0.01593855953216553, 0.015981568336486816, 0.015989760398864745, 0.032860160827636715, 0.016034816741943358, 0.016002048492431642, 0.01601638412475586, 0.016166912078857423, 0.016061439514160156, 0.01595903968811035, 0.01594265556335449, 0.01595084762573242, 0.016054271697998047, 0.015973376274108886, 0.016026624679565428, 0.015806464195251464, 0.01588633632659912, 0.01600819206237793, 0.015847423553466796, 0.01580953598022461, 0.01677926445007324, 0.01621708869934082, 0.016022527694702148, 0.01581056022644043, 0.015900671958923338, 0.015827967643737794, 0.015796223640441896, 0.015764479637145997, 0.0158023681640625, 0.015936511993408203, 0.01624575996398926, 0.016366592407226564, 0.016100351333618163, 0.016248832702636717, 0.016079872131347657, 0.015961088180541993, 0.01600102424621582, 0.015949824333190917, 0.015988736152648925, 0.015970303535461427, 0.01600716781616211, 0.015951871871948242, 0.016069631576538086, 0.01604198455810547, 0.015946751594543458, 0.015880191802978515, 0.01581772804260254, 0.015804415702819825, 0.01582592010498047, 0.01582592010498047, 0.01581875228881836, 0.015952896118164063, 0.01597644805908203, 0.016021503448486327, 0.01614028739929199, 0.016011327743530274, 0.015975359916687012, 0.01595903968811035, 0.01601740837097168, 0.01583513641357422, 0.015945728302001954, 0.01602457618713379, 0.01602457618713379, 0.01597542381286621, 0.015996928215026854, 0.015920127868652344]",tokens/s,60.733282219666066,,,,,,,,,,,,,,,,,,,,,,,,,,,,